@inproceedings{ke-etal-2022-domain,
title = "Domain-Aware Contrastive Knowledge Transfer for Multi-domain Imbalanced Data",
author = "Ke, Zixuan and
Kachuee, Mohammad and
Lee, Sungjin",
editor = "Barnes, Jeremy and
De Clercq, Orph{\'e}e and
Barriere, Valentin and
Tafreshi, Shabnam and
Alqahtani, Sawsan and
Sedoc, Jo{\~a}o and
Klinger, Roman and
Balahur, Alexandra",
booktitle = "Proceedings of the 12th Workshop on Computational Approaches to Subjectivity, Sentiment {\&} Social Media Analysis",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/Author-page-Marten-During-lu/2022.wassa-1.3/",
doi = "10.18653/v1/2022.wassa-1.3",
pages = "25--36",
abstract = "In many real-world machine learning applications, samples belong to a set of domains e.g., for product reviews each review belongs to a product category. In this paper, we study multi-domain imbalanced learning (MIL), the scenario that there is imbalance not only in classes but also in domains. In the MIL setting, different domains exhibit different patterns and there is a varying degree of similarity and divergence among domains posing opportunities and challenges for transfer learning especially when faced with limited or insufficient training data. We propose a novel domain-aware contrastive knowledge transfer method called DCMI to (1) identify the shared domain knowledge to encourage positive transfer among similar domains (in particular from head domains to tail domains); (2) isolate the domain-specific knowledge to minimize the negative transfer from dissimilar domains. We evaluated the performance of DCMI on three different datasets showing significant improvements in different MIL scenarios."
}
Markdown (Informal)
[Domain-Aware Contrastive Knowledge Transfer for Multi-domain Imbalanced Data](https://preview.aclanthology.org/Author-page-Marten-During-lu/2022.wassa-1.3/) (Ke et al., WASSA 2022)
ACL