@inproceedings{zeng-etal-2022-task,
title = "Task-guided Disentangled Tuning for Pretrained Language Models",
author = "Zeng, Jiali and
Jiang, Yufan and
Wu, Shuangzhi and
Yin, Yongjing and
Li, Mu",
editor = "Muresan, Smaranda and
Nakov, Preslav and
Villavicencio, Aline",
booktitle = "Findings of the Association for Computational Linguistics: ACL 2022",
month = may,
year = "2022",
address = "Dublin, Ireland",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.247/",
doi = "10.18653/v1/2022.findings-acl.247",
pages = "3126--3137",
abstract = "Pretrained language models (PLMs) trained on large-scale unlabeled corpus are typically fine-tuned on task-specific downstream datasets, which have produced state-of-the-art results on various NLP tasks. However, the data discrepancy issue in domain and scale makes fine-tuning fail to efficiently capture task-specific patterns, especially in low data regime. To address this issue, we propose Task-guided Disentangled Tuning (TDT) for PLMs, which enhances the generalization of representations by disentangling task-relevant signals from the entangled representations. For a given task, we introduce a learnable confidence model to detect indicative guidance from context, and further propose a disentangled regularization to mitigate the over-reliance problem. Experimental results on GLUE and CLUE benchmarks show that TDT gives consistently better results than fine-tuning with different PLMs, and extensive analysis demonstrates the effectiveness and robustness of our method. Code is available at \url{https://github.com/lemon0830/TDT}."
}
Markdown (Informal)
[Task-guided Disentangled Tuning for Pretrained Language Models](https://preview.aclanthology.org/fix-sig-urls/2022.findings-acl.247/) (Zeng et al., Findings 2022)
ACL