Impartial Multi-task Representation Learning via Variance-invariant Probabilistic Decoding

Dou Hu, Lingwei Wei, Wei Zhou, Songlin Hu


Abstract
Multi-task learning (MTL) enhances efficiency by sharing representations across tasks, but task dissimilarities often cause partial learning, where some tasks dominate while others are neglected. Existing methods mainly focus on balancing loss or gradients but fail to fundamentally address this issue due to the representation discrepancy in latent space. In this paper, we propose variance-invariant probabilistic decoding for multi-task learning (VIP-MTL), a framework that ensures impartial learning by harmonizing representation spaces across tasks. VIP-MTL decodes shared representations into task-specific probabilistic distributions and applies variance normalization to constrain these distributions to a consistent scale. Experiments on two language benchmarks show that VIP-MTL outperforms 12 representative methods under the same multi-task settings, especially in heterogeneous task combinations and data-constrained scenarios. Further analysis shows that VIP-MTL is robust to sampling distributions, efficient on optimization process, and scale-invariant to task losses. Additionally, the learned task-specific representations are more informative, enhancing the language understanding abilities of pre-trained language models under the multi-task paradigm.
Anthology ID:
2025.acl-long.975
Volume:
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
July
Year:
2025
Address:
Vienna, Austria
Editors:
Wanxiang Che, Joyce Nabende, Ekaterina Shutova, Mohammad Taher Pilehvar
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
19883–19897
Language:
URL:
https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.975/
DOI:
Bibkey:
Cite (ACL):
Dou Hu, Lingwei Wei, Wei Zhou, and Songlin Hu. 2025. Impartial Multi-task Representation Learning via Variance-invariant Probabilistic Decoding. In Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 19883–19897, Vienna, Austria. Association for Computational Linguistics.
Cite (Informal):
Impartial Multi-task Representation Learning via Variance-invariant Probabilistic Decoding (Hu et al., ACL 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingestion-acl-25/2025.acl-long.975.pdf