On the Compositional Generalization in Versatile Open-domain Dialogue

Tingchen Fu, Xueliang Zhao, Lemao Liu, Rui Yan


Abstract
Previous research has demonstrated the potential of multi-task learning to foster a conversational agent’s ability to acquire a variety of skills. However, these approaches either suffer from interference among different datasets (also known as negative transfer), or fail to effectively reuse knowledge and skills learned from other datasets. In contrast to previous works, we develop a sparsely activated modular network: (1) We propose a well-rounded set of operators and instantiate each operator with an independent module; (2) We formulate dialogue generation as the execution of a generated programme which recursively composes and assembles modules. Extensive experiments on 9 datasets verify the efficacy of our methods through automatic evaluation and human evaluation. Notably, our model outperforms state-of-the-art supervised approaches on 4 datasets with only 10% training data thanks to the modular architecture and multi-task learning.
Anthology ID:
2023.acl-long.760
Volume:
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Month:
July
Year:
2023
Address:
Toronto, Canada
Editors:
Anna Rogers, Jordan Boyd-Graber, Naoaki Okazaki
Venue:
ACL
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
13585–13605
Language:
URL:
https://aclanthology.org/2023.acl-long.760
DOI:
10.18653/v1/2023.acl-long.760
Bibkey:
Cite (ACL):
Tingchen Fu, Xueliang Zhao, Lemao Liu, and Rui Yan. 2023. On the Compositional Generalization in Versatile Open-domain Dialogue. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 13585–13605, Toronto, Canada. Association for Computational Linguistics.
Cite (Informal):
On the Compositional Generalization in Versatile Open-domain Dialogue (Fu et al., ACL 2023)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-2024-clasp/2023.acl-long.760.pdf
Video:
 https://preview.aclanthology.org/ingest-2024-clasp/2023.acl-long.760.mp4