@inproceedings{du-ji-2021-sidecontrol-controlled,
title = "{S}ide{C}ontrol: Controlled Open-domain Dialogue Generation via Additive Side Networks",
author = "Du, Wanyu and
Ji, Yangfeng",
editor = "Moens, Marie-Francine and
Huang, Xuanjing and
Specia, Lucia and
Yih, Scott Wen-tau",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2021",
month = nov,
year = "2021",
address = "Punta Cana, Dominican Republic",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2021.findings-emnlp.188/",
doi = "10.18653/v1/2021.findings-emnlp.188",
pages = "2175--2194",
abstract = "Transformer-based pre-trained language models boost the performance of open-domain dialogue systems. Prior works leverage Transformer-based pre-trained language models to generate texts with desired attributes in two general approaches: (1) gradient-based methods: updating all latent representations of pre-trained models with gradients from attribute models; (2) weighted-decoding methods: re-ranking beam candidates from pre-trained models with attribute functions. However, gradient-based methods lead to high computation cost and can easily get overfitted on small training sets, while weighted-decoding methods are inherently constrained by the low-variance high-bias pre-trained model. In this work, we propose a novel approach to control the generation of Transformer-based pre-trained language models: the SideControl framework, which leverages a novel control attributes loss to incorporate useful control signals, and is shown to perform well with very limited training samples. We evaluate our proposed method on two benchmark open-domain dialogue datasets, and results show that the SideControl framework has better controllability, higher generation quality and better sample-efficiency than existing gradient-based and weighted-decoding baselines."
}
Markdown (Informal)
[SideControl: Controlled Open-domain Dialogue Generation via Additive Side Networks](https://preview.aclanthology.org/add-emnlp-2024-awards/2021.findings-emnlp.188/) (Du & Ji, Findings 2021)
ACL