@inproceedings{wang-etal-2023-dialog,
title = "Dialog Action-Aware Transformer for Dialog Policy Learning",
author = "Wang, Huimin and
Kwan, Wai Chung and
Wong, Kam-Fai",
editor = "Stoyanchev, Svetlana and
Joty, Shafiq and
Schlangen, David and
Dusek, Ondrej and
Kennington, Casey and
Alikhani, Malihe",
booktitle = "Proceedings of the 24th Annual Meeting of the Special Interest Group on Discourse and Dialogue",
month = sep,
year = "2023",
address = "Prague, Czechia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2023.sigdial-1.12/",
doi = "10.18653/v1/2023.sigdial-1.12",
pages = "142--148",
abstract = "Recent works usually address Dialog policy learning DPL by training a reinforcement learning (RL) agent to determine the best dialog action. However, existing works on deep RL require a large volume of agent-user interactions to achieve acceptable performance. In this paper, we propose to make full use of the plain text knowledge from the pre-trained language model to accelerate the RL agent`s learning speed. Specifically, we design a dialog action-aware transformer encoder (DaTrans), which integrates a new fine-tuning procedure named masked last action task to encourage DaTrans to be dialog-aware and distill action-specific features. Then, DaTrans is further optimized in an RL setting with ongoing interactions and evolves through exploration in the dialog action space toward maximizing long-term accumulated rewards. The effectiveness and efficiency of the proposed model are demonstrated with both simulator evaluation and human evaluation."
}
Markdown (Informal)
[Dialog Action-Aware Transformer for Dialog Policy Learning](https://preview.aclanthology.org/add-emnlp-2024-awards/2023.sigdial-1.12/) (Wang et al., SIGDIAL 2023)
ACL