@inproceedings{dai-etal-2018-credit,
title = "From Credit Assignment to Entropy Regularization: Two New Algorithms for Neural Sequence Prediction",
author = "Dai, Zihang and
Xie, Qizhe and
Hovy, Eduard",
editor = "Gurevych, Iryna and
Miyao, Yusuke",
booktitle = "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2018",
address = "Melbourne, Australia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/P18-1155/",
doi = "10.18653/v1/P18-1155",
pages = "1672--1682",
abstract = "In this work, we study the credit assignment problem in reward augmented maximum likelihood (RAML) learning, and establish a theoretical equivalence between the token-level counterpart of RAML and the entropy regularized reinforcement learning. Inspired by the connection, we propose two sequence prediction algorithms, one extending RAML with fine-grained credit assignment and the other improving Actor-Critic with a systematic entropy regularization. On two benchmark datasets, we show the proposed algorithms outperform RAML and Actor-Critic respectively, providing new alternatives to sequence prediction."
}
Markdown (Informal)
[From Credit Assignment to Entropy Regularization: Two New Algorithms for Neural Sequence Prediction](https://preview.aclanthology.org/jlcl-multiple-ingestion/P18-1155/) (Dai et al., ACL 2018)
ACL