@inproceedings{sharma-etal-2017-speeding,
title = "Speeding up Reinforcement Learning-based Information Extraction Training using Asynchronous Methods",
author = "Sharma, Aditya and
Parekh, Zarana and
Talukdar, Partha",
editor = "Palmer, Martha and
Hwa, Rebecca and
Riedel, Sebastian",
booktitle = "Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing",
month = sep,
year = "2017",
address = "Copenhagen, Denmark",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/D17-1281/",
doi = "10.18653/v1/D17-1281",
pages = "2658--2663",
abstract = "RLIE-DQN is a recently proposed Reinforcement Learning-based Information Extraction (IE) technique which is able to incorporate external evidence during the extraction process. RLIE-DQN trains a single agent sequentially, training on one instance at a time. This results in significant training slowdown which is undesirable. We leverage recent advances in parallel RL training using asynchronous methods and propose RLIE-A3C. RLIE-A3C trains multiple agents in parallel and is able to achieve upto 6x training speedup over RLIE-DQN, while suffering no loss in average accuracy."
}
Markdown (Informal)
[Speeding up Reinforcement Learning-based Information Extraction Training using Asynchronous Methods](https://preview.aclanthology.org/fix-sig-urls/D17-1281/) (Sharma et al., EMNLP 2017)
ACL