Abstract
Pruning hypotheses during dynamic programming is commonly used to speed up inference in settings such as parsing. Unlike prior work, we train a pruning policy under an objective that measures end-to-end performance: we search for a fast and accurate policy. This poses a difficult machine learning problem, which we tackle with the lols algorithm. lols training must continually compute the effects of changing pruning decisions: we show how to make this efficient in the constituency parsing setting, via dynamic programming and change propagation algorithms. We find that optimizing end-to-end performance in this way leads to a better Pareto frontier—i.e., parsers which are more accurate for a given runtime.- Anthology ID:
- Q17-1019
- Volume:
- Transactions of the Association for Computational Linguistics, Volume 5
- Month:
- Year:
- 2017
- Address:
- Cambridge, MA
- Venue:
- TACL
- SIG:
- Publisher:
- MIT Press
- Note:
- Pages:
- 263–278
- Language:
- URL:
- https://aclanthology.org/Q17-1019
- DOI:
- 10.1162/tacl_a_00060
- Cite (ACL):
- Tim Vieira and Jason Eisner. 2017. Learning to Prune: Exploring the Frontier of Fast and Accurate Parsing. Transactions of the Association for Computational Linguistics, 5:263–278.
- Cite (Informal):
- Learning to Prune: Exploring the Frontier of Fast and Accurate Parsing (Vieira & Eisner, TACL 2017)
- PDF:
- https://preview.aclanthology.org/ingestion-script-update/Q17-1019.pdf
- Data
- Penn Treebank