@inproceedings{ge-etal-2018-fluency,
title = "Fluency Boost Learning and Inference for Neural Grammatical Error Correction",
author = "Ge, Tao and
Wei, Furu and
Zhou, Ming",
editor = "Gurevych, Iryna and
Miyao, Yusuke",
booktitle = "Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)",
month = jul,
year = "2018",
address = "Melbourne, Australia",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/P18-1097/",
doi = "10.18653/v1/P18-1097",
pages = "1055--1065",
abstract = "Most of the neural sequence-to-sequence (seq2seq) models for grammatical error correction (GEC) have two limitations: (1) a seq2seq model may not be well generalized with only limited error-corrected data; (2) a seq2seq model may fail to completely correct a sentence with multiple errors through normal seq2seq inference. We attempt to address these limitations by proposing a fluency boost learning and inference mechanism. Fluency boosting learning generates fluency-boost sentence pairs during training, enabling the error correction model to learn how to improve a sentence`s fluency from more instances, while fluency boosting inference allows the model to correct a sentence incrementally with multiple inference steps until the sentence`s fluency stops increasing. Experiments show our approaches improve the performance of seq2seq models for GEC, achieving state-of-the-art results on both CoNLL-2014 and JFLEG benchmark datasets."
}
Markdown (Informal)
[Fluency Boost Learning and Inference for Neural Grammatical Error Correction](https://preview.aclanthology.org/add-emnlp-2024-awards/P18-1097/) (Ge et al., ACL 2018)
ACL