@inproceedings{guan-etal-2020-refining,
title = "Refining Data for Text Generation",
author = "Guan, Wenyu and
Liu, Qianying and
Li, Tianyi and
Li, Sujian",
booktitle = "Proceedings of the 19th Chinese National Conference on Computational Linguistics",
month = oct,
year = "2020",
address = "Haikou, China",
publisher = "Chinese Information Processing Society of China",
url = "https://aclanthology.org/2020.ccl-1.82",
pages = "881--891",
abstract = "Recent work on data-to-text generation has made progress under the neural encoder-decoder architectures. However, the data input size is often enormous, while not all data records are important for text generation and inappropriate input may bring noise into the final output. To solve this problem, we propose a two-step approach which first selects and orders the important data records and then generates text from the noise-reduced data. Here we propose a learning to rank model to rank the importance of each record which is supervised by a relation extractor. With the noise-reduced data as input, we implement a text generator which sequentially models the input data records and emits a summary. Experiments on the ROTOWIRE dataset verifies the effectiveness of our proposed method in both performance and efficiency.",
language = "English",
}
<?xml version="1.0" encoding="UTF-8"?>
<modsCollection xmlns="http://www.loc.gov/mods/v3">
<mods ID="guan-etal-2020-refining">
<titleInfo>
<title>Refining Data for Text Generation</title>
</titleInfo>
<name type="personal">
<namePart type="given">Wenyu</namePart>
<namePart type="family">Guan</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Qianying</namePart>
<namePart type="family">Liu</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Tianyi</namePart>
<namePart type="family">Li</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<name type="personal">
<namePart type="given">Sujian</namePart>
<namePart type="family">Li</namePart>
<role>
<roleTerm authority="marcrelator" type="text">author</roleTerm>
</role>
</name>
<originInfo>
<dateIssued>2020-oct</dateIssued>
</originInfo>
<typeOfResource>text</typeOfResource>
<language>
<languageTerm type="text">English</languageTerm>
<languageTerm type="code" authority="iso639-2b">eng</languageTerm>
</language>
<relatedItem type="host">
<titleInfo>
<title>Proceedings of the 19th Chinese National Conference on Computational Linguistics</title>
</titleInfo>
<originInfo>
<publisher>Chinese Information Processing Society of China</publisher>
<place>
<placeTerm type="text">Haikou, China</placeTerm>
</place>
</originInfo>
<genre authority="marcgt">conference publication</genre>
</relatedItem>
<abstract>Recent work on data-to-text generation has made progress under the neural encoder-decoder architectures. However, the data input size is often enormous, while not all data records are important for text generation and inappropriate input may bring noise into the final output. To solve this problem, we propose a two-step approach which first selects and orders the important data records and then generates text from the noise-reduced data. Here we propose a learning to rank model to rank the importance of each record which is supervised by a relation extractor. With the noise-reduced data as input, we implement a text generator which sequentially models the input data records and emits a summary. Experiments on the ROTOWIRE dataset verifies the effectiveness of our proposed method in both performance and efficiency.</abstract>
<identifier type="citekey">guan-etal-2020-refining</identifier>
<location>
<url>https://aclanthology.org/2020.ccl-1.82</url>
</location>
<part>
<date>2020-oct</date>
<extent unit="page">
<start>881</start>
<end>891</end>
</extent>
</part>
</mods>
</modsCollection>
%0 Conference Proceedings
%T Refining Data for Text Generation
%A Guan, Wenyu
%A Liu, Qianying
%A Li, Tianyi
%A Li, Sujian
%S Proceedings of the 19th Chinese National Conference on Computational Linguistics
%D 2020
%8 oct
%I Chinese Information Processing Society of China
%C Haikou, China
%G English
%F guan-etal-2020-refining
%X Recent work on data-to-text generation has made progress under the neural encoder-decoder architectures. However, the data input size is often enormous, while not all data records are important for text generation and inappropriate input may bring noise into the final output. To solve this problem, we propose a two-step approach which first selects and orders the important data records and then generates text from the noise-reduced data. Here we propose a learning to rank model to rank the importance of each record which is supervised by a relation extractor. With the noise-reduced data as input, we implement a text generator which sequentially models the input data records and emits a summary. Experiments on the ROTOWIRE dataset verifies the effectiveness of our proposed method in both performance and efficiency.
%U https://aclanthology.org/2020.ccl-1.82
%P 881-891
Markdown (Informal)
[Refining Data for Text Generation](https://aclanthology.org/2020.ccl-1.82) (Guan et al., CCL 2020)
ACL
- Wenyu Guan, Qianying Liu, Tianyi Li, and Sujian Li. 2020. Refining Data for Text Generation. In Proceedings of the 19th Chinese National Conference on Computational Linguistics, pages 881–891, Haikou, China. Chinese Information Processing Society of China.