Danyang Liu
2023
Visual Storytelling with Question-Answer Plans
Danyang Liu
|
Mirella Lapata
|
Frank Keller
Findings of the Association for Computational Linguistics: EMNLP 2023
Visual storytelling aims to generate compelling narratives from image sequences. Existing models often focus on enhancing the representation of the image sequence, e.g., with external knowledge sources or advanced graph structures. Despite recent progress, the stories are often repetitive, illogical, and lacking in detail. To mitigate these issues, we present a novel framework which integrates visual representations with pretrained language models and planning. Our model translates the image sequence into a visual prefix, a sequence of continuous embeddings which language models can interpret. It also leverages a sequence of question-answer pairs as a blueprint plan for selecting salient visual concepts and determining how they should be assembled into a narrative. Automatic and human evaluation on the VIST benchmark demonstrates that blueprint-based models generate stories that are more coherent, interesting, and natural compared to competitive baselines and state-of-the-art systems.
2022
Automatically Discarding Straplines to Improve Data Quality for Abstractive News Summarization
Amr Keleg
|
Matthias Lindemann
|
Danyang Liu
|
Wanqiu Long
|
Bonnie L. Webber
Proceedings of NLP Power! The First Workshop on Efficient Benchmarking in NLP
Recent improvements in automatic news summarization fundamentally rely on large corpora of news articles and their summaries. These corpora are often constructed by scraping news websites, which results in including not only summaries but also other kinds of texts. Apart from more generic noise, we identify straplines as a form of text scraped from news websites that commonly turn out not to be summaries. The presence of these non-summaries threatens the validity of scraped corpora as benchmarks for news summarization. We have annotated extracts from two news sources that form part of the Newsroom corpus (Grusky et al., 2018), labeling those which were straplines, those which were summaries, and those which were both. We present a rule-based strapline detection method that achieves good performance on a manually annotated test set. Automatic evaluation indicates that removing straplines and noise from the training data of a news summarizer results in higher quality summaries, with improvements as high as 7 points ROUGE score.
2020
MIND: A Large-scale Dataset for News Recommendation
Fangzhao Wu
|
Ying Qiao
|
Jiun-Hung Chen
|
Chuhan Wu
|
Tao Qi
|
Jianxun Lian
|
Danyang Liu
|
Xing Xie
|
Jianfeng Gao
|
Winnie Wu
|
Ming Zhou
Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics
News recommendation is an important technique for personalized news service. Compared with product and movie recommendations which have been comprehensively studied, the research on news recommendation is much more limited, mainly due to the lack of a high-quality benchmark dataset. In this paper, we present a large-scale dataset named MIND for news recommendation. Constructed from the user click logs of Microsoft News, MIND contains 1 million users and more than 160k English news articles, each of which has rich textual content such as title, abstract and body. We demonstrate MIND a good testbed for news recommendation through a comparative study of several state-of-the-art news recommendation methods which are originally developed on different proprietary datasets. Our results show the performance of news recommendation highly relies on the quality of news content understanding and user interest modeling. Many natural language processing techniques such as effective text representation methods and pre-trained language models can effectively improve the performance of news recommendation. The MIND dataset will be available at https://msnews.github.io.
Search
Co-authors
- Mirella Lapata 1
- Frank Keller 1
- Fangzhao Wu 1
- Ying Qiao 1
- Jiun-Hung Chen 1
- show all...