Hierarchical Trivia Fact Extraction from Wikipedia Articles
Jingun Kwon, Hidetaka Kamigaito, Young-In Song, Manabu Okumura
Abstract
Recently, automatic trivia fact extraction has attracted much research interest. Modern search engines have begun to provide trivia facts as the information for entities because they can motivate more user engagement. In this paper, we propose a new unsupervised algorithm that automatically mines trivia facts for a given entity. Unlike previous studies, the proposed algorithm targets at a single Wikipedia article and leverages its hierarchical structure via top-down processing. Thus, the proposed algorithm offers two distinctive advantages: it does not incur high computation time, and it provides a domain-independent approach for extracting trivia facts. Experimental results demonstrate that the proposed algorithm is over 100 times faster than the existing method which considers Wikipedia categories. Human evaluation demonstrates that the proposed algorithm can mine better trivia facts regardless of the target entity domain and outperforms the existing methods.- Anthology ID:
- 2020.coling-main.424
- Volume:
- Proceedings of the 28th International Conference on Computational Linguistics
- Month:
- December
- Year:
- 2020
- Address:
- Barcelona, Spain (Online)
- Editors:
- Donia Scott, Nuria Bel, Chengqing Zong
- Venue:
- COLING
- SIG:
- Publisher:
- International Committee on Computational Linguistics
- Note:
- Pages:
- 4825–4834
- Language:
- URL:
- https://preview.aclanthology.org/ingest_wac_2008/2020.coling-main.424/
- DOI:
- 10.18653/v1/2020.coling-main.424
- Cite (ACL):
- Jingun Kwon, Hidetaka Kamigaito, Young-In Song, and Manabu Okumura. 2020. Hierarchical Trivia Fact Extraction from Wikipedia Articles. In Proceedings of the 28th International Conference on Computational Linguistics, pages 4825–4834, Barcelona, Spain (Online). International Committee on Computational Linguistics.
- Cite (Informal):
- Hierarchical Trivia Fact Extraction from Wikipedia Articles (Kwon et al., COLING 2020)
- PDF:
- https://preview.aclanthology.org/ingest_wac_2008/2020.coling-main.424.pdf