Jeonghwan Choi
2025
Learning to Verify Summary Facts with Fine-Grained LLM Feedback
Jihwan Oh
|
Jeonghwan Choi
|
Nicole Hee-Yoen Kim
|
Taewon Yun
|
Hwanjun Song
Proceedings of the 31st International Conference on Computational Linguistics
Training automatic summary fact verifiers often faces the challenge of a lack of human-labeled data. In this paper, we explore alternative way of leveraging Large Language Model (LLM) generated feedback to address the inherent limitation of using human-labeled data. We introduce FineSumFact, a large-scale dataset containing fine-grained factual feedback on summaries. We employ 10 distinct LLMs for diverse summary generation and Llama-3-70B-Instruct for feedback. We utilize this dataset to fine-tune the lightweight open-source model Llama-3-8B-Instruct, optimizing resource efficiency while maintaining high performance. Our experimental results reveal that the model trained on extensive LLM-generated datasets surpasses that trained on smaller human-annotated datasets when evaluated using human-generated test sets. Fine-tuning fact verification models with LLM feedback can be more effective and cost-efficient than using human feedback. The dataset is available at https://github.com/DISL-Lab/FineSumFact.
Word2Passage: Word-level Importance Re-weighting for Query Expansion
Jeonghwan Choi
|
Minjeong Ban
|
Minseok Kim
|
Hwanjun Song
Findings of the Association for Computational Linguistics: ACL 2025
Retrieval-augmented generation (RAG) enhances the quality of LLM generation by providing relevant chunks, but retrieving accurately from external knowledge remains challenging due to missing contextually important words in query. We present Word2Passage, a novel approach that improves retrieval accuracy by optimizing word importance in query expansion. Our method generates references at word, sentence, and passage levels for query expansion, then determines word importance by considering both their reference level origin and characteristics derived from query types and corpus analysis. Specifically, our method assigns distinct importance scores to words based on whether they originate from word, sentence, or passage-level references. Extensive experiments demonstrate that Word2Passage outperforms existing methods across various datasets and LLM configurations, effectively enhancing both retrieval accuracy and generation quality. The code is publicly available at https://github.com/DISL-Lab/Word2Passage
Search
Fix author
Co-authors
- Hwanjun Song 2
- Minjeong Ban 1
- Nicole Hee-Yoen Kim 1
- Minseok Kim 1
- Jihwan Oh 1
- show all...