2025
pdf
bib
abs
Generating Long-form Story Using Dynamic Hierarchical Outlining with Memory-Enhancement
Qianyue Wang
|
Jinwu Hu
|
Zhengping Li
|
Yufeng Wang
|
Daiyuan Li
|
Yu Hu
|
Mingkui Tan
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Long-form story generation task aims to produce coherent and sufficiently lengthy text, essential for applications such as novel writingand interactive storytelling. However, existing methods, including LLMs, rely on rigid outlines or lack macro-level planning, making it difficult to achieve both contextual consistency and coherent plot development in long-form story generation. To address this issues, we propose Dynamic Hierarchical Outlining with Memory-Enhancement long-form story generation method, named DOME, to generate the long-form story with coherent content and plot. Specifically, the Dynamic Hierarchical Outline(DHO) mechanism incorporates the novel writing theory into outline planning and fuses the plan and writing stages together, improving the coherence of the plot by ensuring the plot completeness and adapting to the uncertainty during story generation. A Memory-Enhancement Module (MEM) based on temporal knowledge graphs is introduced to store and access the generated content, reducing contextual conflicts and improving story coherence. Finally, we propose a Temporal Conflict Analyzer leveraging temporal knowledge graphs to automatically evaluate the contextual consistency of long-form story. Experiments demonstrate that DOME significantly improves the fluency, coherence, and overall quality of generated long stories compared to state-of-the-art methods.
2024
pdf
bib
abs
UniTabNet: Bridging Vision and Language Models for Enhanced Table Structure Recognition
Zhenrong Zhang
|
Shuhang Liu
|
Pengfei Hu
|
Jiefeng Ma
|
Jun Du
|
Jianshu Zhang
|
Yu Hu
Findings of the Association for Computational Linguistics: EMNLP 2024
In the digital era, table structure recognition technology is a critical tool for processing and analyzing large volumes of tabular data. Previous methods primarily focus on visual aspects of table structure recovery but often fail to effectively comprehend the textual semantics within tables, particularly for descriptive textual cells. In this paper, we introduce UniTabNet, a novel framework for table structure parsing based on the image-to-text model. UniTabNet employs a “divide-and-conquer” strategy, utilizing an image-to-text model to decouple table cells and integrating both physical and logical decoders to reconstruct the complete table structure. We further enhance our framework with the Vision Guider, which directs the model’s focus towards pertinent areas, thereby boosting prediction accuracy. Additionally, we introduce the Language Guider to refine the model’s capability to understand textual semantics in table images. Evaluated on prominent table structure datasets such as PubTabNet, PubTables1M, WTW, and iFLYTAB, UniTabNet achieves a new state-of-the-art performance, demonstrating the efficacy of our approach. The code will also be made publicly available.
2016
pdf
bib
Exploring Semantic Representation in Brain Activity Using Word Embeddings
Yu-Ping Ruan
|
Zhen-Hua Ling
|
Yu Hu
Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing
pdf
bib
Intra-Topic Variability Normalization based on Linear Projection for Topic Classification
Quan Liu
|
Wu Guo
|
Zhen-Hua Ling
|
Hui Jiang
|
Yu Hu
Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies
2015
pdf
bib
Learning Semantic Word Embeddings based on Ordinal Knowledge Constraints
Quan Liu
|
Hui Jiang
|
Si Wei
|
Zhen-Hua Ling
|
Yu Hu
Proceedings of the 53rd Annual Meeting of the Association for Computational Linguistics and the 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)
2006
pdf
bib
Exploring variant definitions of pointer length in MDL
Aris Xanthos
|
Yu Hu
|
John Goldsmith
Proceedings of the Eighth Meeting of the ACL Special Interest Group on Computational Phonology and Morphology at HLT-NAACL 2006
2005
pdf
bib
Using Morphology and Syntax Together in Unsupervised Learning
Yu Hu
|
Irina Matveeva
|
John Goldsmith
|
Colin Sprague
Proceedings of the Workshop on Psychocomputational Models of Human Language Acquisition
pdf
bib
Refining the SED Heuristic for Morpheme Discovery: Another Look at Swahili
Yu Hu
|
Irina Matveeva
|
John Goldsmith
|
Colin Sprague
Proceedings of the Workshop on Psychocomputational Models of Human Language Acquisition