@inproceedings{yang-etal-2025-rethinking,
title = "Rethinking Long Context Generation from the Continual Learning Perspective",
author = "Yang, Zeyuan and
Xiong, Fangzhou and
Li, Peng and
Liu, Yang",
editor = "Rambow, Owen and
Wanner, Leo and
Apidianaki, Marianna and
Al-Khalifa, Hend and
Eugenio, Barbara Di and
Schockaert, Steven",
booktitle = "Proceedings of the 31st International Conference on Computational Linguistics",
month = jan,
year = "2025",
address = "Abu Dhabi, UAE",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2025.coling-main.131/",
pages = "1922--1933",
abstract = "Due to the limited context window, Large Language Models (LLMs) struggle with processing long contexts. Although fine-tuning can extend the context window, it incurs substantial computation costs. In contrast, recent tuning-free approaches reallocate the attention mechanism or incorporate temporary trainable parameters. In this work, by jointly modeling instance-level generation with a limited context window and learning over sequential data, we rethink the long context generation of LLMs from a continual learning perspective. In practice, we inspect existing representative approaches and analyze their synergy with continual learning strategies. Moreover, we integrate these strategies into current approaches to further boost LLMs' efficiency in processing long contexts. Comprehensive experiments and analysis confirm the feasibility of continual learning insights for improving long-context processing."
}
Markdown (Informal)
[Rethinking Long Context Generation from the Continual Learning Perspective](https://preview.aclanthology.org/fix-sig-urls/2025.coling-main.131/) (Yang et al., COLING 2025)
ACL