@inproceedings{zhang-etal-2023-g,
title = "{G}-{SPEED}: General {SP}arse Efficient Editing {M}o{D}el",
author = "Zhang, Haoke and
Wang, Yue and
Li, Juntao and
Zhou, Xiabing and
Zhang, Min",
editor = "Bouamor, Houda and
Pino, Juan and
Bali, Kalika",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2023",
month = dec,
year = "2023",
address = "Singapore",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-emnlp.142/",
doi = "10.18653/v1/2023.findings-emnlp.142",
pages = "2160--2175",
abstract = "Large Language Models (LLMs) have demonstrated incredible capabilities in understanding, generating, and manipulating languages. Through human-model interactions, LLMs can automatically understand human-issued instructions and output the expected contents, which can significantly increase working efficiency. In various types of real-world demands, editing-oriented tasks account for a considerable proportion, which involves an interactive process that entails the continuous refinement of existing texts to meet specific criteria. Due to the need for multi-round human-model interaction and the generation of complicated editing tasks, there is an emergent need for efficient general editing models. In this paper, we propose \textbf{G}eneral \textbf{SP}arse \textbf{E}fficient \textbf{E}diting Mo\textbf{D}el (\textbf{G-SPEED}), which can fulfill diverse editing requirements through a single model while maintaining low computational costs. Specifically, we first propose a novel unsupervised text editing data clustering algorithm to deal with the data scarcity problem. Subsequently, we introduce a sparse editing model architecture to mitigate the inherently limited learning capabilities of small language models. The experimental outcomes indicate that G-SPEED, with its 508M parameters, can surpass LLMs equipped with 175B parameters. Our code and model checkpoints are available at \url{https://github.com/Banner-Z/G-SPEED}."
}
Markdown (Informal)
[G-SPEED: General SParse Efficient Editing MoDel](https://preview.aclanthology.org/jlcl-multiple-ingestion/2023.findings-emnlp.142/) (Zhang et al., Findings 2023)
ACL
- Haoke Zhang, Yue Wang, Juntao Li, Xiabing Zhou, and Min Zhang. 2023. G-SPEED: General SParse Efficient Editing MoDel. In Findings of the Association for Computational Linguistics: EMNLP 2023, pages 2160–2175, Singapore. Association for Computational Linguistics.