Ivan Lysenko
2025
ATGen: A Framework for Active Text Generation
Akim Tsvigun
|
Daniil Vasilev
|
Ivan Tsvigun
|
Ivan Lysenko
|
Talgat Bektleuov
|
Aleksandr Medvedev
|
Uliana Vinogradova
|
Nikita Severin
|
Mikhail Mozikov
|
Andrey Savchenko
|
Ilya Makarov
|
Grigorev Rostislav
|
Ramil Kuleev
|
Fedor Zhdanov
|
Artem Shelmanov
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 3: System Demonstrations)
Active learning (AL) has demonstrated remarkable potential in reducing the annotation effort required for training machine learning models. However, despite the surging popularity of natural language generation (NLG) tasks in recent years, the application of AL to NLG has been limited. In this paper, we introduce Active Text Generation (ATGen) - a comprehensive framework that bridges AL with text generation tasks, enabling the application of state-of-the-art AL strategies to NLG. Our framework simplifies AL-empowered annotation in NLG tasks using both human annotators and automatic annotation agents based on large language models (LLMs). The framework supports LLMs deployed as a service, such as ChatGPT and Claude, or operated on-premises. Furthermore, ATGen provides a unified platform for smooth implementation and benchmarking of novel AL strategies tailored to NLG tasks. Finally, we present experimental results across multiple text generation tasks where we compare the performance of state-of-the-art AL strategies in various settings. We demonstrate that ATGen can reduce both the effort of human annotators and costs for API calls to automatic annotation agents based on LLMs.
2022
Active Learning for Abstractive Text Summarization
Akim Tsvigun
|
Ivan Lysenko
|
Danila Sedashov
|
Ivan Lazichny
|
Eldar Damirov
|
Vladimir Karlov
|
Artemy Belousov
|
Leonid Sanochkin
|
Maxim Panov
|
Alexander Panchenko
|
Mikhail Burtsev
|
Artem Shelmanov
Findings of the Association for Computational Linguistics: EMNLP 2022
Construction of human-curated annotated datasets for abstractive text summarization (ATS) is very time-consuming and expensive because creating each instance requires a human annotator to read a long document and compose a shorter summary that would preserve the key information relayed by the original document. Active Learning (AL) is a technique developed to reduce the amount of annotation required to achieve a certain level of machine learning model performance. In information extraction and text classification, AL can reduce the amount of labor up to multiple times. Despite its potential for aiding expensive annotation, as far as we know, there were no effective AL query strategies for ATS. This stems from the fact that many AL strategies rely on uncertainty estimation, while as we show in our work, uncertain instances are usually noisy, and selecting them can degrade the model performance compared to passive annotation. We address this problem by proposing the first effective query strategy for AL in ATS based on diversity principles. We show that given a certain annotation budget, using our strategy in AL annotation helps to improve the model performance in terms of ROUGE and consistency scores. Additionally, we analyze the effect of self-learning and show that it can additionally increase the performance of the model.
Search
Fix author
Co-authors
- Artem Shelmanov 2
- Akim Tsvigun 2
- Talgat Bektleuov 1
- Artemy Belousov 1
- Mikhail Burtsev 1
- show all...
- Eldar Damirov 1
- Vladimir Karlov 1
- Ramil Kuleev 1
- Ivan Lazichny 1
- Ilya Makarov 1
- Aleksandr Medvedev 1
- Mikhail Mozikov 1
- Alexander Panchenko 1
- Maxim Panov 1
- Grigorev Rostislav 1
- Leonid Sanochkin 1
- Andrey Savchenko 1
- Danila Sedashov 1
- Nikita Severin 1
- Ivan Tsvigun 1
- Daniil Vasilev 1
- Uliana Vinogradova 1
- Fedor Zhdanov 1