@inproceedings{koshkin-etal-2024-transllama,
title = "{T}rans{LL}a{M}a: {LLM}-based Simultaneous Translation System",
author = "Koshkin, Roman and
Sudoh, Katsuhito and
Nakamura, Satoshi",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Findings of the Association for Computational Linguistics: EMNLP 2024",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/add-emnlp-2024-awards/2024.findings-emnlp.27/",
doi = "10.18653/v1/2024.findings-emnlp.27",
pages = "461--476",
abstract = "Decoder-only large language models (LLMs) have recently demonstrated impressive capabilities in text generation and reasoning. Nonetheless, they have limited applications in simultaneous machine translation (SiMT), currently dominated by encoder-decoder transformers. This study demonstrates that, after fine-tuning on a small dataset comprising causally aligned source and target sentence pairs, a pre-trained open-source LLM can control input segmentation directly by generating a special {\textquotedblleft}wait{\textquotedblright} token. This obviates the need for a separate policy and enables the LLM to perform English-German and English-Russian SiMT tasks with BLEU scores that are comparable to those of specific state-of-the-art baselines. We also evaluated closed-source models such as GPT-4, which displayed encouraging results in performing the SiMT task without prior training (zero-shot), indicating a promising avenue for enhancing future SiMT systems."
}
Markdown (Informal)
[TransLLaMa: LLM-based Simultaneous Translation System](https://preview.aclanthology.org/add-emnlp-2024-awards/2024.findings-emnlp.27/) (Koshkin et al., Findings 2024)
ACL
- Roman Koshkin, Katsuhito Sudoh, and Satoshi Nakamura. 2024. TransLLaMa: LLM-based Simultaneous Translation System. In Findings of the Association for Computational Linguistics: EMNLP 2024, pages 461–476, Miami, Florida, USA. Association for Computational Linguistics.