@inproceedings{mezentsev-oseledets-2025-exploring,
    title = "Exploring the Hidden Capacity of {LLM}s for One-Step Text Generation",
    author = "Mezentsev, Gleb  and
      Oseledets, Ivan",
    editor = "Christodoulopoulos, Christos  and
      Chakraborty, Tanmoy  and
      Rose, Carolyn  and
      Peng, Violet",
    booktitle = "Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing",
    month = nov,
    year = "2025",
    address = "Suzhou, China",
    publisher = "Association for Computational Linguistics",
    url = "https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1165/",
    pages = "22891--22900",
    ISBN = "979-8-89176-332-6",
    abstract = "A recent study showed that large language models (LLMs) can reconstruct surprisingly long texts {---} up to thousands of tokens {---} via autoregressive generation from just one trained input embedding. In this work, we explore whether autoregressive decoding is essential for such reconstruction. We show that frozen LLMs can generate hundreds of accurate tokens in just one token-parallel forward pass, when provided with only two learned embeddings. This reveals a surprising and underexplored multi-token generation capability of autoregressive LLMs. We examine these embeddings and characterize the information they encode. We also empirically show that, although these representations are not unique for a given text, they form connected and local regions in embedding space {---} suggesting the potential to train a practical encoder. The existence of such representations hints that multi-token generation may be natively accessible in off-the-shelf LLMs via a learned input encoder, eliminating heavy retraining and helping to overcome the fundamental bottleneck of autoregressive decoding while reusing already-trained models."
}Markdown (Informal)
[Exploring the Hidden Capacity of LLMs for One-Step Text Generation](https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.1165/) (Mezentsev & Oseledets, EMNLP 2025)
ACL