Promote, Suppress, Iterate: How Language Models Answer One-to-Many Factual Queries

Tianyi Lorena Yan, Robin Jia


Abstract
To answer one-to-many factual queries (e.g., listing cities of a country), a language model (LM) must simultaneously recall knowledge and avoid repeating previous answers. How are these two subtasks implemented and integrated internally? Across multiple datasets, models, and prompt templates, we identify a promote-then-suppress mechanism: the model first recalls all answers, and then suppresses previously generated ones. Specifically, LMs use both the subject and previous answer tokens to perform knowledge recall, with attention propagating subject information and MLPs promoting the answers. Then, attention attends to and suppresses previous answer tokens, while MLPs amplify the suppression signal. Our mechanism is corroborated by extensive experimental evidence: in addition to using early decoding and causal tracing, we analyze how components use different tokens by introducing both Token Lens, which decodes aggregated attention updates from specified tokens, and a knockout method that analyzes changes in MLP outputs after removing attention to specified tokens. Overall, we provide new insights into how LMs’ internal components interact with different input tokens to support complex factual recall.
Anthology ID:
2025.emnlp-main.815
Volume:
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2025
Address:
Suzhou, China
Editors:
Christos Christodoulopoulos, Tanmoy Chakraborty, Carolyn Rose, Violet Peng
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
16122–16145
Language:
URL:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.815/
DOI:
Bibkey:
Cite (ACL):
Tianyi Lorena Yan and Robin Jia. 2025. Promote, Suppress, Iterate: How Language Models Answer One-to-Many Factual Queries. In Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing, pages 16122–16145, Suzhou, China. Association for Computational Linguistics.
Cite (Informal):
Promote, Suppress, Iterate: How Language Models Answer One-to-Many Factual Queries (Yan & Jia, EMNLP 2025)
Copy Citation:
PDF:
https://preview.aclanthology.org/ingest-emnlp/2025.emnlp-main.815.pdf
Checklist:
 2025.emnlp-main.815.checklist.pdf