@inproceedings{peng-sogaard-2024-concept,
title = "Concept Space Alignment in Multilingual {LLM}s",
author = "Peng, Qiwei and
S{\o}gaard, Anders",
editor = "Al-Onaizan, Yaser and
Bansal, Mohit and
Chen, Yun-Nung",
booktitle = "Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing",
month = nov,
year = "2024",
address = "Miami, Florida, USA",
publisher = "Association for Computational Linguistics",
url = "https://preview.aclanthology.org/fix-sig-urls/2024.emnlp-main.315/",
doi = "10.18653/v1/2024.emnlp-main.315",
pages = "5511--5526",
abstract = "Multilingual large language models (LLMs) seem to generalize somewhat across languages. We hypothesize this is a result of implicit vector space alignment. Evaluating such alignment, we see that larger models exhibit very high-quality linear alignments between corresponding concepts in different languages. Our experiments show that multilingual LLMs suffer from two familiar weaknesses: generalization works best for languages with similar typology, and for abstract concepts. For some models, e.g., the Llama-2 family of models, prompt-based embeddings align better than word embeddings, but the projections are less linear {--} an observation that holds across almost all model families, indicating that some of the implicitly learned alignments are broken somewhat by prompt-based methods."
}
Markdown (Informal)
[Concept Space Alignment in Multilingual LLMs](https://preview.aclanthology.org/fix-sig-urls/2024.emnlp-main.315/) (Peng & Søgaard, EMNLP 2024)
ACL
- Qiwei Peng and Anders Søgaard. 2024. Concept Space Alignment in Multilingual LLMs. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 5511–5526, Miami, Florida, USA. Association for Computational Linguistics.