CAVA: A Tool for Cultural Alignment Visualization & Analysis
Nevan Giuliani, Cheng Charles Ma, Prakruthi Pradeep, Daphne Ippolito
Abstract
It is well-known that language models are biased; they have patchy knowledge of countries and cultures that are poorly represented in their training data. We introduce CAVA, a visualization tool for identifying and analyzing country-specific biases in language models.Our tool allows users to identify whether a language model successful captures the perspectives of people of different nationalities. The tool supports analysis of both longform and multiple-choice models responses and comparisons between models.Our open-source code easily allows users to upload any country-based language model generations they wish to analyze.To showcase CAVA’s efficacy, we present a case study analyzing how several popular language models answer survey questions from the World Values Survey.- Anthology ID:
- 2024.emnlp-demo.16
- Volume:
- Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: System Demonstrations
- Month:
- November
- Year:
- 2024
- Address:
- Miami, Florida, USA
- Editors:
- Delia Irazu Hernandez Farias, Tom Hope, Manling Li
- Venue:
- EMNLP
- SIG:
- Publisher:
- Association for Computational Linguistics
- Note:
- Pages:
- 153–161
- Language:
- URL:
- https://preview.aclanthology.org/moar-dois/2024.emnlp-demo.16/
- DOI:
- 10.18653/v1/2024.emnlp-demo.16
- Cite (ACL):
- Nevan Giuliani, Cheng Charles Ma, Prakruthi Pradeep, and Daphne Ippolito. 2024. CAVA: A Tool for Cultural Alignment Visualization & Analysis. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing: System Demonstrations, pages 153–161, Miami, Florida, USA. Association for Computational Linguistics.
- Cite (Informal):
- CAVA: A Tool for Cultural Alignment Visualization & Analysis (Giuliani et al., EMNLP 2024)
- PDF:
- https://preview.aclanthology.org/moar-dois/2024.emnlp-demo.16.pdf