Muhammad Farid Adilazuarda
Also published as: Farid Adilazuarda
2025
NusaAksara: A Multimodal and Multilingual Benchmark for Preserving Indonesian Indigenous Scripts
Muhammad Farid Adilazuarda | Musa Izzanardi Wijanarko | Lucky Susanto | Khumaisa Nur’aini | Derry Tanti Wijaya | Alham Fikri Aji
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Muhammad Farid Adilazuarda | Musa Izzanardi Wijanarko | Lucky Susanto | Khumaisa Nur’aini | Derry Tanti Wijaya | Alham Fikri Aji
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
Indonesia is rich in languages and scripts. However, most NLP progress has been made using romanized text. In this paper, we present NusaAksara, a novel public benchmark for Indonesian languages that includes their original scripts. Our benchmark covers both text and image modalities and encompasses diverse tasks such as image segmentation, OCR, transliteration, translation, and language identification. Our data is constructed by human experts through rigorous steps. NusaAksara covers 8 scripts across 7 languages, including low-resource languages not commonly seen in NLP benchmarks. Although unsupported by Unicode, the Lampung script is included in this dataset. We benchmark our data across several models, from LLMs and VLMs such as GPT-4o, Llama 3.2, and Aya 23 to task-specific systems such as PP-OCR and LangID, and show that most NLP technologies cannot handle Indonesia’s local scripts, with many achieving near-zero performance.
From Surveys to Narratives: Rethinking Cultural Value Adaptation in LLMs
Farid Adilazuarda | Chen Cecilia Liu | Iryna Gurevych | Alham Fikri Aji
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Farid Adilazuarda | Chen Cecilia Liu | Iryna Gurevych | Alham Fikri Aji
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Adapting cultural values in Large Language Models (LLMs) presents significant challenges, particularly due to biases and data limitations. Previous work aligns LLMs with different cultures using survey data, primarily from the World Values Survey (WVS). However, it remains unclear whether this approach effectively captures cultural nuances or produces distinct cultural representations for tasks like offensiveness classification. In this paper, we systematically investigate WVS-based training for cultural value adaptation and find that relying solely on survey data can homogenize cultural norms and interfere with factual knowledge. To address these issues, we propose augmenting WVS with encyclopedic and scenario-based cultural narratives from Wikipedia and NormAd. Our experiments across multiple cultures show that this approach captures more enhances differentiated cultural values and improves downstream classification performances.
MLKV: Multi-Layer Key-Value Heads for Memory Efficient Transformer Decoding
Zayd Muhammad Kawakibi Zuhri | Muhammad Farid Adilazuarda | Ayu Purwarianti | Alham Fikri Aji
Findings of the Association for Computational Linguistics: NAACL 2025
Zayd Muhammad Kawakibi Zuhri | Muhammad Farid Adilazuarda | Ayu Purwarianti | Alham Fikri Aji
Findings of the Association for Computational Linguistics: NAACL 2025
Auto-regressive inference of transformers benefit greatly from Key-Value (KV) caching, but can lead to major memory bottlenecks as model size, batch size, and sequence length grow at scale. We introduce Multi-Layer Key-Value (MLKV) sharing, a novel approach extending KV sharing across transformer layers to reduce memory usage beyond what was possible with Multi-Query Attention (MQA) and Grouped-Query Attention (GQA). Evaluations on various NLP benchmarks and inference metrics using uptrained Pythia-160M variants demonstrate that MLKV significantly reduces memory usage with minimal performance loss, reducing KV cache size down to a factor of 6x compared to MQA. These results highlight MLKV’s potential for efficient deployment of transformer models at scale.
WorldCuisines: A Massive-Scale Benchmark for Multilingual and Multicultural Visual Question Answering on Global Cuisines
Genta Indra Winata | Frederikus Hudi | Patrick Amadeus Irawan | David Anugraha | Rifki Afina Putri | Wang Yutong | Adam Nohejl | Ubaidillah Ariq Prathama | Nedjma Ousidhoum | Afifa Amriani | Anar Rzayev | Anirban Das | Ashmari Pramodya | Aulia Adila | Bryan Wilie | Candy Olivia Mawalim | Cheng Ching Lam | Daud Abolade | Emmanuele Chersoni | Enrico Santus | Fariz Ikhwantri | Garry Kuwanto | Hanyang Zhao | Haryo Akbarianto Wibowo | Holy Lovenia | Jan Christian Blaise Cruz | Jan Wira Gotama Putra | Junho Myung | Lucky Susanto | Maria Angelica Riera Machin | Marina Zhukova | Michael Anugraha | Muhammad Farid Adilazuarda | Natasha Christabelle Santosa | Peerat Limkonchotiwat | Raj Dabre | Rio Alexander Audino | Samuel Cahyawijaya | Shi-Xiong Zhang | Stephanie Yulia Salim | Yi Zhou | Yinxuan Gui | David Ifeoluwa Adelani | En-Shiun Annie Lee | Shogo Okada | Ayu Purwarianti | Alham Fikri Aji | Taro Watanabe | Derry Tanti Wijaya | Alice Oh | Chong-Wah Ngo
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Genta Indra Winata | Frederikus Hudi | Patrick Amadeus Irawan | David Anugraha | Rifki Afina Putri | Wang Yutong | Adam Nohejl | Ubaidillah Ariq Prathama | Nedjma Ousidhoum | Afifa Amriani | Anar Rzayev | Anirban Das | Ashmari Pramodya | Aulia Adila | Bryan Wilie | Candy Olivia Mawalim | Cheng Ching Lam | Daud Abolade | Emmanuele Chersoni | Enrico Santus | Fariz Ikhwantri | Garry Kuwanto | Hanyang Zhao | Haryo Akbarianto Wibowo | Holy Lovenia | Jan Christian Blaise Cruz | Jan Wira Gotama Putra | Junho Myung | Lucky Susanto | Maria Angelica Riera Machin | Marina Zhukova | Michael Anugraha | Muhammad Farid Adilazuarda | Natasha Christabelle Santosa | Peerat Limkonchotiwat | Raj Dabre | Rio Alexander Audino | Samuel Cahyawijaya | Shi-Xiong Zhang | Stephanie Yulia Salim | Yi Zhou | Yinxuan Gui | David Ifeoluwa Adelani | En-Shiun Annie Lee | Shogo Okada | Ayu Purwarianti | Alham Fikri Aji | Taro Watanabe | Derry Tanti Wijaya | Alice Oh | Chong-Wah Ngo
Proceedings of the 2025 Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 1: Long Papers)
Vision Language Models (VLMs) often struggle with culture-specific knowledge, particularly in languages other than English and in underrepresented cultural contexts. To evaluate their understanding of such knowledge, we introduce WorldCuisines, a massive-scale benchmark for multilingual and multicultural, visually grounded language understanding. This benchmark includes a visual question answering (VQA) dataset with text-image pairs across 30 languages and dialects, spanning 9 language families and featuring over 1 million data points, making it the largest multicultural VQA benchmark to date. It includes tasks for identifying dish names and their origins. We provide evaluation datasets in two sizes (12k and 60k instances) alongside a training dataset (1 million instances). Our findings show that while VLMs perform better with correct location context, they struggle with adversarial contexts and predicting specific regional cuisines and languages. To support future research, we release a knowledge base with annotated food entries and images along with the VQA data.
2024
SEACrowd: A Multilingual Multimodal Data Hub and Benchmark Suite for Southeast Asian Languages
Holy Lovenia | Rahmad Mahendra | Salsabil Maulana Akbar | Lester James V. Miranda | Jennifer Santoso | Elyanah Aco | Akhdan Fadhilah | Jonibek Mansurov | Joseph Marvin Imperial | Onno P. Kampman | Joel Ruben Antony Moniz | Muhammad Ravi Shulthan Habibi | Frederikus Hudi | Railey Montalan | Ryan Ignatius | Joanito Agili Lopo | William Nixon | Börje F. Karlsson | James Jaya | Ryandito Diandaru | Yuze Gao | Patrick Amadeus | Bin Wang | Jan Christian Blaise Cruz | Chenxi Whitehouse | Ivan Halim Parmonangan | Maria Khelli | Wenyu Zhang | Lucky Susanto | Reynard Adha Ryanda | Sonny Lazuardi Hermawan | Dan John Velasco | Muhammad Dehan Al Kautsar | Willy Fitra Hendria | Yasmin Moslem | Noah Flynn | Muhammad Farid Adilazuarda | Haochen Li | Johanes Lee | R. Damanhuri | Shuo Sun | Muhammad Reza Qorib | Amirbek Djanibekov | Wei Qi Leong | Quyet V. Do | Niklas Muennighoff | Tanrada Pansuwan | Ilham Firdausi Putra | Yan Xu | Tai Ngee Chia | Ayu Purwarianti | Sebastian Ruder | William Tjhi | Peerat Limkonchotiwat | Alham Fikri Aji | Sedrick Keh | Genta Indra Winata | Ruochen Zhang | Fajri Koto | Zheng-Xin Yong | Samuel Cahyawijaya
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Holy Lovenia | Rahmad Mahendra | Salsabil Maulana Akbar | Lester James V. Miranda | Jennifer Santoso | Elyanah Aco | Akhdan Fadhilah | Jonibek Mansurov | Joseph Marvin Imperial | Onno P. Kampman | Joel Ruben Antony Moniz | Muhammad Ravi Shulthan Habibi | Frederikus Hudi | Railey Montalan | Ryan Ignatius | Joanito Agili Lopo | William Nixon | Börje F. Karlsson | James Jaya | Ryandito Diandaru | Yuze Gao | Patrick Amadeus | Bin Wang | Jan Christian Blaise Cruz | Chenxi Whitehouse | Ivan Halim Parmonangan | Maria Khelli | Wenyu Zhang | Lucky Susanto | Reynard Adha Ryanda | Sonny Lazuardi Hermawan | Dan John Velasco | Muhammad Dehan Al Kautsar | Willy Fitra Hendria | Yasmin Moslem | Noah Flynn | Muhammad Farid Adilazuarda | Haochen Li | Johanes Lee | R. Damanhuri | Shuo Sun | Muhammad Reza Qorib | Amirbek Djanibekov | Wei Qi Leong | Quyet V. Do | Niklas Muennighoff | Tanrada Pansuwan | Ilham Firdausi Putra | Yan Xu | Tai Ngee Chia | Ayu Purwarianti | Sebastian Ruder | William Tjhi | Peerat Limkonchotiwat | Alham Fikri Aji | Sedrick Keh | Genta Indra Winata | Ruochen Zhang | Fajri Koto | Zheng-Xin Yong | Samuel Cahyawijaya
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Southeast Asia (SEA) is a region rich in linguistic diversity and cultural variety, with over 1,300 indigenous languages and a population of 671 million people. However, prevailing AI models suffer from a significant lack of representation of texts, images, and audio datasets from SEA, compromising the quality of AI models for SEA languages. Evaluating models for SEA languages is challenging due to the scarcity of high-quality datasets, compounded by the dominance of English training data, raising concerns about potential cultural misrepresentation. To address these challenges, through a collaborative movement, we introduce SEACrowd, a comprehensive resource center that fills the resource gap by providing standardized corpora in nearly 1,000 SEA languages across three modalities. Through our SEACrowd benchmarks, we assess the quality of AI models on 36 indigenous languages across 13 tasks, offering valuable insights into the current AI landscape in SEA. Furthermore, we propose strategies to facilitate greater AI advancements, maximizing potential utility and resource equity for the future of AI in Southeast Asia.
Towards Measuring and Modeling “Culture” in LLMs: A Survey
Muhammad Farid Adilazuarda | Sagnik Mukherjee | Pradhyumna Lavania | Siddhant Shivdutt Singh | Alham Fikri Aji | Jacki O’Neill | Ashutosh Modi | Monojit Choudhury
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Muhammad Farid Adilazuarda | Sagnik Mukherjee | Pradhyumna Lavania | Siddhant Shivdutt Singh | Alham Fikri Aji | Jacki O’Neill | Ashutosh Modi | Monojit Choudhury
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
We present a survey of more than 90 recent papers that aim to study cultural representation and inclusion in large language models (LLMs). We observe that none of the studies explicitly define “culture, which is a complex, multifaceted concept; instead, they probe the models on some specially designed datasets which represent certain aspects of “culture”. We call these aspects the proxies of culture, and organize them across two dimensions of demographic and semantic proxies. We also categorize the probing methods employed. Our analysis indicates that only certain aspects of “culture,” such as values and objectives, have been studied, leaving several other interesting and important facets, especially the multitude of semantic domains (Thompson et al., 2020) and aboutness (Hershcovich et al., 2022), unexplored. Two other crucial gaps are the lack of robustness of probing techniques and situated studies on the impact of cultural mis- and under-representation in LLM-based applications.
Cultural Conditioning or Placebo? On the Effectiveness of Socio-Demographic Prompting
Sagnik Mukherjee | Muhammad Farid Adilazuarda | Sunayana Sitaram | Kalika Bali | Alham Fikri Aji | Monojit Choudhury
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Sagnik Mukherjee | Muhammad Farid Adilazuarda | Sunayana Sitaram | Kalika Bali | Alham Fikri Aji | Monojit Choudhury
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Socio-demographic prompting is a commonly employed approach to study cultural biases in LLMs as well as for aligning models to certain cultures. In this paper, we systematically probe four LLMs (Llama 3, Mistral v0.2, GPT-3.5 Turbo and GPT4) with prompts that are conditioned on culturally sensitive and non-sensitive cues, on datasets that are supposed to be culturally sensitive (EtiCor and CALI) or neutral (MMLU and ETHICS). We observe that all models except GPT4 show significant variations in their responses on both kinds of datasets for both kinds of prompts, casting doubt on the robustness of the culturally-conditioned prompting as a method for eliciting cultural bias in models that are not sufficiently stable with respect to arbitrary prompting cues. Further, we also show that some of the supposedly culturally neutral datasets have a non-trivial fraction of culturally sensitive questions/tasks.
LinguAlchemy: Fusing Typological and Geographical Elements for Unseen Language Generalization
Muhammad Farid Adilazuarda | Samuel Cahyawijaya | Genta Indra Winata | Ayu Purwarianti | Alham Fikri Aji
Findings of the Association for Computational Linguistics: EMNLP 2024
Muhammad Farid Adilazuarda | Samuel Cahyawijaya | Genta Indra Winata | Ayu Purwarianti | Alham Fikri Aji
Findings of the Association for Computational Linguistics: EMNLP 2024
Pretrained language models (PLMs) have shown remarkable generalization toward multiple tasks and languages. Nonetheless, the generalization of PLMs towards unseen languages is poor, resulting in significantly worse language performance, or even generating nonsensical responses that are comparable to a random baseline. This limitation has been a longstanding problem of PLMs raising the problem of diversity and equal access to language modeling technology. In this work, we solve this limitation by introducing LinguAlchemy, a regularization technique that incorporates various aspects of languages covering typological, geographical, and phylogenetic constraining the resulting representation of PLMs to better characterize the corresponding linguistics constraints. LinguAlchemy significantly improves the accuracy performance of mBERT and XLM-R on unseen languages by ~18% and ~2%, respectively compared to fully finetuned models and displaying a high degree of unseen language generalization. We further introduce AlchemyScale and AlchemyTune, extension of LinguAlchemy which adjusts the linguistic regularization weights automatically, alleviating the need for hyperparameter search. LinguAlchemy enables better cross-lingual generalization to unseen languages which is vital for better inclusivity and accessibility of PLMs.
2023
Findings of the 1st Shared Task on Multi-lingual Multi-task Information Retrieval at MRL 2023
Francesco Tinner | David Ifeoluwa Adelani | Chris Emezue | Mammad Hajili | Omer Goldman | Muhammad Farid Adilazuarda | Muhammad Dehan Al Kautsar | Aziza Mirsaidova | Müge Kural | Dylan Massey | Chiamaka Chukwuneke | Chinedu Mbonu | Damilola Oluwaseun Oloyede | Kayode Olaleye | Jonathan Atala | Benjamin A. Ajibade | Saksham Bassi | Rahul Aralikatte | Najoung Kim | Duygu Ataman
Proceedings of the 3rd Workshop on Multi-lingual Representation Learning (MRL)
Francesco Tinner | David Ifeoluwa Adelani | Chris Emezue | Mammad Hajili | Omer Goldman | Muhammad Farid Adilazuarda | Muhammad Dehan Al Kautsar | Aziza Mirsaidova | Müge Kural | Dylan Massey | Chiamaka Chukwuneke | Chinedu Mbonu | Damilola Oluwaseun Oloyede | Kayode Olaleye | Jonathan Atala | Benjamin A. Ajibade | Saksham Bassi | Rahul Aralikatte | Najoung Kim | Duygu Ataman
Proceedings of the 3rd Workshop on Multi-lingual Representation Learning (MRL)
2022
Search
Fix author
Co-authors
- Alham Fikri Aji 8
- Ayu Purwarianti 5
- Samuel Cahyawijaya 4
- Genta Indra Winata 4
- Lucky Susanto 3
- David Ifeoluwa Adelani 2
- Muhammad Dehan Al Kautsar 2
- Monojit Choudhury 2
- Jan Christian Blaise Cruz 2
- Frederikus Hudi 2
- Peerat Limkonchotiwat 2
- Holy Lovenia 2
- Sagnik Mukherjee 2
- Derry Tanti Wijaya 2
- Daud Abolade 1
- Elyanah Aco 1
- Aulia Adila 1
- Benjamin A. Ajibade 1
- Salsabil Maulana Akbar 1
- Patrick Amadeus 1
- Afifa Amriani 1
- David Anugraha 1
- Michael Anugraha 1
- Rahul Aralikatte 1
- Jonathan Atala 1
- Duygu Ataman 1
- Rio Alexander Audino 1
- Kalika Bali 1
- Saksham Bassi 1
- Emmanuele Chersoni 1
- Tai Ngee Chia 1
- Chiamaka Chukwuneke 1
- Raj Dabre 1
- R. Damanhuri 1
- Anirban Das 1
- Ryandito Diandaru 1
- Amirbek Djanibekov 1
- Quyet V. Do 1
- Chris Chinenye Emezue 1
- Akhdan Fadhilah 1
- Noah Flynn 1
- Pascale Fung 1
- Yuze Gao 1
- Omer Goldman 1
- Yinxuan Gui 1
- Iryna Gurevych 1
- Muhammad Ravi Shulthan Habibi 1
- Mammad Hajili 1
- Willy Fitra Hendria 1
- Sonny Lazuardi Hermawan 1
- Ryan Ignatius 1
- Fariz Ikhwantri 1
- Joseph Marvin Imperial 1
- Patrick Amadeus Irawan 1
- James Jaya 1
- Onno P. Kampman 1
- Börje F. Karlsson 1
- Sedrick Keh 1
- Maria Khelli 1
- Najoung Kim 1
- Fajri Koto 1
- Müge Kural 1
- Garry Kuwanto 1
- Cheng Ching Lam 1
- Pradhyumna Lavania 1
- Johanes Lee 1
- En-Shiun Annie Lee 1
- Wei Qi Leong 1
- Haochen Li 1
- Chen Cecilia Liu 1
- Joanito Agili Lopo 1
- Rahmad Mahendra 1
- Jonibek Mansurov 1
- Dylan Massey 1
- Candy Olivia Mawalim 1
- Chinedu Mbonu 1
- Lester James Validad Miranda 1
- Aziza Mirsaidova 1
- Ashutosh Modi 1
- Joel Ruben Antony Moniz 1
- Jann Railey Montalan 1
- Yasmin Moslem 1
- Niklas Muennighoff 1
- Junho Myung 1
- Chong-Wah Ngo 1
- William Nixon 1
- Adam Nohejl 1
- Khumaisa Nur’aini 1
- Alice Oh 1
- Shogo Okada 1
- Kayode Olaleye 1
- Damilola Oluwaseun Oloyede 1
- Nedjma Ousidhoum 1
- Jacki O’Neill 1
- Tanrada Pansuwan 1
- Ivan Halim Parmonangan 1
- Ashmari Pramodya 1
- Ubaidillah Ariq Prathama 1
- Ilham Firdausi Putra 1
- Jan Wira Gotama Putra 1
- Rifki Afina Putri 1
- Muhammad Reza Qorib 1
- Maria Angelica Riera Machin 1
- Sebastian Ruder 1
- Reynard Adha Ryanda 1
- Anar Rzayev 1
- Stephanie Yulia Salim 1
- Natasha Christabelle Santosa 1
- Jennifer Santoso 1
- Enrico Santus 1
- Siddhant Shivdutt Singh 1
- Sunayana Sitaram 1
- Shuo Sun 1
- Francesco Tinner 1
- William Tjhi 1
- Dan John Velasco 1
- Bin Wang 1
- Taro Watanabe 1
- Chenxi Whitehouse 1
- Haryo Akbarianto Wibowo 1
- Musa Izzanardi Wijanarko 1
- Bryan Wilie 1
- Yan Xu 1
- Zheng Xin Yong 1
- Wang Yutong 1
- Wenyu Zhang 1
- Ruochen Zhang 1
- Shi-Xiong Zhang 1
- Hanyang Zhao 1
- Yi Zhou 1
- Marina Zhukova 1
- Zayd Muhammad Kawakibi Zuhri 1