FAC2E: Better Understanding Large Language Model Capabilities by Dissociating Language and Cognition

Xiaoqiang Wang, Lingfei Wu, Tengfei Ma, Bang Liu


Abstract
Large language models (LLMs) are primarily evaluated by overall performance on various text understanding and generation tasks. However, such a paradigm fails to comprehensively differentiate the fine-grained language and cognitive skills, rendering the lack of sufficient interpretation to LLMs’ capabilities. In this paper, we present FAC2E, a framework for Fine-grAined and Cognition-grounded LLMs’ Capability Evaluation. Specifically, we formulate LLMs’ evaluation in a multi-dimensional and explainable manner by dissociating the language-related capabilities and the cognition-related ones. Besides, through extracting the intermediate reasoning from LLMs, we further break down the process of applying a specific capability into three sub-steps: recalling relevant knowledge, utilizing knowledge, and solving problems. Finally, FAC2E evaluates each sub-step of each fine-grained capability, providing a two-faceted diagnosis for LLMs. Utilizing FAC2E, we identify a common shortfall in knowledge utilization among models and propose a straightforward, knowledge-enhanced method to mitigate this issue. Our results not only showcase promising performance enhancements but also highlight a direction for future LLM advancements.
Anthology ID:
2024.emnlp-main.734
Volume:
Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing
Month:
November
Year:
2024
Address:
Miami, Florida, USA
Editors:
Yaser Al-Onaizan, Mohit Bansal, Yun-Nung Chen
Venue:
EMNLP
SIG:
Publisher:
Association for Computational Linguistics
Note:
Pages:
13228–13243
Language:
URL:
https://aclanthology.org/2024.emnlp-main.734
DOI:
10.18653/v1/2024.emnlp-main.734
Bibkey:
Cite (ACL):
Xiaoqiang Wang, Lingfei Wu, Tengfei Ma, and Bang Liu. 2024. FAC2E: Better Understanding Large Language Model Capabilities by Dissociating Language and Cognition. In Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, pages 13228–13243, Miami, Florida, USA. Association for Computational Linguistics.
Cite (Informal):
FAC2E: Better Understanding Large Language Model Capabilities by Dissociating Language and Cognition (Wang et al., EMNLP 2024)
Copy Citation:
PDF:
https://preview.aclanthology.org/dois-2013-emnlp/2024.emnlp-main.734.pdf