Latent vs Explicit Knowledge Representation: How ChatGPT Answers Questions about Low-Frequency Entities

Arianna Graciotti, Valentina Presutti, Rocco Tripodi


Abstract
In this paper, we present an evaluation of two different approaches to the free-form Question Answering (QA) task. The main difference between the two approaches is that one is based on latent representations of knowledge, and the other uses explicit knowledge representation. For the evaluation, we developed DynaKnowledge, a new benchmark composed of questions concerning Wikipedia low-frequency entities. We wanted to ensure, on the one hand, that the questions are answerable and, on the other, that the models can provide information about very specific facts. The evaluation that we conducted highlights that the proposed benchmark is particularly challenging. The best model answers correctly only on 50% of the questions. Analysing the results, we also found that ChatGPT shows low reliance on low-frequency entity questions, manifesting a popularity bias. On the other hand, a simpler model based on explicit knowledge is less affected by this bias. With this paper, we want to provide a living benchmark for open-form QA to test knowledge and latent representation models on a dynamic benchmark.
Anthology ID:
2024.lrec-main.888
Volume:
Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024)
Month:
May
Year:
2024
Address:
Torino, Italia
Editors:
Nicoletta Calzolari, Min-Yen Kan, Veronique Hoste, Alessandro Lenci, Sakriani Sakti, Nianwen Xue
Venues:
LREC | COLING
SIG:
Publisher:
ELRA and ICCL
Note:
Pages:
10172–10185
Language:
URL:
https://aclanthology.org/2024.lrec-main.888
DOI:
Bibkey:
Cite (ACL):
Arianna Graciotti, Valentina Presutti, and Rocco Tripodi. 2024. Latent vs Explicit Knowledge Representation: How ChatGPT Answers Questions about Low-Frequency Entities. In Proceedings of the 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation (LREC-COLING 2024), pages 10172–10185, Torino, Italia. ELRA and ICCL.
Cite (Informal):
Latent vs Explicit Knowledge Representation: How ChatGPT Answers Questions about Low-Frequency Entities (Graciotti et al., LREC-COLING 2024)
Copy Citation:
PDF:
https://preview.aclanthology.org/nschneid-patch-4/2024.lrec-main.888.pdf
Optional supplementary material:
 2024.lrec-main.888.OptionalSupplementaryMaterial.zip