Baihui Zheng
2025
Chinese SafetyQA: A Safety Short-form Factuality Benchmark for Large Language Models
Yingshui Tan
|
Boren Zheng
|
Baihui Zheng
|
Kerui Cao
|
Huiyun Jing
|
Jincheng Wei
|
Jiaheng Liu
|
Yancheng He
|
Wenbo Su
|
Xiaoyong Zhu
|
Bo Zheng
|
Kaifu Zhang
Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)
With the rapid advancement of Large Language Models (LLMs), significant safety concerns have emerged. Fundamentally, the safety of large language models is closely linked to the accuracy, comprehensiveness, and clarity of their understanding of safety knowledge, particularly in domains such as law, policy and ethics. This factuality ability is crucial in determining whether these models can be deployed and applied safely and compliantly within specific regions. To address these challenges and better evaluate the factuality ability of LLMs to answer short question, we introduce the Chinese SafetyQA benchmark. Chinese SafetyQA has several properties (i.e., Chinese, Diverse, High-quality, Static, Easy-to-evaluate, safety-related, harmless). Based on Chinese SafetyQA, we perform a comprehensive evaluation on the factuality abilities of existing LLMs and analyze how these capabilities relate to LLM abilities, e.g., RAG ability and robustness against attacks.
Search
Fix author
Co-authors
- Kerui Cao 1
- Yancheng He 1
- Huiyun Jing 1
- Jiaheng Liu 1
- Wenbo Su 1
- show all...
Venues
- acl1