Majid Nili Ahmadabadi
2026
DeepQuestion: Systematic Generation of Real-World Challenges for Evaluating LLMs Performance
Ali Khoramfar | Ali Ramezani | Mohammad Mahdi Mohajeri | Mohammad Javad Dousti | Majid Nili Ahmadabadi | Heshaam Faili
Proceedings of the Fifteenth Language Resources and Evaluation Conference
Ali Khoramfar | Ali Ramezani | Mohammad Mahdi Mohajeri | Mohammad Javad Dousti | Majid Nili Ahmadabadi | Heshaam Faili
Proceedings of the Fifteenth Language Resources and Evaluation Conference
While Large Language Models (LLMs) achieve near-human performance on standard benchmarks, their capabilities often fail to generalize to complex, real-world problems. To bridge this gap, we introduce DeepQuestion, a scalable, automated framework that systematically elevates the cognitive complexity of existing datasets through controlled task transformations grounded in explicit cognitive hierarchies. Based on Bloom’s taxonomy, DeepQuestion generates (1) scenario-based problems to test the application of knowledge in noisy, realistic contexts, and (2) instruction-based prompts that require models to create new questions from a given solution path, assessing synthesis and evaluation skills. Our extensive evaluation across ten leading open-source and proprietary models, covering both general-purpose and reasoning LLMs, reveals a stark performance decline—with accuracy dropping by up to 70%—as tasks ascend the cognitive hierarchy across evaluation settings. These findings underscore that current benchmarks overestimate true reasoning abilities and highlight the critical need for cognitively diverse evaluations to guide future LLM development.