Muye Huang
2025
Diagram-Driven Course Questions Generation
Xinyu Zhang
|
Lingling Zhang
|
Yanrui Wu
|
Muye Huang
|
Wenjun Wu
|
Bo Li
|
Shaowei Wang
|
Basura Fernando
|
Jun Liu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Visual Question Generation (VQG) research focuses predominantly on natural images while neglecting the diagram, which is a critical component in educational materials. To meet the needs of pedagogical assessment, we propose the Diagram-Driven Course Questions Generation (DDCQG) task and construct DiagramQG, a comprehensive dataset with 15,720 diagrams and 25,798 questions across 37 subjects and 371 courses. Our approach employs course and input text constraints to generate course-relevant questions about specific diagram elements. We reveal three challenges of DDCQG: domain-specific knowledge requirements across courses, long-tail distribution in course coverage, and high information density in diagrams. To address these, we propose the Hierarchical Knowledge Integration framework (HKI-DDCQG), which utilizes trainable CLIP for identifying relevant diagram patches, leverages frozen vision-language models for knowledge extraction, and generates questions with trainable T5. Experiments demonstrate that HKI-DDCQG outperforms existing models on DiagramQG while maintaining strong generalizability across natural image datasets, establishing a strong baseline for DDCQG.
Search
Fix author
Co-authors
- Basura Fernando 1
- Bo Li 1
- Jun Liu 1
- Shaowei Wang 1
- Yanrui Wu 1
- show all...