Hanghang Fan
2025
BLCU-ICALL at BEA 2025 Shared Task: Multi-Strategy Evaluation of AI Tutors
Jiyuan An
|
Xiang Fu
|
Bo Liu
|
Xuquan Zong
|
Cunliang Kong
|
Shuliang Liu
|
Shuo Wang
|
Zhenghao Liu
|
Liner Yang
|
Hanghang Fan
|
Erhong Yang
Proceedings of the 20th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2025)
This paper describes our approaches for the BEA-2025 Shared Task on assessing pedagogical ability and attributing tutor identities in AI-powered tutoring systems. We explored three methodological paradigms: in-context learning (ICL), supervised fine-tuning (SFT), and reinforcement learning from human feedback (RLHF). Results indicate clear methodological strengths: SFT is highly effective for structured classification tasks such as mistake identification and feedback actionability, while ICL with advanced prompting excels at open-ended tasks involving mistake localization and instructional guidance. Additionally, fine-tuned models demonstrated strong performance in identifying tutor authorship. Our findings highlight the importance of aligning methodological strategy and task structure, providing insights toward more effective evaluations of educational AI systems.
2023
Leveraging Prefix Transfer for Multi-Intent Text Revision
Ruining Chong
|
Cunliang Kong
|
Liu Wu
|
Zhenghao Liu
|
Ziye Jin
|
Liner Yang
|
Yange Fan
|
Hanghang Fan
|
Erhong Yang
Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)
Text revision is a necessary process to improve text quality. During this process, writers constantly edit texts out of different edit intentions. Identifying edit intention for a raw text is always an ambiguous work, and most previous work on revision systems mainly focuses on editing texts according to one specific edit intention. In this work, we aim to build a multi-intent text revision system that could revise texts without explicit intent annotation. Our system is based on prefix-tuning, which first gets prefixes for every edit intent, and then trains a prefix transfer module, enabling the system to selectively leverage the knowledge from various prefixes according to the input text. We conduct experiments on the IteraTeR dataset, and the results show that our system outperforms baselines. The system can significantly improve the SARI score with more than 3% improvements, which thrives on the learned editing intention prefixes.
Search
Fix author
Co-authors
- Cunliang Kong (孔存良) 2
- Zhenghao Liu 2
- Liner Yang (杨麟儿) 2
- Erhong Yang (杨尔弘) 2
- Jiyuan An 1
- show all...