Jiaxian He
2025
BrailleLLM: Braille Instruction Tuning with Large Language Models for Braille Domain Tasks
Tianyuan Huang
|
Zepeng Zhu
|
Hangdi Xing
|
Zirui Shao
|
Zhi Yu
|
Chaoxiong Yang
|
Jiaxian He
|
Xiaozhong Liu
|
Jiajun Bu
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Braille plays a vital role in education and information accessibility for visually impaired individuals. However, Braille information processing faces challenges such as data scarcity and ambiguities in mixed-text contexts. We construct English and Chinese Braille Mixed Datasets (EBMD/CBMD) with mathematical formulas to support diverse Braille domain research, and propose a syntax tree-based augmentation method tailored for Braille data. To address the underperformance of traditional fine-tuning methods in braille-related tasks, we investigate Braille Knowledge-Based Fine-Tuning (BKFT), which reduces the learning difficulty of Braille contextual features. BrailleLLM employs BKFT via instruction tuning to achieve unified Braille translation, formula-to-Braille conversion, and mixed-text translation. Experiments demonstrate that BKFT achieves significant performance improvements over conventional fine-tuning in Braille translation scenarios. Our open-sourced datasets and methodologies establish a foundation for low-resource multilingual Braille research.
Search
Fix author
Co-authors
- Jiajun Bu 1
- Tianyuan Huang 1
- Xiaozhong Liu 1
- Zirui Shao 1
- Hangdi Xing 1
- show all...