BrailleLLM: Braille Instruction Tuning with Large Language Models for Braille Domain Tasks
Huang, Tianyuan, Zhu, Zepeng, Xing, Hangdi, Shao, Zirui, Yu, Zhi, Yang, Chaoxiong, He, Jiaxian, Liu, Xiaozhong, Bu, Jiajun
–arXiv.org Artificial Intelligence
Braille plays a vital role in education and information accessibility for visually impaired individuals. However, Braille information processing faces challenges such as data scarcity and ambiguities in mixed-text contexts. We construct English and Chinese Braille Mixed Datasets (EBMD/CBMD) with mathematical formulas to support diverse Braille domain research, and propose a syntax tree-based augmentation method tailored for Braille data. To address the underperformance of traditional fine-tuning methods in Braille-related tasks, we investigate Braille Knowledge-Based Fine-Tuning (BKFT), which reduces the learning difficulty of Braille contextual features. BrailleLLM employs BKFT via instruction tuning to achieve unified Braille translation, formula-to-Braille conversion, and mixed-text translation. Experiments demonstrate that BKFT achieves significant performance improvements over conventional fine-tuning in Braille translation scenarios. Our open-sourced datasets and methodologies establish a foundation for low-resource multilingual Braille research.
arXiv.org Artificial Intelligence
Oct-22-2025
- Country:
- Asia > China
- Shanghai > Shanghai (0.04)
- Zhejiang Province > Hangzhou (0.04)
- Europe > France (0.04)
- South America > Chile
- Asia > China
- Genre:
- Research Report (1.00)
- Industry:
- Education (0.46)
- Health & Medicine (0.49)
- Social Sector (0.34)
- Technology: