BrailleLLM: Braille Instruction Tuning with Large Language Models for Braille Domain Tasks

📅 2025-10-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Braille information processing faces two major challenges: data scarcity and ambiguity in mixed-text content (particularly mathematical notation). To address these, we propose BKFT, a braille-knowledge-driven instruction-tuning framework. BKFT introduces the first bilingual (English–Chinese) braille dataset featuring mixed text—including LaTeX-encoded mathematical formulas—and incorporates a syntax-tree structural enhancement strategy to improve data efficiency in low-resource settings. It unifies braille translation, mathematical formula-to-braille conversion, and mixed-text braille rendering into a single multi-task instruction-learning paradigm, explicitly injecting braille grammar and encoding rules (e.g., UEB and CBE standards) into model training. Experiments demonstrate that BKFT significantly outperforms conventional fine-tuning baselines on braille translation tasks. This work marks the first successful application of instruction tuning for multi-task braille language modeling. We publicly release both the dataset and code, establishing critical infrastructure for low-resource, multilingual braille research.

Technology Category

Application Category

📝 Abstract
Braille plays a vital role in education and information accessibility for visually impaired individuals. However, Braille information processing faces challenges such as data scarcity and ambiguities in mixed-text contexts. We construct English and Chinese Braille Mixed Datasets (EBMD/CBMD) with mathematical formulas to support diverse Braille domain research, and propose a syntax tree-based augmentation method tailored for Braille data. To address the underperformance of traditional fine-tuning methods in Braille-related tasks, we investigate Braille Knowledge-Based Fine-Tuning (BKFT), which reduces the learning difficulty of Braille contextual features. BrailleLLM employs BKFT via instruction tuning to achieve unified Braille translation, formula-to-Braille conversion, and mixed-text translation. Experiments demonstrate that BKFT achieves significant performance improvements over conventional fine-tuning in Braille translation scenarios. Our open-sourced datasets and methodologies establish a foundation for low-resource multilingual Braille research.
Problem

Research questions and friction points this paper is trying to address.

Addresses data scarcity and ambiguities in Braille information processing
Improves Braille translation performance through knowledge-based fine-tuning
Enables unified Braille processing including translation and formula conversion
Innovation

Methods, ideas, or system contributions that make the work stand out.

Constructed English and Chinese Braille datasets with formulas
Proposed syntax tree-based augmentation for Braille data
Employed Braille Knowledge-Based Fine-Tuning via instruction tuning
🔎 Similar Papers
No similar papers found.
T
Tianyuan Huang
Zhejiang Key Laboratory of Accessible Perception and Intelligent Systems, Zhejiang University
Z
Zepeng Zhu
Zhejiang Key Laboratory of Accessible Perception and Intelligent Systems, Zhejiang University
Hangdi Xing
Hangdi Xing
Student, Zhejiang University
Document UnderstandingVision-Language Models
Zirui Shao
Zirui Shao
Zhejiang University
Z
Zhi Yu
Zhejiang Key Laboratory of Accessible Perception and Intelligent Systems, Zhejiang University
C
Chaoxiong Yang
Zhejiang Key Laboratory of Accessible Perception and Intelligent Systems, Zhejiang University
J
Jiaxian He
Zhejiang Key Laboratory of Accessible Perception and Intelligent Systems, Zhejiang University
Xiaozhong Liu
Xiaozhong Liu
School of Informatics and Computing, Indiana University Bloomington
Information RetrievalNatural Language ProcessingDigital LibrarySemantic WebMetadata
J
Jiajun Bu
Zhejiang Key Laboratory of Accessible Perception and Intelligent Systems, Zhejiang University