🤖 AI Summary
This work addresses the scarcity of morpheme segmentation annotations for low-resource languages. We propose a Transformer-based multitask learning framework that jointly models morpheme segmentation and word-level glossing, and—novelly—incorporates synthetic data generated via zero-shot or in-context learning from large language models (LLMs). To enhance cross-lingual generalization and alleviate data scarcity, the framework shares document-level representations across languages. Evaluated on the SIGMORPHON 2023 multilingual benchmark, our approach achieves significant improvements in word-level segmentation accuracy and morpheme-level F1 score, especially for extremely low-resource languages such as Old Church Slavonic and Ainu. These results demonstrate the effectiveness and strong generalization capability of synergistically combining multitask learning with LLM-generated synthetic data for morphological analysis in low-resource settings.
📝 Abstract
We introduce a transformer-based morpheme segmentation system that augments a low-resource training signal through multitask learning and LLM-generated synthetic data. Our framework jointly predicts morphological segments and glosses from orthographic input, leveraging shared linguistic representations obtained through a common documentary process to enhance model generalization. To further address data scarcity, we integrate synthetic training data generated by large language models (LLMs) using in-context learning. Experimental results on the SIGMORPHON 2023 dataset show that our approach significantly improves word-level segmentation accuracy and morpheme-level F1-score across multiple low-resource languages.