ELMI: Interactive and Intelligent Sign Language Translation of Lyrics for Song Signing

📅 2024-09-15
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current sign language song translation faces challenges including semantic distortion, syntactic misalignment, and difficulty synchronizing affective expression with musical rhythm, compounded by a lack of accessible, domain-specific tools. This paper introduces the first paradigm integrating real-time audio-visual synchronization editing with large language model (LLM)-guided multidimensional dialogue—covering semantics, facial expression, and prosody—for sign language song translation. The system supports gloss-based visual morpheme editing, time-aligned audio-video embedding, and Deaf-centered interaction design. An LLM-driven collaborative translation interface, augmented by user behavior feedback, ensures cultural sensitivity and practical usability. In a field study involving 13 professional sign language interpreters, the tool significantly enhanced translators’ confidence and autonomy: all participants successfully completed original song translations, and their feedback was constructive, motivating, and information-rich.

Technology Category

Application Category

📝 Abstract
d/Deaf and hearing song-signers have become prevalent across video-sharing platforms, but translating songs into sign language remains cumbersome and inaccessible. Our formative study revealed the challenges song-signers face, including semantic, syntactic, expressive, and rhythmic considerations in translations. We present ELMI, an accessible song-signing tool that assists in translating lyrics into sign language. ELMI enables users to edit glosses line-by-line, with real-time synced lyric and music video snippets. Users can also chat with a large language model-driven AI to discuss meaning, glossing, emoting, and timing. Through an exploratory study with 13 song-signers, we examined how ELMI facilitates their workflows and how song-signers leverage and receive an LLM-driven chat for translation. Participants successfully adopted ELMI to song-signing, with active discussions throughout. They also reported improved confidence and independence in their translations, finding ELMI encouraging, constructive, and informative. We discuss research and design implications for accessible and culturally sensitive song-signing translation tools.
Problem

Research questions and friction points this paper is trying to address.

Addresses inaccessibility in song-signing translation
Integrates AI for real-time lyric translation assistance
Enhances confidence in d/Deaf song-signers' translations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Interactive sign language translation tool
Real-time synced lyric editing
AI-driven chat for translation support
🔎 Similar Papers
No similar papers found.
S
Suhyeon Yoo
University of Toronto, Toronto, Canada (conducted this work as a research intern at NAVER AI Lab)
K
Khai-Nghi Truong
University of Toronto, Toronto, Canada
Young-Ho Kim
Young-Ho Kim
Research Scientist at NAVER AI Lab
Human-Computer InteractionPersonal InformaticsInclusive AILarge Language Models