EmojiVoice: Towards long-term controllable expressivity in robot speech

📅 2025-06-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address monotonous expressiveness and declining user engagement in long-term voice interactions with social robots, this paper proposes a lightweight, offline-deployable expressive text-to-speech (TTS) method. Our approach builds upon an enhanced Matcha-TTS architecture and introduces *emoji-prompting*—a novel mechanism that jointly embeds emoji semantics and models dynamic prosody for phase-level, temporally adaptive emotional control. To our knowledge, this is the first work enabling real-time, long-duration, fine-grained expressive control on resource-constrained robotic platforms. Experiments in storytelling tasks demonstrate significant improvements in speech naturalness and perceived expressiveness. Furthermore, we uncover scenario-dependent expressive preferences—e.g., between assistant and educational use cases—thereby empirically delineating the method’s applicability boundaries. The proposed framework bridges the gap between expressive TTS and practical deployment constraints in social robotics.

Technology Category

Application Category

📝 Abstract
Humans vary their expressivity when speaking for extended periods to maintain engagement with their listener. Although social robots tend to be deployed with ``expressive'' joyful voices, they lack this long-term variation found in human speech. Foundation model text-to-speech systems are beginning to mimic the expressivity in human speech, but they are difficult to deploy offline on robots. We present EmojiVoice, a free, customizable text-to-speech (TTS) toolkit that allows social roboticists to build temporally variable, expressive speech on social robots. We introduce emoji-prompting to allow fine-grained control of expressivity on a phase level and use the lightweight Matcha-TTS backbone to generate speech in real-time. We explore three case studies: (1) a scripted conversation with a robot assistant, (2) a storytelling robot, and (3) an autonomous speech-to-speech interactive agent. We found that using varied emoji prompting improved the perception and expressivity of speech over a long period in a storytelling task, but expressive voice was not preferred in the assistant use case.
Problem

Research questions and friction points this paper is trying to address.

Enable long-term expressive speech in robots
Provide offline-deployable expressive TTS for robots
Control robot speech expressivity via emoji-prompting
Innovation

Methods, ideas, or system contributions that make the work stand out.

Emoji-prompting for fine-grained expressivity control
Lightweight Matcha-TTS backbone for real-time speech
Customizable TTS toolkit for variable expressive speech
P
Paige Tuttos'i
School of Computing Science, Simon Fraser University
Shivam Mehta
Shivam Mehta
Research Scientist Netflix - PhD @ KTH Royal Institute of Technology & WASP AI
Probabilistic Machine LearningDeep LearningSpeech SynthesisGenerative ModelsGen AI
Z
Zachary Syvenky
School of Computing Science, Simon Fraser University
B
Bermet Burkanova
School of Computing Science, Simon Fraser University
G
G. Henter
Division of Speech Music and Hearing, KTH Royal Institute of Technology
A
Angelica Lim
School of Computing Science, Simon Fraser University