🤖 AI Summary
Traditional large language model (LLM)-based writing feedback is static, imposes high cognitive load, and fails to foster higher-order writing awareness or substantive revision. Method: This study proposes a “static feedback–voice dialogue” transformation paradigm, converting LLM-generated textual feedback into interactive, queryable, example-supported, and clarifiable voice-based dialogue prompts. The system integrates automatic speech recognition (ASR), text-to-speech (TTS), and multimodal LLMs, validated through a formative experimental design. Contribution/Results: Compared to text-only interaction, voice-driven dialogue significantly enhances writers’ attention to structural and logical issues, deepens reflective engagement, and improves the quality of subsequent revisions. This work provides the first systematic empirical evidence of voice interaction’s superiority in intelligent writing assistance, offering both theoretical grounding and an implementable interaction paradigm for next-generation, human-like, low-cognitive-load, and highly engaging writing tutors.
📝 Abstract
Writing well requires not only expressing ideas but also refining them through revision, a process facilitated by reflection. Prior research suggests that feedback delivered through dialogues, such as those in writing center tutoring sessions, can help writers reflect more thoughtfully on their work compared to static feedback. Recent advancements in multi-modal large language models (LLMs) now offer new possibilities for supporting interactive and expressive voice-based reflection in writing. In particular, we propose that LLM-generated static feedback can be repurposed as conversation starters, allowing writers to seek clarification, request examples, and ask follow-up questions, thereby fostering deeper reflection on their writing. We argue that voice-based interaction can naturally facilitate this conversational exchange, encouraging writers' engagement with higher-order concerns, facilitating iterative refinement of their reflections, and reduce cognitive load compared to text-based interactions. To investigate these effects, we propose a formative study exploring how text vs. voice input influence writers' reflection and subsequent revisions. Findings from this study will inform the design of intelligent and interactive writing tools, offering insights into how voice-based interactions with LLM-powered conversational agents can support reflection and revision.