🤖 AI Summary
This study addresses the limited environment-aware visual agency of voice-based assistants in remote visual assistance for people with visual impairments. By analyzing conversational data from real-world tasks where visually impaired users collaborated with either human helpers or multimodal voice agents, the research offers the first fine-grained characterization of context-sensitive proactive visual practices employed by human remote assistants. Findings reveal that current voice agents lack mechanisms to generate environment-triggered visual actions, rendering them unable to replicate the dynamic, visually guided initiations characteristic of human assistants. This fundamental limitation underscores a critical gap in achieving truly proactive assistance and highlights essential directions for future design of multimodal agents capable of situated, responsive visual guidance.
📝 Abstract
Does human-AI assistance unfold in the same way as human-human assistance? This research explores what can be learned from the expertise of blind individuals and sighted volunteers to inform the design of multimodal voice agents and address the enduring challenge of proactivity. Drawing on granular analysis of two representative fragments from a larger corpus, we contrast the practices co-produced by an experienced human remote sighted assistant and a blind participant-as they collaborate to find a stain on a blanket over the phone-with those achieved when the same participant worked with a multimodal voice agent on the same task, a few moments earlier. This comparison enables us to specify precisely which fundamental proactive practices the agent did not enact in situ. We conclude that, so long as multimodal voice agents cannot produce environmentally occasioned vision-based actions, they will lack a key resource relied upon by human remote sighted assistants.