🤖 AI Summary
Wearable AI systems often rely on user-initiated triggers or predefined task logic, neglecting real-time cognitive states—leading to ill-timed interventions. Method: This paper proposes a proactive assistance framework grounded in dynamic visual working memory (WM) modeling. Leveraging multimodal wearable sensors—including smart glasses—it captures physiological and behavioral signals in real time; integrates cognitive theory to construct an encoding-interference WM representation model for fine-grained perception of cognitive load and attentional state; and introduces, for the first time in wearables, a WM-informed interruption cost–benefit trade-off framework to optimize intervention timing. The system integrates sensor fusion, WM modeling, real-time attention recognition, and LLM-augmented collaborative reasoning. Contribution/Results: A 12-participant user study demonstrates significant improvements over a pure-LLM baseline: +32.7% accuracy in assistance timing (p < 0.01) and significantly higher user engagement (p < 0.01), validating the efficacy and feasibility of cognition-aligned proactive intervention.
📝 Abstract
Wearable AI systems aim to provide timely assistance in daily life, but existing approaches often rely on user initiation or predefined task knowledge, neglecting users' current mental states. We introduce ProMemAssist, a smart glasses system that models a user's working memory (WM) in real-time using multi-modal sensor signals. Grounded in cognitive theories of WM, our system represents perceived information as memory items and episodes with encoding mechanisms, such as displacement and interference. This WM model informs a timing predictor that balances the value of assistance with the cost of interruption. In a user study with 12 participants completing cognitively demanding tasks, ProMemAssist delivered more selective assistance and received higher engagement compared to an LLM baseline system. Qualitative feedback highlights the benefits of WM modeling for nuanced, context-sensitive support, offering design implications for more attentive and user-aware proactive agents.