๐ค AI Summary
In semi-autonomous driving, drivers often exhibit delayed detection of unperceived hazards (e.g., unlabeled obstacles) and attentional tunneling during takeover. To address this, we propose a multimodal attention-guided situation awareness enhancement framework. Methodologically, it integrates real-time eye-tracking with context-aware dynamic saliency modeling to generate synchronized audiovisual cues, whose timing and modality weights are optimized according to human factors principles. Our key contribution is the first joint modeling of scene-semantic saliency and individual gaze state to enable proactive, adaptive attention regulation. Experimental results demonstrate that the framework significantly improves driver detection speed for unknown hazards (average reduction of 32% in reaction time) and recognition accuracy (+27.5%), effectively mitigating attentional tunneling and enhancing safety and coordination efficiency during humanโmachine handover.
๐ Abstract
The advent of autonomous driving systems promises to transform transportation by enhancing safety, efficiency, and comfort. As these technologies evolve toward higher levels of autonomy, the need for integrated systems that seamlessly support human involvement in decision-making becomes increasingly critical. Certain scenarios necessitate human involvement, including those where the vehicle is unable to identify an object or element in the scene, and as such cannot take independent action. Therefore, situational awareness is essential to mitigate potential risks during a takeover, where a driver must assume control and autonomy from the vehicle. The need for driver attention is important to avoid collisions with external agents and ensure a smooth transition during takeover operations. This paper explores the integration of attention redirection techniques, such as gaze manipulation through targeted visual and auditory cues, to help drivers maintain focus on emerging hazards and reduce target fixation in semi-autonomous driving scenarios. We propose a conceptual framework that combines real-time gaze tracking, context-aware saliency analysis, and synchronized visual and auditory alerts to enhance situational awareness, proactively address potential hazards, and foster effective collaboration between humans and autonomous systems.