🤖 AI Summary
VR-induced motion sickness primarily stems from sensory conflict between visual motion (optical flow) and vestibular perception, limiting VR applications in highly dynamic scenes. To address this, we propose U-MAD—a lightweight, real-time AI framework that introduces learnable diffusion models for optical flow perception modulation. Unlike prior approaches, U-MAD operates directly at the image level without requiring scene priors, 3D reconstruction, or mesh editing. It comprises three core components: a diffusion-driven optical flow estimation network, a frame-level real-time processing architecture, and a perception-guided motion attenuation algorithm—collectively reducing mean optical flow magnitude and enhancing temporal stability. User studies demonstrate that U-MAD significantly alleviates motion sickness symptoms and improves immersive comfort. Moreover, it supports plug-and-play integration and is compatible with procedurally generated environments.
📝 Abstract
Cybersickness remains a critical barrier to the widespread adoption of Virtual Reality (VR), particularly in scenarios involving intense or artificial motion cues. Among the key contributors is excessive optical flow-perceived visual motion that, when unmatched by vestibular input, leads to sensory conflict and discomfort. While previous efforts have explored geometric or hardware based mitigation strategies, such methods often rely on predefined scene structures, manual tuning, or intrusive equipment. In this work, we propose U-MAD, a lightweight, real-time, AI-based solution that suppresses perceptually disruptive optical flow directly at the image level. Unlike prior handcrafted approaches, this method learns to attenuate high-intensity motion patterns from rendered frames without requiring mesh-level editing or scene specific adaptation. Designed as a plug and play module, U-MAD integrates seamlessly into existing VR pipelines and generalizes well to procedurally generated environments. The experiments show that U-MAD consistently reduces average optical flow and enhances temporal stability across diverse scenes. A user study further confirms that reducing visual motion leads to improved perceptual comfort and alleviated cybersickness symptoms. These findings demonstrate that perceptually guided modulation of optical flow provides an effective and scalable approach to creating more user-friendly immersive experiences. The code will be released at https://github.com/XXXXX (upon publication).