Ascribe New Dimensions to Scientific Data Visualization with VR

📅 2025-04-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Traditional 2D visualization techniques inadequately represent and analyze inherently 3D, multiscale scientific images, hindering microstructural understanding and quantitative analysis. To address this, we propose ASCRI BE-VR—a novel platform integrating AI-driven semantic segmentation with immersive VR interaction in a closed-loop paradigm. It employs a lightweight iterative feedback learning model for high-accuracy, fully automated segmentation of multi-source 3D volumetric data (e.g., CT, MRI), and leverages real-time VR rendering (Meta Quest) with natural gesture-based interaction to establish a human-in-the-loop digital twin visualization system. The platform achieves millisecond-scale responsiveness for hundred-GB-scale volumetric datasets. In materials science microstructure identification tasks, it improves analytical efficiency by over 3.2× and segmentation accuracy by +18.7%. ASCRI BE-VR establishes a new methodological framework for interpretable exploration of complex scientific imagery.

Technology Category

Application Category

📝 Abstract
For over half a century, the computer mouse has been the primary tool for interacting with digital data, yet it remains a limiting factor in exploring complex, multi-scale scientific images. Traditional 2D visualization methods hinder intuitive analysis of inherently 3D structures. Virtual Reality (VR) offers a transformative alternative, providing immersive, interactive environments that enhance data comprehension. This article introduces ASCRIBE-VR, a VR platform of Autonomous Solutions for Computational Research with Immersive Browsing &Exploration, which integrates AI-driven algorithms with scientific images. ASCRIBE-VR enables multimodal analysis, structural assessments, and immersive visualization, supporting scientific visualization of advanced datasets such as X-ray CT, Magnetic Resonance, and synthetic 3D imaging. Our VR tools, compatible with Meta Quest, can consume the output of our AI-based segmentation and iterative feedback processes to enable seamless exploration of large-scale 3D images. By merging AI-generated results with VR visualization, ASCRIBE-VR enhances scientific discovery, bridging the gap between computational analysis and human intuition in materials research, connecting human-in-the-loop with digital twins.
Problem

Research questions and friction points this paper is trying to address.

Overcoming limitations of 2D tools for 3D data exploration
Enhancing intuitive analysis of complex scientific images
Bridging AI-driven computation and immersive VR visualization
Innovation

Methods, ideas, or system contributions that make the work stand out.

VR platform integrates AI-driven algorithms
Supports multimodal and immersive 3D visualization
Combines AI segmentation with VR exploration
🔎 Similar Papers
No similar papers found.
Daniela Ushizima
Daniela Ushizima
Lawrence Berkeley National Laboratory, UC Berkeley, UC San Francisco
Computer VisionMachine LearningQuantitative MicroscopyVisualization & Explainable AI
G
Guilherme Melo dos Santos
Applied Math and Computational Research Division, Lawrence Berkeley National Laboratory
Zineb Sordo
Zineb Sordo
Lawrence National Berkeley Laboratory
Deep Learning and Computer vision
R
Ronald Pandolfi
Applied Math and Computational Research Division, Lawrence Berkeley National Laboratory
Jeffrey Donatelli
Jeffrey Donatelli
Staff Scientist, Lawrence Berkeley National Laboratory