REOcc: Camera-Radar Fusion with Radar Feature Enrichment for 3D Occupancy Prediction

📅 2025-11-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Pure vision-based 3D occupancy prediction suffers from limited performance in complex scenes, while camera-radar fusion is hindered by radar sparsity and noise, leading to insufficient feature representation. To address these challenges, this paper proposes REOcc, a novel camera-radar fusion network. Its core innovation is the Radar Densification and Amplification (RDA) module, which employs learnable voxel-wise interpolation and context-aware enhancement to significantly improve radar feature density and robustness. Furthermore, a cross-modal feature interaction mechanism is introduced to enable fine-grained alignment and complementary fusion of image semantics and radar geometry. Evaluated on the Occ3D-nuScenes benchmark, REOcc substantially outperforms pure vision methods: notably, it achieves a 12.7% mIoU gain for dynamic objects (e.g., vehicles and pedestrians), demonstrating the effectiveness of its multimodal complementary modeling.

Technology Category

Application Category

📝 Abstract
Vision-based 3D occupancy prediction has made significant advancements, but its reliance on cameras alone struggles in challenging environments. This limitation has driven the adoption of sensor fusion, among which camera-radar fusion stands out as a promising solution due to their complementary strengths. However, the sparsity and noise of the radar data limits its effectiveness, leading to suboptimal fusion performance. In this paper, we propose REOcc, a novel camera-radar fusion network designed to enrich radar feature representations for 3D occupancy prediction. Our approach introduces two main components, a Radar Densifier and a Radar Amplifier, which refine radar features by integrating spatial and contextual information, effectively enhancing spatial density and quality. Extensive experiments on the Occ3D-nuScenes benchmark demonstrate that REOcc achieves significant performance gains over the camera-only baseline model, particularly in dynamic object classes. These results underscore REOcc's capability to mitigate the sparsity and noise of the radar data. Consequently, radar complements camera data more effectively, unlocking the full potential of camera-radar fusion for robust and reliable 3D occupancy prediction.
Problem

Research questions and friction points this paper is trying to address.

Enhancing 3D occupancy prediction in challenging environments
Addressing radar data sparsity and noise for sensor fusion
Improving camera-radar fusion performance with enriched radar features
Innovation

Methods, ideas, or system contributions that make the work stand out.

Camera-radar fusion network enriches radar features
Radar Densifier integrates spatial information to enhance density
Radar Amplifier refines contextual information to improve quality
🔎 Similar Papers
No similar papers found.
C
Chaehee Song
School of Electrical Engineering, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, South Korea
S
Sanmin Kim
Department of Automobile and IT Convergence, Kookmin University, Seoul 02707, South Korea
H
Hyeonjun Jeong
Cho Chun Shik Graduate School of Mobility, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, South Korea
J
Juyeb Shin
Robotics Program, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, South Korea
J
Joonhee Lim
Robotics Program, Korea Advanced Institute of Science and Technology (KAIST), Daejeon 34141, South Korea
Dongsuk Kum
Dongsuk Kum
KAIST
Vehicle Dynamics & Control