4DRadar-GS: Self-Supervised Dynamic Driving Scene Reconstruction with 4D Radar

πŸ“… 2025-09-16
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
In dynamic driving scenes, self-supervised 3D reconstruction and novel-view synthesis suffer from inaccurate motion estimation, temporal inconsistency, and incomplete reconstruction of dynamic objects. To address these challenges, this paper proposes the first 4D radar-enhanced self-supervised framework. Our key contributions are: (1) a Gaussian point cloud initialization method leveraging 4D radar’s velocity and spatial measurements to enable precise dynamic object segmentation and metric depth recovery; and (2) a velocity-guided fine-grained trajectory tracking model that jointly optimizes differentiable rendering and point cloud motion under scene flow supervision, ensuring strong temporal consistency. Evaluated on the OmniHD-Scenes dataset, our method significantly improves dynamic object completeness and inter-frame consistency, achieving state-of-the-art performance for dynamic driving scene reconstruction.

Technology Category

Application Category

πŸ“ Abstract
3D reconstruction and novel view synthesis are critical for validating autonomous driving systems and training advanced perception models. Recent self-supervised methods have gained significant attention due to their cost-effectiveness and enhanced generalization in scenarios where annotated bounding boxes are unavailable. However, existing approaches, which often rely on frequency-domain decoupling or optical flow, struggle to accurately reconstruct dynamic objects due to imprecise motion estimation and weak temporal consistency, resulting in incomplete or distorted representations of dynamic scene elements. To address these challenges, we propose 4DRadar-GS, a 4D Radar-augmented self-supervised 3D reconstruction framework tailored for dynamic driving scenes. Specifically, we first present a 4D Radar-assisted Gaussian initialization scheme that leverages 4D Radar's velocity and spatial information to segment dynamic objects and recover monocular depth scale, generating accurate Gaussian point representations. In addition, we propose a Velocity-guided PointTrack (VGPT) model, which is jointly trained with the reconstruction pipeline under scene flow supervision, to track fine-grained dynamic trajectories and construct temporally consistent representations. Evaluated on the OmniHD-Scenes dataset, 4DRadar-GS achieves state-of-the-art performance in dynamic driving scene 3D reconstruction.
Problem

Research questions and friction points this paper is trying to address.

Reconstructing dynamic driving scenes with accurate motion estimation
Overcoming incomplete representations in self-supervised 3D reconstruction
Addressing temporal consistency issues in dynamic object reconstruction
Innovation

Methods, ideas, or system contributions that make the work stand out.

4D Radar-assisted Gaussian initialization scheme
Velocity-guided PointTrack model for tracking
Self-supervised 3D reconstruction framework
πŸ”Ž Similar Papers
No similar papers found.
X
Xiao Tang
School of Automotive Studies, Tongji University, Shanghai, China
G
Guirong Zhuo
School of Automotive Studies, Tongji University, Shanghai, China
C
Cong Wang
State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences
B
Boyuan Zheng
School of Automotive Studies, Tongji University, Shanghai, China
M
Minqing Huang
School of Automotive Studies, Tongji University, Shanghai, China
Lianqing Zheng
Lianqing Zheng
Tongji University Ph.D student
BEV/OCCVLA4D Radar PerceptionMultimodal FusionData Closed-Loop
L
Long Chen
State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences
Shouyi Lu
Shouyi Lu
TongJi University
AIGCImage EditingPoint Cloud Super-ResolutionPose EstimationPlace Recognition