PhotonSplat: 3D Scene Reconstruction and Colorization from SPAD Sensors

📅 2025-06-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address severe motion blur induced by high-speed motion and the reconstruction challenges posed by SPAD’s binary imaging modality, this paper proposes the first end-to-end neural rendering framework tailored for SPAD imaging. Methodologically, we design a SPAD-aware NeRF variant incorporating 3D spatial filtering for effective binary-image denoising; introduce a reference-free generative prior jointly with single-frame blur-map guidance for robust color recovery; and extend the framework to support dynamic scene modeling. Our key contributions are: (1) the first deep integration of SPAD physics, binary measurement characteristics, and neural radiance fields; (2) significant improvements in geometric accuracy and appearance fidelity on real-world PhotonScenes data under motion-blurred conditions; and (3) reconstructions that directly enable downstream vision tasks—including segmentation, detection, and editing—without additional post-processing.

Technology Category

Application Category

📝 Abstract
Advances in 3D reconstruction using neural rendering have enabled high-quality 3D capture. However, they often fail when the input imagery is corrupted by motion blur, due to fast motion of the camera or the objects in the scene. This work advances neural rendering techniques in such scenarios by using single-photon avalanche diode (SPAD) arrays, an emerging sensing technology capable of sensing images at extremely high speeds. However, the use of SPADs presents its own set of unique challenges in the form of binary images, that are driven by stochastic photon arrivals. To address this, we introduce PhotonSplat, a framework designed to reconstruct 3D scenes directly from SPAD binary images, effectively navigating the noise vs. blur trade-off. Our approach incorporates a novel 3D spatial filtering technique to reduce noise in the renderings. The framework also supports both no-reference using generative priors and reference-based colorization from a single blurry image, enabling downstream applications such as segmentation, object detection and appearance editing tasks. Additionally, we extend our method to incorporate dynamic scene representations, making it suitable for scenes with moving objects. We further contribute PhotonScenes, a real-world multi-view dataset captured with the SPAD sensors.
Problem

Research questions and friction points this paper is trying to address.

Reconstruct 3D scenes from noisy SPAD binary images
Address noise vs. blur trade-off in neural rendering
Enable colorization and dynamic scene representation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses SPAD arrays for high-speed 3D capture
Introduces PhotonSplat for binary image reconstruction
Incorporates 3D filtering and dynamic scene support
🔎 Similar Papers
No similar papers found.
S
Sai Sri Teja
Department of Electrical Engineering, Indian Institute of Technology, Madras, India
S
Sreevidya Chintalapati
Department of Electrical Engineering, Indian Institute of Technology, Madras, India
V
Vinayak Gupta
Department of Electrical Engineering, Indian Institute of Technology, Madras, India
Mukund Varma T
Mukund Varma T
University of California - San Diego
Neural RenderingComputer GraphicsComputer Vision
Haejoon Lee
Haejoon Lee
University of Michigan, Ann Arbor
multi-agent systemsmulti-robot systemsresilient network control
A
Aswin Sankaranarayanan
Department of Electrical and Computer Engineering, Carnegie Mellon University, Pittsburgh, PA, 15213
Kaushik Mitra
Kaushik Mitra
Department of Electrical Engineering, IIT Madras
Computational ImagingComputer VisionMachine Learning