🤖 AI Summary
This work investigates whether VGGT (Vision Geometry Foundation Transformer), trained without explicit geometric supervision, implicitly acquires an understanding of camera geometry and scene structure—not merely relying on appearance-driven data priors. To this end, we conduct a systematic analysis of its internal representations via feature-space probing, attention-pattern visualization, input-space masking and perturbation experiments, and comparative evaluation against traditional multi-stage geometric pipelines. We establish, for the first time, that a purely data-driven, single-stage Transformer can spontaneously emerge geometric reasoning capabilities; specifically, its global self-attention layers implicitly perform cross-view correspondence matching and encode epipolar geometric constraints. Experiments demonstrate VGGT’s strong geometric robustness under occlusion, appearance variation, and camera configuration changes. Moreover, its implicit geometric understanding synergizes with data priors to enhance 3D perception performance.
📝 Abstract
The Visual Geometry Grounded Transformer (VGGT) is a 3D foundation model that infers camera geometry and scene structure in a single feed-forward pass. Trained in a supervised, single-step fashion on large datasets, VGGT raises a key question: does it build upon geometric concepts like traditional multi-view methods, or does it rely primarily on learned appearance-based data-driven priors? In this work, we conduct a systematic analysis of VGGT's internal mechanisms to uncover whether geometric understanding emerges within its representations. By probing intermediate features, analyzing attention patterns, and performing interventions, we examine how the model implements its functionality. Our findings reveal that VGGT implicitly performs correspondence matching within its global attention layers and encodes epipolar geometry, despite being trained without explicit geometric constraints. We further investigate VGGT's dependence on its learned data priors. Using spatial input masking and perturbation experiments, we assess its robustness to occlusions, appearance variations, and camera configurations, comparing it with classical multi-stage pipelines. Together, these insights highlight how VGGT internalizes geometric structure while using learned data-driven priors.