How DNNs break the Curse of Dimensionality: Compositionality and Symmetry Learning

📅 2024-07-08
🏛️ arXiv.org
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses how deep neural networks (DNNs) overcome the “curse of dimensionality” in high-dimensional learning. Method: It proposes a modeling paradigm grounded in function composition structures (h∘g) and symmetry quotient mappings, integrating covering number theory with F₁-norm-based learnability characterization, and combining Sobolev/Barron norm-constrained optimization with symmetry-driven dimensionality reduction. Contribution/Results: The work establishes novel generalization bounds for composite functions under symmetry, rigorously proving that DNN global minima achieve ε-accuracy approximation using only O(1/ε²) samples. It uncovers a phase transition mechanism arising from asymmetric learning—where g performs smooth dimensionality reduction and h exhibits low regularity—and experimentally validates both the predicted phase transition and dimension-independent convergence rates. These results provide new theoretical foundations for sample-efficient, high-dimensional learning with DNNs.

Technology Category

Application Category

📝 Abstract
We show that deep neural networks (DNNs) can efficiently learn any composition of functions with bounded $F_{1}$-norm, which allows DNNs to break the curse of dimensionality in ways that shallow networks cannot. More specifically, we derive a generalization bound that combines a covering number argument for compositionality, and the $F_{1}$-norm (or the related Barron norm) for large width adaptivity. We show that the global minimizer of the regularized loss of DNNs can fit for example the composition of two functions $f^{*}=hcirc g$ from a small number of observations, assuming $g$ is smooth/regular and reduces the dimensionality (e.g. $g$ could be the quotient map of the symmetries of $f^{*}$), so that $h$ can be learned in spite of its low regularity. The measures of regularity we consider is the Sobolev norm with different levels of differentiability, which is well adapted to the $F_{1}$ norm. We compute scaling laws empirically and observe phase transitions depending on whether $g$ or $h$ is harder to learn, as predicted by our theory.
Problem

Research questions and friction points this paper is trying to address.

DNNs break dimensionality curse via function composition learning.
Generalization bound combines covering number and F1-norm adaptivity.
DNNs fit composite functions from limited observations efficiently.
Innovation

Methods, ideas, or system contributions that make the work stand out.

DNNs break dimensionality curse via compositionality.
Generalization bound combines covering number and F1-norm.
Empirical scaling laws show phase transitions in learning.
🔎 Similar Papers
No similar papers found.
Arthur Jacot
Arthur Jacot
Assistant Professor, Courant Institute of Mathematical Sciences, NYU
Deep Learning
S
Seok Hoan Choi
Courant Institute, New York University, New York, NY 10012
Y
Yuxiao Wen
Courant Institute, New York University, New York, NY 10012