🤖 AI Summary
This paper investigates necessary and sufficient conditions for the composition operator (C_h f = f circ h) to map Riesz bases or orthonormal bases between (L^2(Omega_1)) and (L^2(Omega_2)). Focusing on the geometric perturbation induced by the mapping (h: Omega_2 o Omega_1) on standard orthonormal bases, we establish the first complete characterization: (C_h) preserves Riesz bases—i.e., maps any Riesz basis of (L^2(Omega_1)) onto a Riesz basis of (L^2(Omega_2))—if and only if (h) is a bi-Lipschitz bijection. Moreover, (C_h) preserves orthonormal bases if and only if (h) is additionally measure-preserving (up to equivalence). This equivalence reveals a fundamental connection between geometric regularity of (h) and stability of basis structures under composition. Building upon this characterization, we propose a novel paradigm for constructing complete, optimally approximating sequences via bijective neural networks. Our framework provides a rigorous functional-analytic foundation for invertible neural network approximation and yields verifiable constructive criteria for basis preservation.
📝 Abstract
We investigate perturbations of orthonormal bases of $L^2$ via a composition operator $C_h$ induced by a mapping $h$. We provide a comprehensive characterization of the mapping $h$ required for the perturbed sequence to form an orthonormal or Riesz basis. Restricting our analysis to differentiable mappings, we reveal that all Riesz bases of the given form are induced by bi-Lipschitz mappings. In addition, we discuss implications of these results for approximation theory, highlighting the potential of using bijective neural networks to construct complete sequences with favorable approximation properties.