Optimal neighbourhood selection in structural equation models

📅 2023-06-04
📈 Citations: 1
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the sample efficiency bottleneck in neighborhood selection for linear structural equation models (SEMs) under structured dependencies. We propose a novel klBSS estimator that overcomes the performance degradation of classical BSS and Lasso estimators in the presence of path cancellation—a common phenomenon in causal graphs. For the first time, we theoretically establish that even when hidden path cancellation occurs, the underlying structural constraints significantly reduce the sample complexity required for consistent neighborhood selection. The klBSS estimator is applicable to generalized structured SEMs and achieves a strictly tighter sample complexity bound than standard BSS. Through rigorous statistical analysis within a causal graph learning framework—and corroborated by extensive simulations—we demonstrate that explicit exploitation of structural constraints yields substantial empirical gains. This work provides a more efficient and robust subroutine for structure learning in high-dimensional graphical models.
📝 Abstract
We study the optimal sample complexity of neighbourhood selection in linear structural equation models, and compare this to best subset selection (BSS) for linear models under general design. We show by example that -- even when the structure is emph{unknown} -- the existence of underlying structure can reduce the sample complexity of neighbourhood selection. This result is complicated by the possibility of path cancellation, which we study in detail, and show that improvements are still possible in the presence of path cancellation. Finally, we support these theoretical observations with experiments. The proof introduces a modified BSS estimator, called klBSS, and compares its performance to BSS. The analysis of klBSS may also be of independent interest since it applies to arbitrary structured models, not necessarily those induced by a structural equation model. Our results have implications for structure learning in graphical models, which often relies on neighbourhood selection as a subroutine.
Problem

Research questions and friction points this paper is trying to address.

KL-BSS improves neighborhood selection in structural equation models
It recovers model support with fewer samples than classical methods
The method addresses limitations of both Lasso and best subset selection
Innovation

Methods, ideas, or system contributions that make the work stand out.

KL-BSS improves neighbourhood selection in SEM
Method leverages underlying structure using existing solvers
Achieves support recovery with fewer samples than BSS/Lasso
🔎 Similar Papers
No similar papers found.
M
Ming Gao
University of Chicago
W
Wai Ming Tai
University of Chicago
Bryon Aragam
Bryon Aragam
University of Chicago
machine learningstatisticsartificial intelligenceunsupervised learning