🤖 AI Summary
This paper studies smooth nonconvex optimization problems subject to simple polyhedral and convex inequality constraints. Under a local error bound condition—rather than the stronger Mangasarian–Fromovitz or linear independence constraint qualification—it proposes the inexact Moreau–envelope Lagrangian (iMELa) algorithm. At each main iteration, iMELa solves a convex subproblem incorporating a proximal term, synergistically combining Moreau envelope smoothing, Lagrangian relaxation, and proximal gradient techniques. We establish, for the first time under a local error bound, that iMELa converges to an ε-KKT point with optimal gradient complexity Õ(ε⁻²), thereby overcoming theoretical limitations of classical penalty and augmented Lagrangian methods. Crucially, this result holds under significantly weaker constraint regularity assumptions, enhancing both the theoretical efficiency and practical applicability of algorithms for nonconvex constrained optimization.
📝 Abstract
In this paper, we study the inexact Moreau envelope Lagrangian (iMELa) method for solving smooth non-convex optimization problems over a simple polytope with additional convex inequality constraints. By incorporating a proximal term into the traditional Lagrangian function, the iMELa method approximately solves a convex optimization subproblem over the polyhedral set at each main iteration. Under the assumption of a local error bound condition for subsets of the feasible set defined by subsets of the constraints, we establish that the iMELa method can find an $epsilon$-Karush-Kuhn-Tucker point with $ ilde O(epsilon^{-2})$ gradient oracle complexity.