π€ AI Summary
Formal safety verification tools are lacking for neural feedback closed-loop systems.
Method: This paper proposes a backward under-approximation method for reachable sets of nonlinear discrete-time systems. It constructs an over-approximation model of system dynamics and reformulates reachable set computation as a tractable mixed-integer linear programming (MILP) problem.
Contribution/Results: The approach enables, for the first time, rigorous backward under-approximation of nonlinear discrete closed-loop systems incorporating neural network controllers. We prove the algorithmβs completeness, guaranteeing that the computed set strictly contains the true backward reachable set. Numerical experiments demonstrate its effectiveness in verifying goal-achievement properties, significantly expanding the scope of safety properties amenable to formal verification in learning-based control systems.
π Abstract
Learning-enabled planning and control algorithms are increasingly popular, but they often lack rigorous guarantees of performance or safety. We introduce an algorithm for computing underapproximate backward reachable sets of nonlinear discrete time neural feedback loops. We then use the backward reachable sets to check goal-reaching properties. Our algorithm is based on overapproximating the system dynamics function to enable computation of underapproximate backward reachable sets through solutions of mixed-integer linear programs. We rigorously analyze the soundness of our algorithm and demonstrate it on a numerical example. Our work expands the class of properties that can be verified for learning-enabled systems.