Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Provably-Safe Neural Network Training Using Hybrid Zonotope Reachability Analysis

Created by
  • Haebom

Author

Long Kiu Chung, Shreyas Kousik

Outline

To address the challenge of tightening constraints on neural network outputs in safety-critical control applications, this paper proposes a neural network training method that derives accurate images of nonconvex input sets for neural networks with ReLU nonlinearities to avoid nonconvex unsafe regions. This is achieved through reachability analysis using an extended hybrid zonotope set representation that enables differentiable collision detection via mixed-integer linear programming (MILP). The method is proven to be efficient and fast for networks with up to 240 neurons, and its computational complexity is dominated by inverse operations on matrices that scale linearly with the number of neurons and the complexity of the input and unsafe sets. We demonstrate its practicality in training forward-invariant neural network controllers for affine dynamic systems with nonconvex input sets and in generating safe reach-avoidance plans for black-box dynamic systems.

Takeaways, Limitations

Takeaways: An efficient and fast method for training neural networks that avoid nonconvex unsafe regions is presented. Its effectiveness is demonstrated on networks with up to 240 neurons. The feasibility of generating safe control and planning for affine dynamic systems and black-box systems is demonstrated. Efficient reachability analysis is performed using differentiable collision detection using an extended hybrid zonotope.
Limitations: Performance verification is needed for networks with more than 240 neurons. Since computational complexity is linearly proportional to the number of neurons and the complexity of the input/uncertainty set, scalability may be limited for large networks or complex sets. Generalization to neural networks with various types of nonlinearities is also needed.
👍