Iteratively Enhanced Semidefinite Relaxations for Efficient Neural Network Verification
DOI:
https://doi.org/10.1609/aaai.v37i12.26744Keywords:
GeneralAbstract
We propose an enhanced semidefinite program (SDP) relaxation to enable the tight and efficient verification of neural networks (NNs). The tightness improvement is achieved by introducing a nonlinear constraint to existing SDP relaxations previously proposed for NN verification. The efficiency of the proposal stems from the iterative nature of the proposed algorithm in that it solves the resulting non-convex SDP by recursively solving auxiliary convex layer-based SDP problems. We show formally that the solution generated by our algorithm is tighter than state-of-the-art SDP-based solutions for the problem. We also show that the solution sequence converges to the optimal solution of the non-convex enhanced SDP relaxation. The experimental results on standard benchmarks in the area show that our algorithm achieves the state-of-the-art performance whilst maintaining an acceptable computational cost.Downloads
Published
2023-06-26
How to Cite
Lan, J., Zheng, Y., & Lomuscio, A. (2023). Iteratively Enhanced Semidefinite Relaxations for Efficient Neural Network Verification. Proceedings of the AAAI Conference on Artificial Intelligence, 37(12), 14937-14945. https://doi.org/10.1609/aaai.v37i12.26744
Issue
Section
AAAI Special Track on Safe and Robust AI