Iteratively Enhanced Semidefinite Relaxations for Efficient Neural Network Verification

Authors

  • Jianglin Lan University of Glasgow
  • Yang Zheng University of California San Diego
  • Alessio Lomuscio Imperial College London

DOI:

https://doi.org/10.1609/aaai.v37i12.26744

Keywords:

General

Abstract

We propose an enhanced semidefinite program (SDP) relaxation to enable the tight and efficient verification of neural networks (NNs). The tightness improvement is achieved by introducing a nonlinear constraint to existing SDP relaxations previously proposed for NN verification. The efficiency of the proposal stems from the iterative nature of the proposed algorithm in that it solves the resulting non-convex SDP by recursively solving auxiliary convex layer-based SDP problems. We show formally that the solution generated by our algorithm is tighter than state-of-the-art SDP-based solutions for the problem. We also show that the solution sequence converges to the optimal solution of the non-convex enhanced SDP relaxation. The experimental results on standard benchmarks in the area show that our algorithm achieves the state-of-the-art performance whilst maintaining an acceptable computational cost.

Downloads

Published

2023-06-26

How to Cite

Lan, J., Zheng, Y., & Lomuscio, A. (2023). Iteratively Enhanced Semidefinite Relaxations for Efficient Neural Network Verification. Proceedings of the AAAI Conference on Artificial Intelligence, 37(12), 14937-14945. https://doi.org/10.1609/aaai.v37i12.26744

Issue

Section

AAAI Special Track on Safe and Robust AI