Abstract Interpretation of Decision Tree Ensemble Classifiers

Authors

  • Francesco Ranzato University of Padova
  • Marco Zanella University of Padova

DOI:

https://doi.org/10.1609/aaai.v34i04.5998

Abstract

We study the problem of formally and automatically verifying robustness properties of decision tree ensemble classifiers such as random forests and gradient boosted decision tree models. A recent stream of works showed how abstract interpretation, which is ubiquitously used in static program analysis, can be successfully deployed to formally verify (deep) neural networks. In this work we push forward this line of research by designing a general and principled abstract interpretation-based framework for the formal verification of robustness and stability properties of decision tree ensemble models. Our abstract interpretation-based method may induce complete robustness checks of standard adversarial perturbations and output concrete adversarial attacks. We implemented our abstract verification technique in a tool called silva, which leverages an abstract domain of not necessarily closed real hyperrectangles and is instantiated to verify random forests and gradient boosted decision trees. Our experimental evaluation on the MNIST dataset shows that silva provides a precise and efficient tool which advances the current state of the art in tree ensembles verification.

Downloads

Published

2020-04-03

How to Cite

Ranzato, F., & Zanella, M. (2020). Abstract Interpretation of Decision Tree Ensemble Classifiers. Proceedings of the AAAI Conference on Artificial Intelligence, 34(04), 5478-5486. https://doi.org/10.1609/aaai.v34i04.5998

Issue

Section

AAAI Technical Track: Machine Learning