Verification and Validation of AI Systems Using Explanations

Authors

  • Saaduddin Mahmud University of Massachusetts Amherst
  • Sandhya Saisubramanian Oregon State University
  • Shlomo Zilberstein University of Massachusetts Amherst

DOI:

https://doi.org/10.1609/aaaiss.v4i1.31774

Abstract

Verification and validation of AI systems, particularly learning-enabled systems, is hard because often they lack formal specifications and rely instead on incomplete data and human subjective feedback. Aligning the behavior of such systems with the intended objectives and values of human designers and stakeholders is very challenging, and deploying AI systems that are misaligned can be risky. We propose to use both existing and new forms of explanations to improve the verification and validation of AI systems. Toward that goal, we preseant a framework, the agent explains its behavior and a critic signals whether the explanation passes a test. In cases where the explanation fails, the agent offers alternative explanations to gather feedback, which is then used to improve the system's alignment. We discuss examples of this approach that proved to be effective, and how to extend the scope of explanations and minimize human effort involved in this process.

Downloads

Published

2024-11-08

How to Cite

Mahmud, S., Saisubramanian, S., & Zilberstein, S. (2024). Verification and Validation of AI Systems Using Explanations. Proceedings of the AAAI Symposium Series, 4(1), 76-80. https://doi.org/10.1609/aaaiss.v4i1.31774

Issue

Section

AI Trustworthiness and Risk Assessment for Challenging Contexts (ATRACC)