From Bench to Bedside: Implementing AI Ethics as Policies for AI Trustworthiness
DOI:
https://doi.org/10.1609/aaaiss.v4i1.31778Abstract
It is well known that successful human-AI collaboration depends on the perceived trustworthiness of the AI. We argue that a key to securing trust in such collaborations is ensuring that the AI competently addresses ethics' foundational role in engagements. Specifically, developers need to identify, address, and implement mechanisms for accommodating ethical components of AI choices. We propose an approach that instantiates ethics semantically as ontology-based moral policies. To accommodate the wide variation and interpretation of ethics, we capture such variations into ethics sets, which are situationally specific aggregations of relevant moral policies. We are extending our ontology-based policy management systems with new representations and capabilities to allow trustworthy AI-human ethical collaborative behavior. Moreover, we believe that such AI-human ethical encounters demand that trustworthiness is bi-directional – humans need to be able to assess and calibrate their actions to be consistent with the trustworthiness of AI in a given context, and AIs need to be able to do the same with respect to humans.Downloads
Published
2024-11-08
How to Cite
Bradshaw, J. M., Bunch, L., Prietula, M., Queen, E., Uszok, A., & Venable, K. B. (2024). From Bench to Bedside: Implementing AI Ethics as Policies for AI Trustworthiness. Proceedings of the AAAI Symposium Series, 4(1), 102–105. https://doi.org/10.1609/aaaiss.v4i1.31778
Issue
Section
AI Trustworthiness and Risk Assessment for Challenging Contexts (ATRACC) - Short Papers