Why Bad Coffee? Explaining BDI Agent Behaviour with Valuings (Extended Abstract)

Why Bad Coffee? Explaining BDI Agent Behaviour with Valuings (Extended Abstract)

Michael Winikoff, Galina Sidorenko, Virginia Dignum, Frank Dignum

Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence
Journal Track. Pages 5782-5786. https://doi.org/10.24963/ijcai.2022/810

An important issue in deploying an autonomous system is how to enable human users and stakeholders to develop an appropriate level of trust in the system. It has been argued that a crucial mechanism to enable appropriate trust is the ability of a system to explain its behaviour. Obviously, such explanations need to be comprehensible to humans. Due to the perceived similarity in functioning between humans and autonomous systems, we argue that it makes sense to build on the results of extensive research in social sciences that explores how humans explain their behaviour. Using similar concepts for explanation is argued to help with comprehensibility, since the concepts are familiar. Following work in the social sciences, we propose the use of a folk-psychological model that utilises beliefs, desires, and ``valuings''. We propose a formal framework for constructing explanations of the behaviour of an autonomous system, present an (implemented) algorithm for giving explanations, and present evaluation results.
Keywords:
AI Ethics, Trust, Fairness: Explainability and Interpretability
Agent-based and Multi-agent Systems: Agent Theories and Models
AI Ethics, Trust, Fairness: Trustworthy AI
AI Ethics, Trust, Fairness: Values
AI Ethics, Trust, Fairness: General