Explainable Reinforcement Learning via a Causal World Model

Explainable Reinforcement Learning via a Causal World Model

Zhongwei Yu, Jingqing Ruan, Dengpeng Xing

Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence
Main Track. Pages 4540-4548. https://doi.org/10.24963/ijcai.2023/505

Generating explanations for reinforcement learning (RL) is challenging as actions may produce long-term effects on the future. In this paper, we develop a novel framework for explainable RL by learning a causal world model without prior knowledge of the causal structure of the environment. The model captures the influence of actions, allowing us to interpret the long-term effects of actions through causal chains, which present how actions influence environmental variables and finally lead to rewards. Different from most explanatory models which suffer from low accuracy, our model remains accurate while improving explainability, making it applicable in model-based learning. As a result, we demonstrate that our causal model can serve as the bridge between explainability and learning.
Keywords:
Machine Learning: ML: Explainable/Interpretable machine learning
Machine Learning: ML: Causality
Machine Learning: ML: Reinforcement learning