A Symbolic Approach to Explaining Bayesian Network Classifiers

A Symbolic Approach to Explaining Bayesian Network Classifiers

Andy Shih, Arthur Choi, Adnan Darwiche

Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence
Main track. Pages 5103-5111. https://doi.org/10.24963/ijcai.2018/708

We propose an approach for explaining Bayesian network classifiers, which is based on compiling such classifiers into decision functions that have a tractable and symbolic form. We introduce two types of explanations for why a classifier may have classified an instance positively or negatively and suggest algorithms for computing these explanations. The first type of explanation identifies a minimal set of the currently active features that is responsible for the current classification, while the second type of explanation identifies a minimal set of features whose current state (active or not) is sufficient for the classification. We consider in particular the compilation of Naive and Latent-Tree Bayesian network classifiers into Ordered Decision Diagrams (ODDs), providing a context for evaluating our proposal using case studies and experiments based on classifiers from the literature.
Keywords:
Knowledge Representation and Reasoning: Diagnosis and Abductive Reasoning
Uncertainty in AI: Bayesian Networks
Machine Learning: Interpretability
Knowledge Representation and Reasoning: Tractable Languages and Knowledge compilation