While Bayesian network (BN) can achieve accurate predictions even with erroneous or incomplete evidence, explaining the inferences remains a challenge. Existing approaches fall short because they do not exploit variable interactions and cannot account for compensations during inferences. This paper proposes the Explaining BN Inferences (EBI) procedure for explaining how variables interact to reach conclusions. EBI explains the value of a target node in terms of the influential nodes in the target's Markov blanket under specific contexts, where the Markov nodes include the target's parents, children, and the children's other parents. Working back from the target node, EBI shows the derivation of each intermediate variable, and finally explains how missing and erroneous evidence values are compensated. We validated EBI on a variety of problem domains, including mushroom classification, water purification and web page recommendation. The experiments show that EBI generates high quality, concise and comprehensible explanations for BN inferences, in particular the underlying compensation mechanism that enables BN to outperform alternative prediction systems, such as decision tree.
Bayesian networks, Explanations, Inferences, Compensations, Error values, Missing values
Databases and Information Systems | Numerical Analysis and Scientific Computing
Data Management and Analytics
YAP, Ghim-Eng; TAN, Ah-Hwee; and PANG, Hwee Hwa.
Explaining Inferences in Bayesian Networks. (2007). Applied Intelligence. 29, (3), 263-278. Research Collection School Of Information Systems.
Available at: http://ink.library.smu.edu.sg/sis_research/1247
Copyright Owner and License
Creative Commons License
This work is licensed under a Creative Commons Attribution-Noncommercial-No Derivative Works 4.0 License.