Using ontologies to enhance human understandability of global post-hoc explanations of black-box models

作者:

Highlights:

• We propose Trepan Reloaded for explaining black-box models with decision trees.

• Trepan Reloaded extends Trepan using ontologies to guide the decision tree generation.

• Improved human understanding of generated explanations is confirmed in a user study.

• We have published (anonymised) data and code to make our study fully reproducible.

摘要

The interest in explainable artificial intelligence has grown strongly in recent years because of the need to convey safety and trust in the ‘how’ and ‘why’ of automated decision-making to users. While a plethora of approaches has been developed, only a few focus on how to use domain knowledge and how this influences the understanding of explanations by users. In this paper, we show that by using ontologies we can improve the human understandability of global post-hoc explanations, presented in the form of decision trees. In particular, we introduce Trepan Reloaded, which builds on Trepan, an algorithm that extracts surrogate decision trees from black-box models. Trepan Reloaded includes ontologies, that model domain knowledge, in the process of extracting explanations to improve their understandability. We tested the understandability of the extracted explanations by humans in a user study with four different tasks. We evaluate the results in terms of response times and correctness, subjective ease of understanding and confidence, and similarity of free text responses. The results show that decision trees generated with Trepan Reloaded, taking into account domain knowledge, are significantly more understandable throughout than those generated by standard Trepan. The enhanced understandability of post-hoc explanations is achieved with little compromise on the accuracy with which the surrogate decision trees replicate the behaviour of the original neural network models.

论文关键词:Human-understandable explainable AI,Global explanations,Ontologies,Neural-symbolic learning and reasoning,Knowledge extraction,Concept refinement

论文评审过程:Received 23 April 2020, Revised 24 December 2020, Accepted 8 February 2021, Available online 15 February 2021, Version of Record 22 February 2021.

论文官网地址:https://doi.org/10.1016/j.artint.2021.103471