Using state abstractions to compute personalized contrastive explanations for AI agent behavior

作者:

摘要

There is a growing interest within the AI research community in developing autonomous systems capable of explaining their behavior to users. However, the problem of computing explanations for users of different levels of expertise has received little research attention. We propose an approach for addressing this problem by representing the user's understanding of the task as an abstraction of the domain model that the planner uses. We present algorithms for generating minimal explanations in cases where this abstract human model is not known. We reduce the problem of generating an explanation to a search over the space of abstract models and show that while the complete problem is NP-hard, a greedy algorithm can provide good approximations of the optimal solution. We empirically show that our approach can efficiently compute explanations for a variety of problems and also perform user studies to test the utility of state abstractions in explanations.

论文关键词:Explanations for plans,Abstractions,Contrastive explanations

论文评审过程:Received 9 March 2020, Revised 14 July 2021, Accepted 19 July 2021, Available online 30 July 2021, Version of Record 17 August 2021.

论文官网地址:https://doi.org/10.1016/j.artint.2021.103570