A survey of inverse reinforcement learning
作者:Stephen Adams, Tyler Cody, Peter A. Beling
摘要
Learning from demonstration, or imitation learning, is the process of learning to act in an environment from examples provided by a teacher. Inverse reinforcement learning (IRL) is a specific form of learning from demonstration that attempts to estimate the reward function of a Markov decision process from examples provided by the teacher. The reward function is often considered the most succinct description of a task. In simple applications, the reward function may be known or easily derived from properties of the system and hard coded into the learning process. However, in complex applications, this may not be possible, and it may be easier to learn the reward function by observing the actions of the teacher. This paper provides a comprehensive survey of the literature on IRL. This survey outlines the differences between IRL and two similar methods - apprenticeship learning and inverse optimal control. Further, this survey organizes the IRL literature based on the principal method, describes applications of IRL algorithms, and provides areas of future research.
论文关键词:Reinforcement learning, Inverse reinforcement learning, Inverse optimal control, Apprenticeship learning, Learning from demonstration
论文评审过程:
论文官网地址:https://doi.org/10.1007/s10462-021-10108-x