State representation modeling for deep reinforcement learning based recommendation

作者:

Highlights:

摘要

Reinforcement learning techniques have recently been introduced to interactive recommender systems to capture the dynamic patterns of user behavior during the interaction with recommender systems and perform planning to optimize long-term performance. Most existing research work focuses on designing policy and learning algorithms of the recommender agent but seldom cares about the state representation of the environment, which is indeed essential for the recommendation decision making. In this paper, we first formulate the interactive recommender system problem with a deep reinforcement learning recommendation framework. Within this framework, we then carefully design four state representation schemes for learning the recommendation policy. Inspired by recent advances in feature interaction modeling in user response prediction, we discover that explicitly modeling user–item interactions in state representation can largely help the recommendation policy perform effective reinforcement learning. Extensive experiments on four real-world datasets are conducted under both the offline and simulated online evaluation settings. The experimental results demonstrate the proposed state representation schemes lead to better performance over the state-of-the-art methods.

论文关键词:State representation modeling,Deep reinforcement learning,Recommendation

论文评审过程:Received 15 March 2019, Revised 17 June 2020, Accepted 19 June 2020, Available online 4 July 2020, Version of Record 30 July 2020.

论文官网地址:https://doi.org/10.1016/j.knosys.2020.106170