Teach and Explore: A Multiplex Information-guided Effective and Efficient Reinforcement Learning for Sequential Recommendation

Surong Yan, Chenglong Shi*, Haosen Wang, Lei Chen, Ling Jiang, Ruilin Guo, Kwei Jay Lin*

*此作品的通信作者

研究成果: 期刊稿件文章同行評審

摘要

Casting sequential recommendation (SR) as a reinforcement learning (RL) problem is promising and some RL-based methods have been proposed for SR. However, these models are sub-optimal due to the following limitations: (a) they fail to leverage the supervision signals in the RL training to capture users' explicit preferences, leading to slow convergence; and (b) they do not utilize auxiliary information (e.g., knowledge graph) to avoid blindness when exploring users' potential interests. To address the above-mentioned limitations, we propose a multiplex information-guided RL model (MELOD), which employs a novel RL training framework with Teach and Explore components for SR. We adopt a Teach component to accurately capture users' explicit preferences and speed up RL convergence. Meanwhile, we design a dynamic intent induction network (DIIN) as a policy function to generate diverse predictions. We utilize the DIIN for the Explore component to mine users' potential interests by conducting a sequential and knowledge information joint-guided exploration. Moreover, a sequential and knowledge-aware reward function is designed to achieve stable RL training. These components significantly improve MELOD's performance and convergence against existing RL algorithms to achieve effectiveness and efficiency. Experimental results on seven real-world datasets show that our model significantly outperforms state-of-the-art methods.

原文英語
文章編號120
期刊ACM Transactions on Information Systems
42
發行號5
DOIs
出版狀態已出版 - 29 04 2024

文獻附註

Publisher Copyright:
© 2024 Copyright held by the owner/author(s). Publication rights licensed to ACM.

指紋

深入研究「Teach and Explore: A Multiplex Information-guided Effective and Efficient Reinforcement Learning for Sequential Recommendation」主題。共同形成了獨特的指紋。

引用此