Publication Type
Journal Article
Version
publishedVersion
Publication Date
11-2023
Abstract
Conversational recommender systems (CRS) endow traditional recommender systems with the capability of dynamically obtaining users’ short-term preferences for items and attributes through interactive dialogues. There are three core challenges for CRS, including the intelligent decisions for what attributes to ask, which items to recommend, and when to askor recommend, at each conversation turn. Previous methods mainly leverage reinforcement learning (RL) to learn conversational recommendation policies for solving one or two of these three decision-making problems in CRS with separated conversation and recommendation components. These approaches restrict the scalability and generality of CRS and fall short of preserving a stable training procedure. In the light of these challenges, we tackle these three decision-making problems in CRS as a unified policy learning task. In order to leverage different features that are important to each sub-problem and facilitate better unified policy learning in CRS, we propose two novel multi-agent RL-based frameworks, namely Independent and Hierarchical Multi-Agent UNIfied COnversational RecommeNders (IMAUNICORNandHMA-UNICORN),respectively. In specific, two low-level agents enrich the state representations for attribute prediction and item recommendation, by combining the long-term user preference information from the historical interaction data and the shortterm user preference information from the conversation history. A high-level meta agent is responsible for coordinating the low-level agents to adaptively make the final decision. Experimental results on four benchmark CRS datasets and a real-world E-Commerce application show that the proposed frameworks significantly outperform state-of-the-art methods. Extensive analyses further demonstrate the superior scalability of the MARL frameworks on the multi-round conversational recommendation.
Keywords
Conversational recommender system, multi-agent reinforcement learning, graph representation learning
Discipline
Databases and Information Systems
Research Areas
Data Science and Engineering
Areas of Excellence
Digital transformation
Publication
IEEE Transactions on Knowledge and Data Engineering
Volume
35
Issue
11
First Page
11541
Last Page
11555
ISSN
1041-4347
Identifier
10.1109/TKDE.2022.3225109
Publisher
Institute of Electrical and Electronics Engineers
Citation
DENG, Yang; LI, Yaliang; DING, Bolin; and LAM, Wai.
Leveraging long short-term user preference in conversational recommendation via multi-agent reinforcement learning. (2023). IEEE Transactions on Knowledge and Data Engineering. 35, (11), 11541-11555.
Available at: https://ink.library.smu.edu.sg/sis_research/9088
Copyright Owner and License
Authors
Creative Commons License
This work is licensed under a Creative Commons Attribution-NonCommercial-No Derivative Works 4.0 International License.
Additional URL
https://doi.org/10.1109/TKDE.2022.3225109