Host: The Japanese Society for Artificial Intelligence
Name : The 102th SIG-SLUD
Number : 102
Location : [in Japanese]
Date : November 28, 2024 - November 29, 2024
Pages 104-108
Conversational recommender systems aim to provide personalized recommendations through interactive conversations with users. A key challenge is to effectively extract and integrate relevant information from both dialogue history and item descriptions for accurate recommendations. Our previous work used a large language model (LLM) to independently generate dialogue summaries and item recommendation descriptions, which were then fed into a score predictor for recommendation. However, this separate processing restricted the model's ability to accurately associate user preferences expressed in the dialogue with relevant item attributes. To address this limitation, we propose a novel approach that uses Direct Preference Optimization (DPO) to fine-tune the LLM. By jointly considering dialogue history and item descriptions during fine-tuning, our method enables the model to generate summaries and recommendation descriptions that are more intricately linked, leading to more effective extraction of user preferences and, ultimately, improved recommendation accuracy in dialogue-based tourist attraction recommendation systems.