JSAI Technical Report, SIG-SLUD
Online ISSN : 2436-4576
Print ISSN : 0918-5682
102nd (Nov.2024)
Conference information

Improving Information Extraction Performance Using DPO(Direct Preference Optimization) for Dialogue-Based Tourist Spot Recommendation
Manato TAJIRIMichimasa INABA
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Pages 104-108

Details
Abstract

Conversational recommender systems aim to provide personalized recommendations through interactive conversations with users. A key challenge is to effectively extract and integrate relevant information from both dialogue history and item descriptions for accurate recommendations. Our previous work used a large language model (LLM) to independently generate dialogue summaries and item recommendation descriptions, which were then fed into a score predictor for recommendation. However, this separate processing restricted the model's ability to accurately associate user preferences expressed in the dialogue with relevant item attributes. To address this limitation, we propose a novel approach that uses Direct Preference Optimization (DPO) to fine-tune the LLM. By jointly considering dialogue history and item descriptions during fine-tuning, our method enables the model to generate summaries and recommendation descriptions that are more intricately linked, leading to more effective extraction of user preferences and, ultimately, improved recommendation accuracy in dialogue-based tourist attraction recommendation systems.

Content from these authors
© 2024 The Japaense Society for Artificial Intelligence
Previous article Next article
feedback
Top