Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
38th (2024)
Session ID : 1G3-GS-6-04
Conference information

Study of Speech-to-Text Dialogue Model Using Continuous Expressions
*Hyuga NAKAGUROSeiya KAWANOAngel Garcia CONTRERASKoichiro YOSHINO
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

Large language models (LLMs) are flexible and can handle various natural language processing tasks. Many spoken dialogue systems are realized by linking a dialogue model built using an LLM with other modules, such as speech recognition or synthesis systems. However, such a cascaded model with multiple modules is complicated and tends to propagate errors from the previous module. The model can also not consider sensitive expressions in the non-verbal representation of dialogue because the discrete representation, such as texts, is used to connect modules. This research aims to solve these problems by converting the input speech into a vector of continuous expressions and connecting it to a dialogue model. The experimental results show that the generated sentences do not fully take the dialogue context into account, and there is room for improvement, but the natural sentence generation is learned, suggesting that a dialogue model using continuous expressions is feasible.

Content from these authors
© 2024 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top