Proceedings of the Annual Conference of JSAI
Online ISSN : 2758-7347
34th (2020)
Session ID : 3Q1-GS-9-02
Conference information

Dialogue over context and structured knowledge using a neural network model with an external memory
*Yuri MURAYAMAIchiro KOBAYASHI
Author information
CONFERENCE PROCEEDINGS FREE ACCESS

Details
Abstract

In recent years, sequence to sequence models such as Seq2Seq and Transformer have been commonplace for dialogue architectures. On the other hand, it is necessary for more natural and intellectual dialogue to understand context and use knowledge. However, scientists have argued that such models are limited in their ability to store data over a long time. To retain the long-term information, neural network models with external memories such as End-To-End Memory Networks and Differentiable Neural Computer (DNC) have been proposed. In this work, we extend DNC architectures and propose a model using both context and structured knowledge. We conducted an experiment on a dataset which is composed of a series of coherently linked questions that require a large scale knowledge graph and their answers. The mean test error rate was 69.25% after 20k iterations and a little higher than the original DNC's error rate 69.09%.

Content from these authors
© 2020 The Japanese Society for Artificial Intelligence
Previous article Next article
feedback
Top