The Proceedings of JSME annual Conference on Robotics and Mechatronics (Robomec)
Online ISSN : 2424-3124
2022
Session ID : 2A2-H11
Conference information

Improvement of interpretability and noise robustness of deep predictive learning by modality attention
- Joint Research and Development of Hitachi, Ltd. and Waseda University -
*Hideyuki ICHIWARAHiroshi ITOKenjiro YAMAMOTOHiroki MORITetsuya OGATA
Author information
CONFERENCE PROCEEDINGS RESTRICTED ACCESS

Details
Abstract

In order for robots to perform tasks that humans perform, it is necessary to process multimodal information such as vision and force, just like humans. In this study, we propose modality attention by deep predictive learning that can interpret which modal information is used during the task. A hierarchical model consisting of low-level NNs(Neural Networks) that process each modal information individually and a high-level NN that integrates the modal information is used. Furthermore, by weighting each modal information input to the upper NN with learnable weights and inputting it, the modal information used for motion generation is self-adjustable. We verified the effectiveness of the proposed method in the task of inserting furniture parts that require vision and force. It was confirmed that the modality that attracts attention transitions appropriately, and that stable motion can be generated even if noise occurs in the modality that does not pay attention.

Content from these authors
© 2022 The Japan Society of Mechanical Engineers
Previous article Next article
feedback
Top