人工知能学会全国大会論文集
Online ISSN : 2758-7347
37th (2023)
セッションID: 2U4-IS-2c-05
会議情報

Rapid training of Perceiver in a low-cost computing environment
*Kaiyu SUZUKITomofumi MATSUZAWA
著者情報
会議録・要旨集 フリー

詳細
抄録

Perceiver is a deep learning model that can be applied to a variety of modalities. It can simultaneously process various forms of input and output, such as images, speech, and natural language using the same architecture. However, Perceiver is computationally more expensive than other models. Therefore, training the model in environments with relatively limited fast parallel computational resources is relatively difficult. In this study, we aimed to reduce the computational cost such that learning can be performed in a short time in environments other than large-scale computing systems. To this end, we first show that a speed-up method proposed for Transformer is also effective for Perceiver. In particular, the gated attention unit proposed for FLASH reduces computational complexity without sacrificing accuracy. The proposed acceleration method can achieve accuracy comparable to that of the original model in a limited computing environment. As an introductory example, we conducted experiments using the ImageNet image recognition task and demonstrated that the proposed method can reduce the training time compared to conventional methods without a significant loss of accuracy. This model can be used to input and output any kind of data quickly in a low-cost computing environment.

著者関連情報
© 2023 The Japanese Society for Artificial Intelligence
前の記事 次の記事
feedback
Top