Nonlinear Theory and Its Applications, IEICE
Online ISSN : 2185-4106
ISSN-L : 2185-4106
Special Section on Recent Progress in Nonlinear Theory and Its Applications
Hardware-oriented deep reinforcement learning for edge computing
Yoshiharu YamagishiTatsuya KanekoMegumi Akai-KasayaTetsuya Asai
著者情報
ジャーナル フリー

2021 年 12 巻 3 号 p. 526-544

詳細
抄録

A new deep reinforcement learning enhancement is proposed for edge computing. This work focuses on deep Q-networks (DQNs), which are used in deep reinforcement learning. Although DQNs are typically improved through a software-based approach, hardware-specific knowledge such as that on data paths and pipelines is used for improving a DQN. The DQN performance is improved and the number of resources are reduced through an efficient hardware design that considers the learning flow and parameter search. As the scale of the problem increases, the amount of reduction in the use of resources also increases. For example, when the size of the block catch game is 5 × 10, the memory requirement is reduced by approximately 50% compared to a previous DQN. The proposed hardware-oriented approach can be applied to any software technology. This study facilitates the development of novel technologies that can be realized through edge computing.

著者関連情報
© 2021 The Institute of Electronics, Information and Communication Engineers
前の記事 次の記事
feedback
Top