抄録
We propose an adaptive probability density function (PDF) to select an effective action on reinforcement learning (RL). The uniform distribution function and the normal distribution function of an action are often used to select an action. When these fuctions are used, however, the information of search direction is net considered. The proposed method utilizing the information of it enables RL to reduce the number of trials, which is needed to real environment learning. Furthermore, the proposed method can be applied easily to various methods of RL, for example, actor-critic, stochastic gradient ascent method. The performance of our proposed method is demonstrated by computer simulations.