Host: The Japanese Society for Artificial Intelligence
Name : The 38th Annual Conference of the Japanese Society for Artificial Intelligence
Number : 38
Location : [in Japanese]
Date : May 28, 2024 - May 31, 2024
For more environmentally sustainable development of deep learning (DL) technologies, computational burden for tuning DL architectures should be reduced. This calls for more systematic strategies for finding an optimal set of hyperparameters which achieves a good balance between training speed and generalization performance. As a preliminary step toward this goal, we address the problem of how to tune fully-connected feedforward perceptrons in the so-called ``kernel regime'' in a systematic manner. By combining the existing theoretical tools, such as the Neural Tangent Kernel (NTK), and the analogy of the signal propagation dynamics with absorbing phase transitions, we conduct thorough analysis of the training dynamics of the neural network, including the case with finite depth. As a result, a simple strategy for optimally tuning the initialization hyperparameters and the depth is proposed.