Transactions of the Japanese Society for Artificial Intelligence
Online ISSN : 1346-8030
Print ISSN : 1346-0714
ISSN-L : 1346-0714
Original Paper
Initialization of CNN Models for Training on a Small Dataset Using Importance of Filter Parameters
Jien KatoGuanwen ZhangYu Wang
Author information
JOURNAL FREE ACCESS

2017 Volume 32 Issue 3 Pages C-G42_1-11

Details
Abstract

Deep Convolutional Neural Networks (CNNs) have achieved great success in many computer vision tasks. However, it is still difficult to use them in practical tasks, especially small scale tasks, because of the large quantity of labeled training data that are required in their training process. In this paper, we present two approaches to enable easy adaption of CNNs in small scale tasks: the Minimum Entropy Loss (MEL) approach and the Minimum Reconstruction Error (MRE) approach. The basic idea of these two approaches is to select informative filters in pre-trained CNN models, and reuse them to initialize CNNs that are designed for small scale tasks. Different with popular fine-turning approach which also reuses pre-trained CNNs by conducting further training without changing their model architectures, MEL and MRE lead to an easy usage of pre-trained models in novel model architectures. This makes it a high flexibility when dealing with small scale tasks. We evaluated the performance of the two approaches on practicalsmall scale tasks, and confirmed their high performance and high flexibility.

Content from these authors
© The Japanese Society for Artificial Intelligence 2017
Previous article Next article
feedback
Top