Journal of the Japanese Society for Artificial Intelligence
Online ISSN : 2435-8614
Print ISSN : 2188-2266
Print ISSN:0912-8085 until 2013
Acquisition of Optimal Internal Representation in Neural Networks by Using Information Criteria
Takashi ONODA
Author information
MAGAZINE FREE ACCESS

1998 Volume 13 Issue 3 Pages 434-443

Details
Abstract

Engineers are often confronted with the problem of extracting information about poorly-known processes from data. Discerning the significant patterns in data, as a first step to process understanding, can be greatly facilitated by reducing dimensionality. An artificial neural network can develop a compact representation of the input data. This neural network can be applied discerning the significant patterns in data. The neural network, which can reduce dimensionality and produce a feature space map, contains an internal "bottleneck"layer (containing fewer units than input or output layers) and two additional hidden layers. In the case of this type neural network, the inputs of the network are reproduced at the output layer. An important problem is to determine the optimal neural network architecture to acquire the optimal feature space map. This paper proposes that information criteria are applied to determine the optimal neural network architecture and presents that Neural Network Information Criterion (NNIC) is the most useful for acquisition of the feature space map with a simple simulation. The neural network, which is selected by using NNIC, can adequately reduce dimensionality and produce the optimal feature space map resembling the actual distribution of the underlying system parameters.

Content from these authors
© 1998 The Japaense Society for Artificial Intelligence
Previous article Next article
feedback
Top