Construction of Neural Networks for Supervised Learning

碩士 === 國立中山大學 === 電機工程研究所 === 82 === Neural networks may overcome the difficulties related to noise and uncertainty. Conventionally, a trial-and-error method must be used to find the proper neural network architecture for a given problem wh...

Full description

Bibliographic Details
Main Authors: Jone, Mu Tune, 鍾木騰
Other Authors: Lee, Shie Jue
Format: Others
Language:en_US
Published: 1994
Online Access:http://ndltd.ncl.edu.tw/handle/52804905863790355848
Description
Summary:碩士 === 國立中山大學 === 電機工程研究所 === 82 === Neural networks may overcome the difficulties related to noise and uncertainty. Conventionally, a trial-and-error method must be used to find the proper neural network architecture for a given problem when one is using back-propagation algorithms. Also, the conventional method initializes all weights and thresholds to zero in a neural network, likely resulting in a long training time and poor classification accuracy. We propose an idea of constructing neural networks by making use of decision trees and threshold logic. Decision trees are obtained by an ID3 algorithm. Such a tree is represented by a logic expression for constructing a threshold network which forms the basis of the architecture of the desired neural network. The number of layers and the number of nodes in each layer of the neural network are determined. Initial values for weights and thresholds are also determined. Experiments have shown that a neural network constructed in this manner learns fast and performs efficiently. Determining proper neural network architectures is an issue when one is using back-propagation algorithms. We make use of information entropy to overcome this difficulty . Minimal numbers of hidden layers and nodes in each hidden layer are determined by minimizing the information entropy function associated with the partition induced by hyperplanes. A learning process combining the delta rule and a Monte-Carlo based simulated annealing method is developed to obtain the coefficients of optimal hyperplanes. Cios and Liu proposed an entropy based method to generate neural network architectures for two-class discretization. For multi-class discretization, independent two-class subnetworks are combined, resulting in large networks since hidden nodes and layers cannot be shared among different classes. We propose a generation procedure, also based on entropy measure, to produce a network in which hidden nodes and layers are shared among classes.