Convergence of an Online Split-Complex Gradient Algorithm for Complex-Valued Neural Networks

The online gradient method has been widely used in training neural networks. We consider in this paper an online split-complex gradient algorithm for complex-valued neural networks. We choose an adaptive learning rate during the training procedure. Under certain conditions, by firstly showing the mo...

Full description

Bibliographic Details
Main Authors: Huisheng Zhang, Dongpo Xu, Zhiping Wang
Format: Article
Language:English
Published: Hindawi Limited 2010-01-01
Series:Discrete Dynamics in Nature and Society
Online Access:http://dx.doi.org/10.1155/2010/829692
Description
Summary:The online gradient method has been widely used in training neural networks. We consider in this paper an online split-complex gradient algorithm for complex-valued neural networks. We choose an adaptive learning rate during the training procedure. Under certain conditions, by firstly showing the monotonicity of the error function, it is proved that the gradient of the error function tends to zero and the weight sequence tends to a fixed point. A numerical example is given to support the theoretical findings.
ISSN:1026-0226
1607-887X