Loading…

Deterministic convergence of complex mini-batch gradient learning algorithm for fully complex-valued neural networks

This paper investigates the fully complex mini-batch gradient algorithm for training complex-valued neural networks. Mini-batch gradient method has been widely used in neural network training, however, its convergence analysis is usually restricted to real-valued neural networks and of probability n...

Full description

Saved in:
Bibliographic Details
Published in:Neurocomputing (Amsterdam) 2020-09, Vol.407, p.185-193
Main Authors: Zhang, Huisheng, Zhang, Ying, Zhu, Shuai, Xu, Dongpo
Format: Article
Language:English
Subjects:
Citations: Items that this one cites
Items that cite this one
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:This paper investigates the fully complex mini-batch gradient algorithm for training complex-valued neural networks. Mini-batch gradient method has been widely used in neural network training, however, its convergence analysis is usually restricted to real-valued neural networks and of probability nature. By introducing a new Taylor mean value theorem for analytic functions, in this paper we establish deterministic convergence results for the fully complex mini-batch gradient algorithm under mild conditions. The deterministic convergence here means that the algorithm will deterministically converge, and both the weak convergence and strong convergence will be proved. Benefited from the newly introduced mean value theorem, our results are of global nature in that they are valid for arbitrarily given initial values of the weights. The theoretical findings are validated with a simulation example.
ISSN:0925-2312
1872-8286
DOI:10.1016/j.neucom.2020.04.114