Computer Science and Information Systems 2012 Volume 9, Issue 4, Pages: 1603-1625
https://doi.org/10.2298/CSIS120121043Z
Full text ( 235 KB)
Cited by
Superior performance of using hyperbolic sine activation functions in ZNN illustrated via time-varying matrix square roots finding
Zhang Yunong (School of Information Science and Technology, Sun Yat-sen University Guangzhou, Guangdong, China)
Jin Long (School of Information Science and Technology, Sun Yat-sen University Guangzhou, Guangdong, China)
Ke Zhende (School of Information Science and Technology, Sun Yat-sen University Guangzhou, Guangdong, China)
A special class of recurrent neural network (RNN), termed Zhang neural
network (ZNN) depicted in the implicit dynamics, has recently been proposed
for online solution of time-varying matrix square roots. Such a ZNN model can
be constructed by using monotonically-increasing odd activation functions to
obtain the theoretical time-varying matrix square roots in an error-free
manner. Different choices of activation function arrays may lead to different
performance of the ZNN model. Generally speaking, ZNN model using hyperbolic
sine activation functions may achieve better performance, as compared with
those using other activation functions. In this paper, to pursue the superior
convergence and robustness properties, hyperbolic sine activation functions
are applied to the ZNN model for online solution of time-varying matrix
square roots. Theoretical analysis and computer-simulation results further
demonstrate the superior performance of the ZNN model using hyperbolic sine
activation functions in the context of large model-implementation errors, in
comparison with that using linear activation functions.
Keywords: Zhang neural network, global exponential convergence, hyperbolic sine activation functions, time-varying matrix square roots, implementation errors