A new design formula exploited for accelerating Zhang neural network and its application to time-varying matrix inversion.

Theor. Comput. Sci.(2016)

引用 75|浏览56
暂无评分
摘要
Online solution to time-varying matrix inverse is further investigated by proposing a new design formula, which can accelerate Zhang neural network (ZNN) to finite-time convergence. Compared with the existing recurrent neural networks e.g., the gradient neural networks (GNN), and the original Zhang neural network, the proposed neural network (termed finite-time ZNN, FTZNN) makes a breakthrough in the convergence performance (i.e., from infinite time to finite time). In addition, different from the previous processing method (i.e., choosing a better nonlinear activation to accelerate convergence speed), this paper subtly proposes a new design formula to accelerate the original ZNN model and design the new FTZNN model. Besides, theoretical analyses of the design formula and the FTZNN model are given in detail. Simulative results substantiate the effectiveness and superiorness of the proposed FTZNN model for online time-varying matrix inversion, as compared with the GNN model and the original ZNN model.
更多
查看译文
关键词
Design formula,Time-varying matrix inversion,Zhang neural network,Theoretical analysis,Simulation verification
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要