Extended Kalman filter learning algorithm for hyper-complex multilayer neural networks

Author(s):  
H.C.S. Rughooputh ◽  
S.D.D.V. Rughooputh
2004 ◽  
Vol 4 (3) ◽  
pp. 3653-3667 ◽  
Author(s):  
D. J. Lary ◽  
H. Y. Mussa

Abstract. In this study a new extended Kalman filter (EKF) learning algorithm for feed-forward neural networks (FFN) is used. With the EKF approach, the training of the FFN can be seen as state estimation for a non-linear stationary process. The EKF method gives excellent convergence performances provided that there is enough computer core memory and that the machine precision is high. Neural networks are ideally suited to describe the spatial and temporal dependence of tracer-tracer correlations. The neural network performs well even in regions where the correlations are less compact and normally a family of correlation curves would be required. For example, the CH4-N2O correlation can be well described using a neural network trained with the latitude, pressure, time of year, and CH4 volume mixing ratio (v.m.r.). The neural network was able to reproduce the CH4-N2O correlation with a correlation coefficient between simulated and training values of 0.9997. The neural network Fortran code used is available for download.


1994 ◽  
Vol 05 (01) ◽  
pp. 67-75 ◽  
Author(s):  
BYOUNG-TAK ZHANG

Much previous work on training multilayer neural networks has attempted to speed up the backpropagation algorithm using more sophisticated weight modification rules, whereby all the given training examples are used in a random or predetermined sequence. In this paper we investigate an alternative approach in which the learning proceeds on an increasing number of selected training examples, starting with a small training set. We derive a measure of criticality of examples and present an incremental learning algorithm that uses this measure to select a critical subset of given examples for solving the particular task. Our experimental results suggest that the method can significantly improve training speed and generalization performance in many real applications of neural networks. This method can be used in conjunction with other variations of gradient descent algorithms.


2019 ◽  
Vol 52 (1) ◽  
pp. 508-513
Author(s):  
Andressa Apio ◽  
Jônathan W.V. Dambros ◽  
Fabio C. Diehl ◽  
Marcelo Farenzena ◽  
Jorge O. Trierweiler

Sign in / Sign up

Export Citation Format

Share Document