Generating Artificial Neural Networks for Value Function Approximation in a Domain Requiring a Shifting Strategy

Author(s):  
Ransom K. Winder
2006 ◽  
Vol 17 (4) ◽  
pp. 942-952 ◽  
Author(s):  
D. Wedge ◽  
D. Ingram ◽  
D. Mclean ◽  
C. Mingham ◽  
Z. Bandar

Feed-forward artificial neural networks are universal approximators of continuous functions. This property enables the use of these networks to solve learning tasks. Learning tasks in this paradigm are cast as function approximation problems. The universal approximation results for these networks require at least one hidden layer with non-linear nodes, and also require that the non-linearities be non-polynomial in nature. In this paper a non-polynomial and non-sigmoidal non-linear function is proposed as a suitable activation function for these networks. The usefulness of the proposed activation function is shown on 12 function approximation task. The obtained results demonstrate that the proposed activation function outperforms the logistic / log-sigmoid and the hyperbolic tangent activation functions.


Sign in / Sign up

Export Citation Format

Share Document