@brazovayeye

Genetic Programming Discovers Efficient Learning Rules for the Hidden and Output Layers of Feedforward Neural Networks

, and . Genetic Programming, Proceedings of EuroGP'99, volume 1598 of LNCS, page 120--134. Goteborg, Sweden, Springer-Verlag, (26-27 May 1999)

Abstract

The learning method is critical for obtaining good generalisation in neural networks with limited training data. The Standard BackPropagation (SBP ) training algorithm suffers from several problems such as sensitivity to the initial conditions and very slow convergence. The aim of this work is to use Genetic Programming (GP) to discover new supervised learning algorithms which can overcome some of these problems. In previous research a new learning algorithm for the output layer has been discovered using GP. By comparing this with SBP on different problems better performance was demonstrated. This paper shows that GP can also discover better learning algorithms for the hidden layers to be used in conjunction with the algorithm previously discovered. Comparing these with SBP on different problems we show they provide better performance. This study indicates that there exist many supervised learning algorithms better than SBP and that GP can be used to discover them.

Links and resources

Tags