Loading [MathJax]/extensions/MathMenu.js
Extreme learning machine: a new learning scheme of feedforward neural networks | IEEE Conference Publication | IEEE Xplore

Extreme learning machine: a new learning scheme of feedforward neural networks


Abstract:

It is clear that the learning speed of feedforward neural networks is in general far slower than required and it has been a major bottleneck in their applications for pas...Show More

Abstract:

It is clear that the learning speed of feedforward neural networks is in general far slower than required and it has been a major bottleneck in their applications for past decades. Two key reasons behind may be: 1) the slow gradient-based learning algorithms are extensively used to train neural networks, and 2) all the parameters of the networks are tuned iteratively by using such learning algorithms. Unlike these traditional implementations, this paper proposes a new learning algorithm called extreme learning machine (ELM) for single-hidden layer feedforward neural networks (SLFNs) which randomly chooses the input weights and analytically determines the output weights of SLFNs. In theory, this algorithm tends to provide the best generalization performance at extremely fast learning speed. The experimental results based on real-world benchmarking function approximation and classification problems including large complex applications show that the new algorithm can produce best generalization performance in some cases and can learn much faster than traditional popular learning algorithms for feedforward neural networks.
Date of Conference: 25-29 July 2004
Date Added to IEEE Xplore: 17 January 2005
Print ISBN:0-7803-8359-1
Print ISSN: 1098-7576
Conference Location: Budapest, Hungary
References is not available for this document.

I. Introduction

From a mathematical point of view, research on the approximation capabilities of feedforward neural networks has focused on two aspects: universal approximation on compact input sets and approximation in a finite set. Many researchers have explored the universal approximation capabilities of standard multi-layer feedforward neural networks[1], [2], [3]. In real applications, the neural networks are trained in finite training set. For function approximation in a finite training set, Huang and Babri[4] shows that a sigle-hidden layer feedforward neural network (SLFN) with at most hidden neurons and with almost any nonlinear activation function can learn distinct observations with zero error. It should be noted that the input weights (linking the input layer to the first hidden layer) need to be adjusted in all these previous theoretical research works as well as in almost all practical learning algorithms of feedforward neural networks.

Select All
1.
K. Hornik, "Approximation capabilities of multilayer feedforward networks," Neural Networks, vol. 4, pp. 251-257, 1991.
2.
M. Leshno, V. Y. Lin, A. Pinkus, and S. Schocken, "Multilayer feedforward networks with a nonpolynomial activation function can approximate any function," Neural Networks, vol. 6, pp. 861-867, 1993.
3.
Y. Ito, "Approximation of continuous functions on R by linear combinations of shifted rotations of a sigmoid function with and without scaling," Neural Networks, vol. 5, pp. 105-115, 1992.
4.
G.-B. Huang and H. A. Babri, "Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions," IEEE Transactions on Neural Networks, vol. 9, no. 1, pp. 224-229, 1998.
5.
S. Tamura and M. Tateishi, "Capabilities of a four-layered feedforward neural network: Four layers versus three," IEEE Transactions on Neural Networks, vol. 8, no. 2, pp. 251-255, 1997.
6.
G.-B. Huang, "Learning capability and storage capacity of two-hidden-layer feedforward networks," IEEE Transactions on Neural Networks, vol. 14, no. 2, pp. 274-281, 2003.
7.
G.-B. Huang, Q.-Y. Zhu, and C.-K. Siew, "Real-time learning capability of neural networks," in Technical Report ICIS/45/2003, (School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore), Apr. 2003.
8.
G.-B. Huang, L. Chen, and C.-K. Siew, "Universal approximation using incremental feedforward networks with arbitrary input weights," in Technical Report ICIS/46/2003, (School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore), Oct. 2003.
9.
P. L. Bartlett, "The sample complexity of pattern classification with neural networks: The size of the weights is more important than the size of the network," IEEE Transactions on Information Theory, vol. 44, no. 2, pp. 525-536, 1998.
10.
D. Serre, "Matrices: Theory and applications," Springer-Verlag New York, Inc, 2002.
11.
S. Haykin, "Neural networks: A comprehensive foundation," New Jersey: Prentice Hall, 1999.
12.
P. L. Bartlett, "For valid generalization, the size of the weights is more important than the size of the network," in Advances in Neural Information Processing Systems'1996 (M. Mozer, M. Jordan, and T. Petsche, eds.), vol. 9, pp. 134-140, MIT Press, 1997.
13.
C.-C. Chang and C.-J. Lin, "LIBSVM - a library for support vector machines," in http://www.csie.ntu.edu.tw/~cjlin/libsvm/, Deptartment of Computer Science and Information Engineering, National Taiwan University, Taiwan, 2003.
14.
G. Rätsch, T. Onoda, and K. R. Müller, "An improvement of AdaBoost to avoid overfitting," in Proceedings of the 5th International Conference on Neural Information Processing (ICONIP'1998), 1998.
15.
Y. Freund and R. E. Schapire, "Experiments with a new boosting algorithm," in International Conference on Machine Learning, pp. 148-156, 1996.
16.
D. R. Wilson and T. R. Martinez, "Heterogeneous radial basis function networks," in Proceedings of the International Conference on Neural Networks (ICNN 96), pp. 1263-1267, June 1996.
17.
R. Collobert, S. Bengio, and Y. Bengio, "A parallel mixtures of SVMs for very large scale problems," Neural Computation, vol. 14, pp. 1105-1114, 2002.
18.
G.-B. Huang, Y.-Q. Chen, and H. A. Babri, "Classification ability of single hidden layer feedforward neural networks," IEEE Transactions on Neural Networks, vol. 11, no. 3, pp. 799-801, 2000.
19.
G.-B. Huang, Q.-Y. Zhu, and C.-K. Siew, "Extreme learning machine," in Technical Report ICIS/03/2004, (School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore), Jan. 2004.
Contact IEEE to Subscribe

References

References is not available for this document.