1. Introduction
Feed-forward neural networks using random weights were first suggested by Schmidt et al. in 1992 [1] but were not a widely used method until Huang et al. popularised them as Extreme Learning Machines (ELM) [2], [3] in 2005. The ELM is a multi-layer feed-forward neural network topology and algorithm that offers fast training and flexible nonlinearity for function regression and classification tasks. Its principal benefit is that the network parameters are calculated in a single pass during the training process, which offers a significant improvement in implementation time over conventional back-propagation-trained feed-forward net-works. In its standard form it has an input layer that is fully connected to a hidden layer with conventional non-linear activation functions. The hidden layer is fully connected to an output layer with linear activation functions. The number of hidden units is often much greater than the input layer with a fan-out of 5 to 20 hidden units per input element frequently used. A key feature of ELMs is that the weights connecting the input layer to the hidden layer are set to random values, usually uniformly distributed in some predefined range. This simplifies the requirements for training to one of determining the hidden to output unit weights, which can be achieved in a single pass. By randomly projecting the inputs to a much higher dimensionality, it is possible for the algorithm to find a hyperplane which approximates a desired regression function, or represents a linear separable classification problem [4].