Fundamentals of Statistics contains material of various lectures and courses of H. Lohninger on statistics, data analysis and chemometrics......click here for more. 
Home Multivariate Data Modeling Neural Networks Multilayer Perceptron Back Propagation of Errors  
See also: Multilayer Perceptron  
ANN  Back Propagation of Errors
The basic principles of the back propagation algorithm are:
During the training, the data is presented to the network several thousand times. For each data sample, the current output of the network is calculated and compared to the "true" target value. The error signal δ_{j} of neuron j is computed from the difference between the target and the calculated output. For hidden neurons, this difference is estimated by the weighted error signals of the layer above. The error terms are then used to adjust the weights w_{ij} of the neural network. Thus, the network adjusts its weights after each data sample. This learning process is in fact a gradient descent in the error surface of the weight space  with all its drawbacks. The learning algorithm is slow and prone to getting stuck in a local minimum. For the standard back propagation algorithm, the initial weights of the multilayer perceptron have to be relatively small. They can, for instance, be selected randomly from a small interval around zero. During training they are slowly adapted. Starting with small weights is crucial, because large weights are rigid and cannot be changed quickly. The following interactive example
shows how a multilayer perceptron learns to model data.


Home Multivariate Data Modeling Neural Networks Multilayer Perceptron Back Propagation of Errors 
Last Update: 20121008