Neural Network Toolbox |
Mean Square Error
Like the perceptron learning rule, the least mean square error (LMS) algorithm is an example of supervised training, in which the learning rule is provided with a set of examples of desired network behavior.
Here is an input to the network, and is the corresponding target output. As each input is applied to the network, the network output is compared to the target. The error is calculated as the difference between the target output and the network output. We want to minimize the average of the sum of these errors.
The LMS algorithm adjusts the weights and biases of the ADALINE so as to minimize this mean square error.
Fortunately, the mean square error performance index for the ADALINE network is a quadratic function. Thus, the performance index will either have one global minimum, a weak minimum, or no minimum, depending on the characteristics of the input vectors. Specifically, the characteristics of the input vectors determine whether or not a unique solution exists.
You can learn more about this topic in Chapter 10 of [HDB96].
Single ADALINE (newlin) | LMS Algorithm (learnwh) |
© 1994-2005 The MathWorks, Inc.