This lecture roughly follows the beginning of Bishop Chapter 3. 
For each input, the agent has a target output. The actual output is compared with the target output, and the agent is altered to bring them closer together.
The job of this algorithm is to determine the structure of the inputs.
The world receives the action of the agent. Under certain circumstances, the agent receives a reward. The agent adjusts its action to maximize the reward.
This is the simplest nontrivial model of a neuron.
If we have several trials (Indexed by which could be time, but almost never is in these cases.), then there are vectors and outputs , as well as desired outputs .
Our Goal:
Find an incremental way of finding the optimal
, that is, the values of space minimizing the error functional:

We let change by , where is precisely described as ``small''.
The coordinate of for the trial is computed via:
So that the update rule is given by (1)
Whenever we analyze one of these methods, we are concerned with the following questions:
When we slect , we have to keep in mind that choosing an that is too small may cause slow convergence, while choosing an too large may cause us to skip over the minimum point (see figure 1.1.3)
We want to choose so that the successive approximations converge to some position in weight space.
If we teach the neuron with several trials, the total error is . This is the average of the errors.
It is important to include several trials, since, if space in some trial, doesn't matter, and we can't find a value for it. 
WidrowHopf LMS was first used in adaptive radar beam forming programs in the late 50's and early 60's.
The perceptron problem includes the following:
Given a putative , we get that whenever . is a line in space with dimension and with normal . One side of this hyperplane consists of all that will yield an on position, and the other side consists of all values that will yield an off position.
Since our trials tell us what values actually yield on and off, choosing the and is a matter of choosing a plane that puts the on and off dots on the correct sides (see figure 1.2.2).
For the sake of simplicity, let for a moment. It is easy to take care of the afterards.
We want change thus:
Let . Let space be set to be the smallest positive value where 
We would like to guarantee that , so that
.
This means that:
.
So that:
If we say .
Where is just barely large enough to force the to change sign. (Remember that the was set to the largest possible value that would not change the sign.)
We still want to know :
we here implement the perceptron learning problem as a batch solution:
Given trials indexed (again) by , we have and . and . For some , , and for some, . In order for each of the trials to match the desired value, we need for all . This is a linear programming problem. Every defines a half space in  coordinates defined by , and the possible 's must lie in the intersection of all these spaces.
This document was generated using the LaTeX2HTML translator Version 99.1 release (March 30, 1999)
Copyright © 1993, 1994, 1995, 1996,
Nikos Drakos,
Computer Based Learning Unit, University of Leeds.
Copyright © 1997, 1998, 1999,
Ross Moore,
Mathematics Department, Macquarie University, Sydney.
The command line arguments were:
latex2html split 1 white 8_28.tex
The translation was initiated by Ben Jones on 20000830