I am having trouble understanding the weight update rule for perceptrons:
w(t + 1) = w(t) + y(t)x(t).
Assume we have a linearly separable data set.
At iteration t, where t = 0, 1, 2, ...,
Why does this update rule move the boundary in the right direction?
The perceptron's output is the hard limit of the dot product between the instance and the weight. Let's see how this changes after the update. Since
w(t + 1) = w(t) + y(t)x(t),
then
x(t) ⋅ w(t + 1) = x(t) ⋅ w(t) + x(t) ⋅ (y(t) x(t)) = x(t) ⋅ w(t) + y(t) [x(t) ⋅ x(t))].
Note that:
How does this move the boundary relative to x(t)?