algorithmmachine-learningperceptron

Intuition for perceptron weight update rule


I am having trouble understanding the weight update rule for perceptrons:

w(t + 1) = w(t) + y(t)x(t).

Assume we have a linearly separable data set.

At iteration t, where t = 0, 1, 2, ...,


Why does this update rule move the boundary in the right direction?


Solution

  • The perceptron's output is the hard limit of the dot product between the instance and the weight. Let's see how this changes after the update. Since

    w(t + 1) = w(t) + y(t)x(t),

    then

    x(t) ⋅ w(t + 1) = x(t) ⋅ w(t) + x(t) ⋅ (y(t) x(t)) = x(t) ⋅ w(t) + y(t) [x(t) ⋅ x(t))].


    Note that:


    How does this move the boundary relative to x(t)?