Note that there are some explanatory texts on larger screens.

plurals
  1. PONeural Networks: Why does the perceptron rule only work for linearly separable data?
    text
    copied!<p>I previously <a href="https://stackoverflow.com/questions/13976565/neural-networks-what-does-linearly-separable-mean">asked</a> for an explanation of linearly separable data. Still reading Mitchell's Machine Learning book, I have some trouble understanding why exactly the perceptron rule only works for linearly separable data?</p> <p>Mitchell defines a perceptron as follows: <img src="https://i.stack.imgur.com/CxurJ.png" alt="The perceptron"></p> <p>That is, it is y is 1 or -1 if the sum of the weighted inputs exceeds some threshold. </p> <p>Now, the problem is to determine a weight vector that causes the perceptron to produce the correct output (1 or -1) for each of the given training examples. One way of achieving this is through the perceptron rule:</p> <blockquote> <p>One way to learn an acceptable weight vector is to begin with random weights, then iteratively apply the perceptron to each training example, modify- ing the perceptron weights whenever it misclassifies an example. This process is repeated, iterating through the training examples as many times as needed until the perceptron classifies all training examples correctly. Weights are modified at each step according to the perceptron training rule, which revises the weight wi associated with input xi according to the rule: <img src="https://i.stack.imgur.com/D9LN4.png" alt="The perceptron rule"></p> </blockquote> <p>So, my question is: Why does this only work with linearly separable data? Thanks.</p>
 

Querying!

 
Guidance

SQuiL has stopped working due to an internal error.

If you are curious you may find further information in the browser console, which is accessible through the devtools (F12).

Reload