Question: [ 6 pts ] Suppose S = { ( x _ ( i ) , y _ ( i ) ) } _ ( i
pts Suppose SxiyiinsubRdtimes is a linearly separable training dataset. We saw
in that there exists a winRd such that yi:wxi: for all iinn Recall that the Perceptron
algorithm outputs a hyperplane that separates the positive and negative examples ie yi:wxi:
for all iinn
a Devise a new algorithm called MARGINPERCEPTRON that outputs a widehatw that separates the positive
and negative examples by a margin, that is yi:widehatwxi: for all iinn
b Suppose, as in class, that Rmaxix and Bminw: : for all :yi:wxi: Show using
the technique we used in class to show that MARGINPERCEPTRON in at most BR steps.Suppose we modify the Perceptron algorithm as follows: In the update step, instead of performing wt wt yixi, whenever we make a mistake, we instead perform wt wteta yixi, for some eta ; eta is sometimes referred to as the learning rate or the step size. Show that this modified Perceptron will perform the same number of iterations as the original Perceptron we studied in class, and that it will converge to a vector that points in the same direction as the output of the vanilla Perceptron. Hint: What can you say about the relationship between the signs of w x and eta w x
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
