Question: 4. (Classification 10 pt)) Suppose you are working in a binary classification problem so C t0, 1u say (think of 0 is spam and

4. (Classification 10 pt)) Suppose you are working in a binary classification problem so C " t0, 1u say (think of 0 is spam and 1 as not-spam). Now suppose instead of the 0 1 loss function your loss function is: 3 lp?pxq, yq " $ '& '% 100 if ?pxq " 0, y " 1 1 if ?pxq " 1, y " 0. For any classification rule ? let Rp?q " Eplp?pXq, Y qq denote the risk of the classifier i.e. the average loss made by the classifier on a new "typical" data point pX, Y q. By repeating the calculations we did in class when proving the optimality of the Bayes classifier for 0-1 loss, find the optimal classifier for the above loss function (i.e. find an expression in terms of the conditional probability mass function of Y given X " x).

??

4. (Classification 10 pt)) Suppose you are working in a binary classification

4. (Classification 10 pt) (you might be able to do this only after Thursday's class) Suppose you are working in a binary classification problem so C = {0, 1} say (think of 0 is spam and 1 as not-spam). Now suppose instead of the 0 - 1 loss function your loss function is: CO 1($(x) , y ) = 100 if p(x) = 0, y = 1 if o(x) = 1, y = 0. For any classification rule o let R() = E(1(6(X), Y) ) denote the risk of the classifier i.e. the average loss made by the classifier on a new "typical" data point (X, Y). By repeating the calculations we did in class when proving the optimality of the Bayes classifier for 0-1 loss, find the optimal classifier for the above loss function (i.e. find an expression in terms of the conditional probability mass function of Y given X = x)

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Mathematics Questions!