Question: Suppose you are running a learning experiment on a new algorithm. You have a data set consisting of 2 examples of each of two classes.
Suppose you are running a learning experiment on a new algorithm. You have a data set consisting of 2 examples of each of two classes. Yon plan to use leave-one-nut cross-validation. As a baseline, you run your experimental setup on a simple majority classifier. (A majority classifier is given a set of training data and then always outputs the class that is in the majority in the training set, regardless of the input.) You expect the majority classifier to score about 50% on leave-one-out cross-validation, but to your surprise, it scores zero. Can you explain why?
Step by Step Solution
3.47 Rating (177 Votes )
There are 3 Steps involved in it
If we leave out an example of one ... View full answer
Get step-by-step solutions from verified subject matter experts
Document Format (1 attachment)
21-C-S-A-I (265).docx
120 KBs Word File
