Suppose you are running a learning experiment on a new algorithm. You have a data set consisting

Question:

Suppose you are running a learning experiment on a new algorithm. You have a data set consisting of 2 examples of each of two classes. Yon plan to use leave-one-nut cross-validation. As a baseline, you run your experimental setup on a simple majority classifier. (A majority classifier is given a set of training data and then always outputs the class that is in the majority in the training set, regardless of the input.) You expect the majority classifier to score about 50% on leave-one-out cross-validation, but to your surprise, it scores zero. Can you explain why?

Fantastic news! We've Found the answer you've been seeking!

Step by Step Answer:

Related Book For  book-img-for-question

Artificial Intelligence A Modern Approach

ISBN: 978-0137903955

2nd Edition

Authors: Stuart J. Russell and Peter Norvig

Question Posted: