Question: Exercise 2: Impurity Functions Let D be a set of examples over a feature space X and a set of classes C = {1,2,43,c}, with

 Exercise 2: Impurity Functions Let D be a set of examples

Exercise 2: Impurity Functions Let D be a set of examples over a feature space X and a set of classes C = {1,2,43,c}, with |D = 24. Consider the following illustration of two possible decision trees, ti and t2 - the colors represent the classes present in each document set associated with the nodes of the trees, the numbers denote how many examples of each class are present. (t) (ty) D D21 D2 C4 (a) First, consider only the first split that each of the two trees makes compute A:({D1,1, D1,2)) and A1({D2,1, D2,2}) with the misclassification rate misclass and the entropy criterion lentropy as splitting criterion. Interpret the results: which of {D1,1, D1,2} or {D2,1, D2.2} is the better first split? (b) If we compare tand tz in terms of their misclassification rate on D, which one is the better decision tree? (C) Assuming the splits shown are the only possibilities, which of t1 or ty would the ID3 algorithm construct, and why

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!