Question: Question for Computer Datamining Question 1. 5 marks] The following shows a history of users with attributes CS Major (i.e., majored in computer science), Age,

Question for Computer Datamining

Question for Computer Datamining Question 1. 5 marks] The following shows a

Question 1. 5 marks] The following shows a history of users with attributes "CS Major" (i.e., majored in computer science), "Age", and "Income". We also indicate whether they will buy Bitcoin or not in the last column No. CSMajorAgeIncome BuyBitcoin fair middlefair fair young high old es es no no yes es no no ves ves ves yes no no no no 3 young old young youngfair middlelow low low We want to train a C4.5 decision tree classifier to predict whether a new user will buy Bitcoin or not. We define the value of attribute "BuyBitcoin" to be the label of a record. (a) 4 marks] Please construct a C4.5 decision tree according to the above example. In the decision tree, whenever (1) a node contains at least 80% records with the same label, or (2) a node contains at most 2 records, we stop further processing this node for splitting. Note that you should show the calculation of every attribute's impurity gain at each step of the tree construction (b) [1 marks] Consider a new young user studying CS whose income is fair. Please predict whether this new user will buy Bitcoin or not using the decision tree you trained Question 1. 5 marks] The following shows a history of users with attributes "CS Major" (i.e., majored in computer science), "Age", and "Income". We also indicate whether they will buy Bitcoin or not in the last column No. CSMajorAgeIncome BuyBitcoin fair middlefair fair young high old es es no no yes es no no ves ves ves yes no no no no 3 young old young youngfair middlelow low low We want to train a C4.5 decision tree classifier to predict whether a new user will buy Bitcoin or not. We define the value of attribute "BuyBitcoin" to be the label of a record. (a) 4 marks] Please construct a C4.5 decision tree according to the above example. In the decision tree, whenever (1) a node contains at least 80% records with the same label, or (2) a node contains at most 2 records, we stop further processing this node for splitting. Note that you should show the calculation of every attribute's impurity gain at each step of the tree construction (b) [1 marks] Consider a new young user studying CS whose income is fair. Please predict whether this new user will buy Bitcoin or not using the decision tree you trained

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!