Question: Suppose we have a data set {(x1, y),..., (2N, yn)}. Xi's and labels yi's are both binary: i.e., Li, Yi E {0, 1} for all

 Suppose we have a data set {(x1, y),..., (2N, yn)}. Xi's

Suppose we have a data set {(x1, y),..., (2N, yn)}. Xi's and labels yi's are both binary: i.e., Li, Yi E {0, 1} for all i. We know the generating process of this dataset: for each (Li, Yi), first label yi are generated from the Bernoulli distribution: Yi ~ Bernoulli(1/2). In other words, Pr(yi = 1) = Pr(yi = 0) = 1/2. Then, if yi = 1, then Ti~ Bernoulli(p), if yi = 0, then Xi~ Bernoulli(q). In other words, Pr(xi = 1 | Yi = 1) =p, Pr(di = 1 | Yi = 0) = q. Suppose p > q, we would like to find the Bayes optimal classifier f* :X + Y, which predicts label yi based on ti. (i) (Points: 10) What is the Bayes optimal classifier f*(x)? (ii) (Points: 10) Prove that the classifier has minimal risk among all deterministic classifiers. Suppose we have a data set {(x1, y),..., (2N, yn)}. Xi's and labels yi's are both binary: i.e., Li, Yi E {0, 1} for all i. We know the generating process of this dataset: for each (Li, Yi), first label yi are generated from the Bernoulli distribution: Yi ~ Bernoulli(1/2). In other words, Pr(yi = 1) = Pr(yi = 0) = 1/2. Then, if yi = 1, then Ti~ Bernoulli(p), if yi = 0, then Xi~ Bernoulli(q). In other words, Pr(xi = 1 | Yi = 1) =p, Pr(di = 1 | Yi = 0) = q. Suppose p > q, we would like to find the Bayes optimal classifier f* :X + Y, which predicts label yi based on ti. (i) (Points: 10) What is the Bayes optimal classifier f*(x)? (ii) (Points: 10) Prove that the classifier has minimal risk among all deterministic classifiers

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!