Question: Suppose we have three classes in two dimensions with the following underlying distributions: I Class !1: pxj!1 = N0, I. I Class !2: pxj!2 =

Suppose we have three classes in two dimensions with the following underlying distributions:

I Class !1: p¹xj!1º = N¹0, Iº.

I Class !2: p¹xj!2º = N

 h 11 i

, I



.

I Class !3: p¹xj!3º = 1 2N

 h 0.5 0.5 i

, I



+ 12 N

 h

????0.5 0.5 i

, I



.

Here,N¹, º denotes a two-dimensional Gaussian distribution with mean vector  and covariance matrix

, and I is the identity matrix. Assume class prior probabilities Pr¹!iº = 13, i = 1, 2, 3.

a. Classify the feature x =

h 0.25 0.25 i

based on the MAP decision rule.

b. Suppose the first feature is missing. Classify x =

 

0.25



using the optimal rule derived in Q10.1.

c. Suppose the second feature is missing. Classify x =

h 0.25



i using the optimal rule from Q10.1.

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Pattern Recognition And Machine Learning Questions!

Q:

a