Question: Assume a random vector x E R follows a multivariate Gaussian distribution (i.e., p(x) N(x , E)). If we apply an invertible linear transformation

Assume a random vector x E R Show that mutual information satisfies the following: I(X,Y) H(X) - H(XY) H(Y) - H(YX) H(X) + H(Y) - H(X,Y).

Assume a random vector x E R" follows a multivariate Gaussian distribution (i.e., p(x) N(x , E)). If we apply an invertible linear transformation to convert x into another random vector as y = Ax+ b (A Rn and b R"), prove that the joint distribution p(y) is also a multivariate Gaussian distribution, and compute its mean vector and covariance matrix. Show that mutual information satisfies the following: I(X,Y) H(X) H(XY) H(Y) - H(YX) H(X) + H(Y) - H(X,Y). = = = Compute the distance of a point x0 IR to b) a unit ball ||x|| 1; c) an elliptic surface x Ax = 1, where A E R"X" and A > 0. Hints: Give a numerical procedure if no closed-form solution exists.

Step by Step Solution

3.45 Rating (155 Votes )

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock

lets break it down step by step 1 Proving Gaussian Transformation Step 11 Given A random vector xRn ... View full answer

blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Programming Questions!