Question: Problem 1. (18 points. Consider the linear regression model $Y=X beta+varepsilon$ with pre-determined (i.e., non-stochastic) regressors $X$. All of the usual assumptions hold except, instead

Problem 1. (18 points. Consider the linear regression model $Y=X \beta+\varepsilon$ with pre-determined (i.e., non-stochastic) regressors $X$. All of the usual assumptions hold except, instead of $\operatorname[Var}(\varepsilon)=\sigma^{2} I_{n}$, we have $\operatorname{Var}(\arepsilon)=E\left(\arepsilon \varepsilon^{\prime} ight)=$ $\Sigma_{\varepsilon}$, where $\Sigma_{\varepsilon}$ is a generic symmetric positive definite matrix. In other words, we are assuming that the errors are not necessarily homoskedastic and uncorrelated. (1) (2 points) The traditional least-squares estimator is $\widehat{\beta}_{L S}=\left(X^{\prime} X ight)^{-1} X^{\prime} Y$ Define, now, a new estimator $\widehat{\beta}_{GL S}=\left(X^{\prime} \Sigma_{\varepsilon}^{-1} X ight)^{-1} X^{\prime} \Sigma_{\varepsilon}^{-1} Y$. Call it "generalized least- squares" (GLS) estimator. Show that $E\left(\widehat{\beta}_{GL S} ight)=\beta$, i.e., $\widehat{\beta}_{G L 5}$ is unbiased. (2) (3 points) show that the variance of $\widehat{\beta}_{L S}$ under the new assumption is $\left(X^{\prime} X ight)^{-1} X^{\prime} \Sigma_{\varepsilon} X\left(X^{\prime} X ight)^{-1}$. (3) (3 points) Show that the variance of $\widehat{\beta}_{GL S}$ under the new assumption is $\left(X^{\prime} \Sigma_{\varepsilon}^{-1} X ight)^{-1}$. (4) (4 points) Show that the variance of $\widehat{\beta}_{G L S}$ is not larger than the variance of the traditional least-squares estimator $\widehat{\beta}_{L S}$. In other words, show that $\left(X^{\prime} X ight)^{-1} X^{\prime} \Sigma_{\varepsilon} X\left(X^{\prime] X ight)^{-1}-\left(X^{\prime} \Sigma_{\varepsilon}^{-1} X ight)^{-1} \geq 0$. SP.PB. 187
Step by Step Solution
There are 3 Steps involved in it
Get step-by-step solutions from verified subject matter experts
