Question: We have each data point x = [ [ x _ 1 x _ 2 ] ] ^ T in ^ 2 , and we

We have each data point x=[[ x_1 x_2]]^T in ^2, and we encounter the following data points: We are going to build a linear model (i.e., using least square) to predict the label of each data point x=(x_1, x_2) as follows: =w_0+w_1 x_1+w_2 x_2 where (w_0, w_1, w_2) are unknown weights that we will learn based on the above training dataset. Here w_0 is the intercept. (a) Gradient descent. Starting with (w_0, w_1, w_2)=(0,0,0), we will run the gradient descent algorithm to learn these weights. Run for T=4 iterations and report the value of the weights and the corresponding total error at the end of each iteration. The learning rate is \alpha =0.02. Note. For this question, you can compute each step manually or write a python program for it. If you write a python program, you will need to submit the program together with your answers. 1(b) Closed-form solution. Now we are going to use the closed-form solution to find (w_0, w_1, w_2). Provide the values of (w_0, w_1, w_2). What is the corresponding total error? Note. For this question, you are allowed to use any tool to perform matrix inverse or multiplication.

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!