Question: ( a ) Figure 1 below plots the loss when batch gradient descent is used for training. Which optimizers plots the loss in figure 2

(a) Figure 1 below plots the loss when batch gradient descent is used for training. Which
optimizers plots the loss in figure 2 and 3.[1]
Figure 1
(b) In figure 2, assume that the learning rate used was 0.5. Redraw the plot to show the
effect of increasing the learning rate to 2 and decreasing the learning rate to 0.01.[2]
(c) You are given a simple neural network with a single hidden layer containing two neurons,
and an output layer containing one neuron. All neurons use the sigmoid activation
function. The weights and biases of the network are as follows: [3]
Weights from input to hidden layer: w1=0.5,w2=0.6
Biases in hidden layer: b1=0.1, b2=-0.2
Weights from hidden layer to output: w3=0.7, w4=-0.8
Bias in output layer: b3=0.3
Given an input x =0.75, calculate the output of the network and mean squared loss
if the desired output is 1.25. Use the sigmoid activation function. What will be the
effect in the loss if the loss function used is
L(w, b)=1
2(d y)2+||w2||

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!