Question: 6 . [ Neural Networks ] ( 5 pts ) a . ( 1 pts ) True or False: For any neural network, the validation

6.[Neural Networks](5 pts)
a.(1 pts) True or False: For any neural network, the validation loss will always decrease monotonically with the number of iterations of gradient descent, provided the step size is sufficiently small.
- True
- False
b. Let \( f \) be a fully-connected neural network with input \( x \in \mathbb{R}^{M}, P \) hidden layers with \( K \) nodes per layer and logistic activation functions, and a single logistic output. Let \( g \) be the same network as \( f \), except we insert another hidden layer with \( K \) nodes that have no activation function (or equivalently, the identity activation function), so that \( g \) has \( P+1\) hidden layers. Denote this new layer \( L^{\text {new }}\). Assume that there are no bias terms for any layer nor for the input. (Please select one option for all the following questions.)
i.(1 pts)\( f \) can learn the same decision boundary as \( g \) if the additional linear layer is placed
- immediately after the input.
- immediately before the last sigmoid activation function.
- anywhere in between the above two choices.
- none of the above.
ii.(1 pts) Assume that \( L^{\text {new }}\) is placed in between two other hidden layers in \( g \). How many more parameters does \( g \) learn compared to \( f \)?
-\( K \)
-\( K^{2}\)
-\( K P \)
-\( K M \)
-\(2 K^{2}\)
iii. (1 pts) True or False: After training \( f \) and \( g \) to convergence, \( g \) can have a lower training loss than \( f \).(Use the same assumption as in ii).
- True
- False
iv.(1 pts) True or False: After training \( f \) and \( g \) to convergence, \( f \) can have a lower training loss than \( g \).(Use the same assumption as in ii).
- True
- False
6 . [ Neural Networks ] ( 5 pts ) a . ( 1 pts )

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Programming Questions!