Question: model = models. Sequential ( ) # Layer 1 model.add ( layers . Dense ( 5 0 , activation = 'relu', input # Layer 2
model models. Sequential # Layer model.add layers Dense activation'relu', input # Layer model.addlayers Dense activation'relu' # Layer model.addlayersDense activation'relu' # Layer model.addlayers Dense B activationC model.compile optimizer sgd loss D metri A What is the input shape A in Layer or B What will be the value of B activation function C and loss total number of classes in the dataset is i sigmoid, binarycrossentropy softmax, categoricalcrossentropyA. Consider the MLP below where each perceptron has a linear activation function where the output z is related to its inputs x and y as z ax by c We claim that this network can be replaced by a single perceptron with a possibly different linear activation function. Is this claim true or not? Give quantitative justification for your answer. B We would like to minimize the following quadratic function xx using exactly two steps using gradientdescent with a learning rate n for the first step. What should the learning rate of the secondstep be in order to obtain the minimum in twosteps? Marks
Step by Step Solution
There are 3 Steps involved in it
1 Expert Approved Answer
Step: 1 Unlock
Question Has Been Solved by an Expert!
Get step-by-step solutions from verified subject matter experts
Step: 2 Unlock
Step: 3 Unlock
