Question: Exercises We can explore the connection between exponential families and the softmax in some more depth. Compute the second derivative of the cross - entropy

Exercises
We can explore the connection between exponential families and the softmax in some more depth.
Compute the second derivative of the cross-entropy loss l(y,hat(y)) for the softmax.
Compute the variance of the distribution given by softmax (0) and show that it matches the second derivative compt
Assume that we have three classes which occur with equal probability, i.e., the probability vector is (13,13,13).
What is the problem if we try to design a binary code for it?
Can you design a better code? Hint: what happens if we try to encode two independent observations? What if we ence
Softmax is a misnomer for the mapping introduced above (but everyone in deep learning uses it). The real softmax is defined log(exp(a)+exp(b)).
Prove that RealSoftMax (a,b)>max(a,b).
Prove that this holds for -1 RealSoftMax (a,b), provided that >0.
Show that for we have -1 RealSoftMax (a,b)max(a,b).
What does the soft-min look like?
Extend this to more than two numbers.
Discussions
Exercises We can explore the connection between

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Programming Questions!