Question: L 2 regularization: J R ( w ) = J ( w ; D ) + | | w | | 2 2 L 1

L2 regularization:
JR(w)=J(w;D)+||w||22
L1 regularization:
JR(w)=J(w;D)+
where J(w;D) is the original cost function (cost function without regularization) for training
of a general parametric ML model.
Justify the following facts.
(a)L2 regularization pushes all parameters towards small values (but not necessarily exactly
zero).
(b)L1 tends to favor so-called "sparse" solutions, where only a few of the parameters are
non-zero, and the rest are exactly zero.
 L2 regularization: JR(w)=J(w;D)+||w||22 L1 regularization: JR(w)=J(w;D)+ where J(w;D) is the original

Step by Step Solution

There are 3 Steps involved in it

1 Expert Approved Answer
Step: 1 Unlock blur-text-image
Question Has Been Solved by an Expert!

Get step-by-step solutions from verified subject matter experts

Step: 2 Unlock
Step: 3 Unlock

Students Have Also Explored These Related Databases Questions!