![]() |
#1
|
|||
|
|||
![]()
I have a general question regarding weight decay regularization.
Since ![]() For this to make intuitive sense so that the regularization correctly "charges" the hypothesis a cost for being more complex, it seems to me that all the features must be normalized to have zero mean. Otherwise for example if all the data points are in a ball far from the origin, regularization could fail in the sense that a "good" classifier would have ![]() ![]() I'm not sure about this reasoning, is it correct? Is this a concern in practice? Thanks! |
#2
|
|||
|
|||
![]() |
#3
|
||||
|
||||
![]() Quote:
If one matches the regularization criterion to a given problem, the regularizer may be more specific than general weight decay. For instance, when we discuss SVM next week, the regularizer will indeed exclude ![]() ![]() As emphasized in the lecture, the choice of a regularizer in a real situation is largely heuristic. If you have information in a particular situation that suggests that one form of regularizer is more plausible than the other, then that overrules the general choices that are developed for a different, idealized situation. In all of these cases, the amount of regularization ( ![]() ![]()
__________________
Where everyone thinks alike, no one thinks very much |
#4
|
|||
|
|||
![]()
Ok, I understand now. Thank you.
Including ![]() In either case, performing validation should allow us to narrow down the types of regularizers that make sense for a particular data set and application. |
![]() |
Thread Tools | |
Display Modes | |
|
|