#1




Choice of regularization parameter
Will try and ponder my question from the lecture again, since I wasn't quite dextrous enough to get the point across in chatroom format.
I think the point I was missing is that in Fig. 4.7(b) (last graph on slide 21 of May 10 lecture), the stochastic noise is in fact fixed at zero. I was probably having flashbacks to Fig. 4.3(b), where it's a fixed nonzero value, in which case the behavior of E_out would depend on N as well as lambda, right? So I was wondering for what choice of N the graph was plotted, and how the behavior of the Q_f = { 15, 30, 100 } lines would change with N. And imagining that N=15 as in previous examples, it was surprising that regularization wouldn't help out when Q_f=15! But with zero stochastic noise, the expected deterministic noise is just whatever it is, independent of N, as the fit is the same regardless of what random points you pick. Well, I suppose we'd better have N >Q_f, at least, or we're in trouble! Have I got that right? 
#2




Re: Choice of regularization parameter
Which leads me down the road to a related quandary  won't the red and green lines (with nonzero stochastic noise) in Fig. 4.7(a) move around a bit depending on N? (Because of the dependence as shown in Fig. 4.3(a)?) So haven't you had to choose a value of N to generate those? Or am I missing something?

#3




Re: Choice of regularization parameter
In both plots of Fig 4.7, N=30 (as in fig 4.6)  unfortunately it is not mentioned in the caption to Fig 4.7. The general shape of the plots will not change if N increases, and yes if N is less than the degree of the polynomial you are fitting, then there will be problems and you have to use the pseudoinverse.
If your model complexity matches the target complexity (Q=Q_f) *and* there is no stochastic noise, with at least N=Q+1 data points, you will recover the target function exactly without regularization. Any regularization will therefore necessarily result in an inferior Eout. Hope this helps, Quote:
__________________
Have faith in probability 
#4




Re: Choice of regularization parameter
Yes indeed, thanks. I knew there was an N hiding in there somewhere! Also wasn't keeping the potential distinction between Q/Qf entirely clear in my head, so that also demystifies it a bit.

Thread Tools  
Display Modes  

