![]() |
#1
|
|||
|
|||
![]()
In the lecture 11, it is mentioned that deterministic noise depends on the hypothesis set, and deterministic noise decreases as the hypothesis set becomes more complex, because the set can tackle some of the deterministic noise.
But the experiment performed on slide 7 of lecture 11 suggests that the more complex model fails badly when compared to the less complex model. The above statements seem to contradict in my mind, I cannot put them together. Can anyone please clear my doubt. Thank you. |
#2
|
|||
|
|||
![]()
This is indeed confusing and after spending some time thinking about this point I think I have finally understood it (I hope). Deterministic noise is nothing but the bias in the modelling hypothesis. So the more complex model will indeed have less deterministic noise (small bias). But this does not imply that this model will also have smaller Eout. Because Eout also depends on the variance of the hypothesis and since the variance of the more complex model will be large for small N, this means that Eout will be large for more complex model. But if we have sufficiently large sample size (i.e. large N) then both the variance and the bias (i.e. deterministic noise) will be small for complex model. Hence in this case the more complex model will outperform the simpler model. So the lesson learnt is: Complex model is better than simple model provided we have sufficient data. For small data sets, complex models overfit and it is better to choose simple models.
|
#3
|
||||
|
||||
![]() Quote:
If you want to see the impact of deterministic noise by itself, without having another factor that affects overfitting in play, you should fix the hypothesis set and increase the complexity of the target function.
__________________
Where everyone thinks alike, no one thinks very much |
#4
|
|||
|
|||
![]() Quote:
Can we say the overfit measure, indicates noise? |
#5
|
||||
|
||||
![]()
It does. The specific way it indicates it is still a research topic.
__________________
Where everyone thinks alike, no one thinks very much |
#6
|
|||
|
|||
![]() Quote:
The two effects push overfitting in different directions, this means that if we fix a target, and expand H, the more complex H has a better ability to tackle the deterministic noise, but the complexity of the H in turn, makes for bad overfitting? |
#7
|
||||
|
||||
![]() Quote:
![]()
__________________
Where everyone thinks alike, no one thinks very much |
#8
|
|||
|
|||
![]()
Is there anything interesting to be said about the relationship between overfitting and the difference between
![]() ![]() Edit: I've been assuming that the definitions of bias and variance take expectations over all data sets of a particular fixed size ![]() ![]() ![]() |
#9
|
|||
|
|||
![]()
I think I am still a bit unclear about deterministic noise. Doesn't the amount of noise (deterministic or stochastic) depend on both the bias and variance of the noise? For a given N, the more complex noise will have a higher variance but lower bias. Hence doesn't the amount of noise depend upon N and the relative complexity difference between the target function and the hypothesis (the level of deterministic noise if you will)?
![]() |
#10
|
|||
|
|||
![]()
In general a more complex H implies lower "deterministic noise" but it is important to take into consideration the amount of training data that you have (N) when discussing Eout. In the example shown in lecture 11 the target function was very complex (50th order) and the training data was noiseless. We could see that a simple hypothesis (second order pol) gave a much better Eout than the more complex hypothesis (10th order polynomial). In this case there was only "deterministic noise" and the more complex Hypothesis performed much worse even if the "deterministic noise" was lower for the more complex H.
|
![]() |
Tags |
deterministic noise, lecture 11 |
Thread Tools | |
Display Modes | |
|
|