![]() |
#1
|
|||
|
|||
![]()
Links: [Lecture 13 slides] [all slides] [Lecture 13 video]
Question: (Slide 7/22) 1. Why do we report ![]() ![]() ![]() ![]() Answer: 1. Because from theoretical analysis we know, that the more points we have in the dataset the better the learning outcome is. So it is better to use N points for training, than N-K, although we can't measure, how much better it is. 2. Because we can't report ![]() ![]() |
#2
|
|||
|
|||
![]()
Question: (Slide 7/22) The rule of thumb is N/5. Why do we have N/10 on the last slide?
Answer: On the last slide we use cross-validation, on slide 7/22 we do validation just once, so it is a different game. It is clear, that in cross-validation we should use less points for validation, because in any case we repeat the process and finally end up using all the points, so we better increase ![]() |
#3
|
|||
|
|||
![]()
Question: (Slide 11/22) If you already have all the hypothesis, why do you do validation and choose a model instead of doing aggregation?
Answer: In practice people often use aggregation and often it does perform better. If you have 100 points, you can train on all those points or train on 99 points, leaving n-th point out and then take the average over n. And sometimes you get better results, despite the fact, that after all you still use the same 100 points. The reason is that this process may reduce variance and, thus, is less affected by the noise. |
![]() |
Thread Tools | |
Display Modes | |
|
|