Restricted Learner's Rule of Thumb (Lecture 11)
Hello All:
In minute 23:45 of Lecture 11, the restricted learner's reasoning is based on a rule of thumb, whereby you should have 10 data points for every parameter you want to estimate in your model. Where (in the book or in the other lectures) can I find more information on the justification for this rule of thumb? 
Re: Restricted Learner's Rule of Thumb (Lecture 11)
Quote:
The rule of thumb is a practical observation, so its real justification is simply that it has worked most of the time in practice. Once can justify the form that the number of examples is a multiple of the VC dimension by arguing that having multiple data points to fit per degree of freedom will force that degree of freedom to a 'compromise' that is likely to capture what is common between these data points, i.e., likely to generalize. Whether that multiple is 5 or 10 or 100, however, is an empirical observation that is difficult to reason about in a genuine way. 
Re: Restricted Learner's Rule of Thumb (Lecture 11)
Thank you for the answer, Professor. So I understand that it is not possible (even based on your experience) to give a single value for this multiple because a single value cannot cover all possible modeling situations (target complexity, stochastic/deterministic noise, etc.)?

Re: Restricted Learner's Rule of Thumb (Lecture 11)
Quote:

Re: Restricted Learner's Rule of Thumb (Lecture 11)
On a specific point of jlaurentum, I have come to the intuitive conclusion that noise effectively reduces the size of the set of data, with the more noise, the more more data points needed to achieve the same results.
This is related to the much simpler idea that if you want to estimate a mean average with noisy (i.e. nonzero variance) data, then the accuracy is inversely proportional to the square root of the number of data points. Likewise, I would conjecture that a noisy machine learning problem might be reduced to a near noiseless one by having a very large number of data points (although the quantitative details of this are less clear). In machine learning there is the complication that this intuition only applies to genuine (stochastic) noise. As "deterministic noise" is unvarying, it is not reduced (but the variance is). On reflection, I feel the term "deterministic noise" can be a little misleading, as it is a form of error which merely mimics noise to an observer, but lacks one of its properties (randomness). As an analogy with a physical measurement, it is more similar to a calibration error than to an uncertainty in measurement. 
All times are GMT 7. The time now is 09:50 AM. 
Powered by vBulletin® Version 3.8.3
Copyright ©2000  2021, Jelsoft Enterprises Ltd.
The contents of this forum are to be used ONLY by readers of the Learning From Data book by Yaser S. AbuMostafa, Malik MagdonIsmail, and HsuanTien Lin, and participants in the Learning From Data MOOC by Yaser S. AbuMostafa. No part of these contents is to be communicated or made accessible to ANY other person or entity.