![]() |
|
#1
|
|||
|
|||
![]()
Hello.
I understand that learning is picking out a function from a candidate set of functions that most closely resembles the target function. The feasibility of learning would be related to how close this resemblance is. I understand also that Hoeffding's Inequality is an upper bound to the probability that the in-sample error rate deviates significantly from the real error rate. In the end, this upper bound simply implies that given a large enough sample, estimating the real error rate is feasible. Is there any misconception in anything here so far? So my question is: what does the Hoeffding inequality say about the feasibility of learning? Shouldn't it be feasibility of verifying hypothesis? |
#2
|
||||
|
||||
![]() Quote:
Your understanding is correct. The contrast between learning and verification that you allude to is precisely why the union bound was used in Lecture 2. The notion of the feasibility of learning will be further discussed next week in Lecture 4, where the question is split into two parts. Stayed tuned!
__________________
Where everyone thinks alike, no one thinks very much |
#3
|
|||
|
|||
![]()
Thank you very much for your quick response, Professor Yaser. Your course sets the bar for quality in MOOCs very high. You have my respect and gratitude.
At the beginning of minute 32 in the Lecture 2 video, you say "P can be any probability, but the choice of P can affect the value of ...". Is it ![]() ![]() ![]() ![]() |
#4
|
||||
|
||||
![]() Quote:
![]() ![]() ![]() ![]() ![]()
__________________
Where everyone thinks alike, no one thinks very much |
#5
|
|||
|
|||
![]()
Thanks for the lectures. I am a stats phd student at MD (new to the ideas of ML). A friend recommended your site.
![]() My understanding of lecture 2 is that you are setting up a general framework to answer the question of "Is this model feasible?". In the -tossing 1000 coins 10 times analogy- each of the 1000 coins are the same. i.e. each of the possible h's in H are thought of as being the same in some sense, at least in the goal of finding a crude bound. The prob. distribution placed on the input space X affects the bin content and hence the sample content for any h in the Model, H. Question: In this 1st step framework: a small (overall) bound of say 0.001 implies a g/model is verified as learnable? i.e. any g/H is learnable is you if have a very large sample size and reasonable M? Any comments/corrections from anyone is appreciated. Thanks |
![]() |
Tags |
feasibility of learning, hoeffding's inequality |
Thread Tools | |
Display Modes | |
|
|