LFD Book Forum role of P(X) ?
 User Name Remember Me? Password
 Register FAQ Calendar Mark Forums Read

 Thread Tools Display Modes
#1
04-11-2013, 02:34 AM
 kokrah Junior Member Join Date: Apr 2013 Posts: 3
role of P(X) ?

The Hoeffding bound for the model H in chapter one, only requires that
we make the assumption that the input examples are a random sample
from the bin; so we can generalize the sample error.

What role does the distribution on X play? It appears to me that we don't need
it. (at least the way the issue of feasibility is setup in chapter 1)
ie. true mismatch ~ sample mismatch.

Thanks.
#2
04-11-2013, 02:48 AM
 yaser Caltech Join Date: Aug 2009 Location: Pasadena, California, USA Posts: 1,478
Re: role of P(X) ?

Quote:
 Originally Posted by kokrah The Hoeffding bound for the model H in chapter one, only requires that we make the assumption that the input examples are a random sample from the bin; so we can generalize the sample error. What role does the distribution on X play? It appears to me that we don't need it. (at least the way the issue of feasibility is setup in chapter 1) ie. true mismatch ~ sample mismatch. Thanks.
We need the existence of the input probability distribution so that "a random sample" becomes well defined, but we don't need any particular to do that since any will correspond to a legitimate for the bin.
__________________
Where everyone thinks alike, no one thinks very much
#3
04-11-2013, 07:07 AM
 nkatz Junior Member Join Date: Apr 2013 Posts: 4
Re: role of P(X) ?

So can you say that P(X) populates the bin and determines mu? In that case we would be sampling P(X); is this correct?
#4
04-11-2013, 10:13 AM
 yaser Caltech Join Date: Aug 2009 Location: Pasadena, California, USA Posts: 1,478
Re: role of P(X) ?

Quote:
 Originally Posted by nkatz So can you say that P(X) populates the bin and determines mu? In that case we would be sampling P(X); is this correct?
affects the value of because it affects the probability of each , so the probability of red marbles (i.e., 's where ) changes accordingly. We are sampling according to , except that when we look at the bin abstraction of the situation, we only care about the color of not its identity, so the binary-event probability is sufficient to characterize sampling of the marbles.
__________________
Where everyone thinks alike, no one thinks very much
#5
02-06-2014, 04:57 PM
 netweavercn Junior Member Join Date: Jan 2014 Posts: 7
Re: role of P(X) ?

Quote:
 Originally Posted by yaser affects the value of because it affects the probability of each , so the probability of red marbles (i.e., 's where ) changes accordingly. We are sampling according to , except that when we look at the bin abstraction of the situation, we only care about the color of not its identity, so the binary-event probability is sufficient to characterize sampling of the marbles.

Thanks prof Yaser's reply. A quick question, as we are sampling according to , how effect each ? In other words, determines or Sampling process or both?
#6
02-07-2014, 04:04 AM
 yaser Caltech Join Date: Aug 2009 Location: Pasadena, California, USA Posts: 1,478
Re: role of P(X) ?

Quote:
 Originally Posted by netweavercn Thanks prof Yaser's reply. A quick question, as we are sampling according to , how effect each ? In other words, determines or Sampling process or both?
The answer would be both, since the probability of each affects the sampling process (not the mechanism of it, but the frequency of different outcomes that it produces).
__________________
Where everyone thinks alike, no one thinks very much
#7
09-07-2015, 12:50 PM
 giridhar1202 Junior Member Join Date: Sep 2015 Posts: 2
Re: role of P(X) ?

Quote:
 Originally Posted by yaser affects the value of because it affects the probability of each , so the probability of red marbles (i.e., 's where ) changes accordingly. We are sampling according to , except that when we look at the bin abstraction of the situation, we only care about the color of not its identity, so the binary-event probability is sufficient to characterize sampling of the marbles.

But isn't fixed when you choose a particular hypothesis h. [ Because number of red marbles is equal to the number of points in the input space where hypothesis ( h ) and target function ( f ) disagree. And this, in my opinion, has nothing to do with probability distribution function ]

Thanks,
Giridhar.
#8
09-07-2015, 08:06 PM
 yaser Caltech Join Date: Aug 2009 Location: Pasadena, California, USA Posts: 1,478
Re: role of P(X) ?

Quote:
 Originally Posted by giridhar1202 But isn't fixed when you choose a particular hypothesis h. [ Because number of red marbles is equal to the number of points in the input space where hypothesis ( h ) and target function ( f ) disagree. And this, in my opinion, has nothing to do with probability distribution function
The number of marbles, or the fraction of marbles, is a simplification to make the experiment more intuitive. In reality, each marble has a probability of being picked, namely , that may be different for other marbles. This affects the total probability of red marbles, which is .

To take a simple example, Let's say that there are only two marbles in the bin, one red and one green, but the red marble has a higher probability of being picked than the green marble. In this case, is not 1/2 though the fraction of red marbles is 1/2.
__________________
Where everyone thinks alike, no one thinks very much
#9
04-11-2013, 08:04 AM
 kokrah Junior Member Join Date: Apr 2013 Posts: 3
Re: role of P(X) ?

I see.

Example:
is the target.
is the input space.

If we let 1.
or
2. , where t(1) is the t-distribution with one degree of freedom.

I know from my stat classes that in case 1. a linear model is actually "correct".
(this is great since we usually know nothing about f)
So in this case the distribution of X plays a role in selecting H, and hence
reducing the in sample error. (assuming the quadratic loss fct.)

Questions:
So in either case 1. or 2. the interpretation/computation of the sample error is the same?
I am a little confused since the overall true error
(which we hope the sample error approximates) is defined based on the joint
distribution of (X,Y); which depends on the distribution of X.

Thanks. I hope this class/book can clear up some mis-conceptions about the theoretical framework of the learning problem once and for all
#10
04-11-2013, 09:55 AM
 Elroch Invited Guest Join Date: Mar 2013 Posts: 143
Re: role of P(X) ?

Quote:
 Originally Posted by kokrah The Hoeffding bound for the model H in chapter one, only requires that we make the assumption that the input examples are a random sample from the bin; so we can generalize the sample error. What role does the distribution on X play? It appears to me that we don't need it. (at least the way the issue of feasibility is setup in chapter 1) ie. true mismatch ~ sample mismatch. Thanks.
As I see it, the theory of generalisation relies on the fact that the distribution P(X, y) which gives the examples used to generate a model (both training and test data) is the same as the distribution of examples which we are trying to learn. There are two things that can go wrong. Either P(X) may be different, or P(y | X) may be different. In the first case, the examples may be concentrated in some subset of the input space, and this may be a region where the models work better. Obviously the second case can also lead to misleading conclusions.

[This may appear to be a trivial assumption when sampling from some populations, but it is likely to be non-trivial in many cases where we are attempting to infer future behavior from past behavior in a system whose characteristics may change]

 Thread Tools Display Modes Hybrid Mode

 Posting Rules You may not post new threads You may not post replies You may not post attachments You may not edit your posts BB code is On Smilies are On [IMG] code is On HTML code is Off Forum Rules
 Forum Jump User Control Panel Private Messages Subscriptions Who's Online Search Forums Forums Home General     General Discussion of Machine Learning     Free Additional Material         Dynamic e-Chapters         Dynamic e-Appendices Course Discussions     Online LFD course         General comments on the course         Homework 1         Homework 2         Homework 3         Homework 4         Homework 5         Homework 6         Homework 7         Homework 8         The Final         Create New Homework Problems Book Feedback - Learning From Data     General comments on the book     Chapter 1 - The Learning Problem     Chapter 2 - Training versus Testing     Chapter 3 - The Linear Model     Chapter 4 - Overfitting     Chapter 5 - Three Learning Principles     e-Chapter 6 - Similarity Based Methods     e-Chapter 7 - Neural Networks     e-Chapter 8 - Support Vector Machines     e-Chapter 9 - Learning Aides     Appendix and Notation     e-Appendices

All times are GMT -7. The time now is 02:38 AM.

 Contact Us - LFD Book - Top