LFD Book Forum Q19
 Register FAQ Calendar Mark Forums Read

#11
06-09-2013, 11:33 AM
 sptripathi Junior Member Join Date: Apr 2013 Posts: 8
Re: Q20

Need help in verifying if below understanding is correct ?

The Bayesian:
P(h=f | D) = P(D | h=f) * P(h=f) / P(D)

For this Q, we are given:
P(h=f) is uniform in [0,1]
D: one-person-with-heart-attack
Pick f = c (constant)

To simplify, I assume that h and f are a discrete random-variables with 10 possible values from (0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, 0.9, 1.0)
and each is equally likely with P=1/10. Essentially simplifying here to make P(h=f) a pmf which is actually a pdf.

Now:

P (D | h=f)
= Pr( one-person-with-heart-attack | h=f )
= Probability of one-person-with-heart-attack, given (h=f)
= c

( because if h=f were given, then the Prob of one picked person getting heart-attack is c, as defined by f )

Plug in above to get:
P(h=f | D) = c * P(h=f) / P(D)

Does above sound correct ?
Also P(D) =1 in this case ?

Thanks.
#12
06-09-2013, 03:15 PM
 Dorian Member Join Date: Apr 2013 Posts: 11
Re: Q20

I find this exercise simple but very useful. If one thinks of the series of following measurements (1s and 0s for heart attack or not) one can clearly form an idea how this transforms step-by-step from a uniform distribution to a Bernoulli one.

Does this mean that this example represents one of those cases where the initial prior is irrelevant and we can safely use it for learning? Also, is this some form of reinforcement learning?

thanks,
Dorian.
#13
06-09-2013, 06:23 PM
 yaser Caltech Join Date: Aug 2009 Location: Pasadena, California, USA Posts: 1,478
Re: Q20

Quote:
 Originally Posted by Dorian Does this mean that this example represents one of those cases where the initial prior is irrelevant and we can safely use it for learning? Also, is this some form of reinforcement learning?
In this case, with sufficient number of examples, the prior indeed fades away. Noisy examples blur the line between supervised and reinforcement learning somewhat as the information provided by the output is less definitive than in the noiseless case.
__________________
Where everyone thinks alike, no one thinks very much
#14
06-10-2013, 09:36 PM
 nkatz Junior Member Join Date: Apr 2013 Posts: 4
Re: Q20

I am very confused by this problem. Perhaps this questions will help:
Is P(D|h=f) a function of D or h or both? It looks to me like it's a function of D, but we need to convert it to a function of h to get the posterior... Is this correct?
#15
06-10-2013, 10:21 PM
 yaser Caltech Join Date: Aug 2009 Location: Pasadena, California, USA Posts: 1,478
Re: Q20

Quote:
 Originally Posted by nkatz Is P(D|h=f) a function of D or h or both?
Let us first clarify the notions. The data set has one data point in it which is either (heart attack) or (no heart attack). Being a function of means being a function of that value (), so indeed is a function of , and it so happens in this problem that the value is fixed at . The probability is also a function of (which happens to be the same as according to what we are conditioning on).

Since is fixed, this leaves as a function of only.
__________________
Where everyone thinks alike, no one thinks very much
#16
06-11-2013, 04:46 AM
 Elroch Invited Guest Join Date: Mar 2013 Posts: 143
Re: Q20

There is an analogy that may be enlightening, which I thought of because of the presentation of the first part of this course.

Suppose you have a large number of urns each containing a large number of black and white balls in varying proportions. You are told how many urns there are with each proportion.

Then you go up to one of the urns and take out a ball which you find is black. The question is how likely it is that specific urn has each particular fraction of black balls.

 Thread Tools Display Modes Linear Mode

 Posting Rules You may not post new threads You may not post replies You may not post attachments You may not edit your posts BB code is On Smilies are On [IMG] code is On HTML code is Off Forum Rules
 Forum Jump User Control Panel Private Messages Subscriptions Who's Online Search Forums Forums Home General     General Discussion of Machine Learning     Free Additional Material         Dynamic e-Chapters         Dynamic e-Appendices Course Discussions     Online LFD course         General comments on the course         Homework 1         Homework 2         Homework 3         Homework 4         Homework 5         Homework 6         Homework 7         Homework 8         The Final         Create New Homework Problems Book Feedback - Learning From Data     General comments on the book     Chapter 1 - The Learning Problem     Chapter 2 - Training versus Testing     Chapter 3 - The Linear Model     Chapter 4 - Overfitting     Chapter 5 - Three Learning Principles     e-Chapter 6 - Similarity Based Methods     e-Chapter 7 - Neural Networks     e-Chapter 8 - Support Vector Machines     e-Chapter 9 - Learning Aides     Appendix and Notation     e-Appendices

All times are GMT -7. The time now is 10:01 AM.