#1




Q19
I was wondering if the problem assumes that some learning has been done to determine P(Dh=f) for some population or if the person with the heart attack is the only person in D. Obviously, I may not understand Bayesian analysis.

#2




Re: Q20
Quote:
__________________
Where everyone thinks alike, no one thinks very much 
#3




Re: Q20
Thank you very much for the quick reply. This has been an amazing course!!

#4




Re: Q20
I am still a bit confused about the setup of the problem. Is it correct to assume that what we are trying to determine is the underlying probability of somebody picked at random from the population to have a heart attack out of a single sample? If so, shouldn't be relevant? If a single point is all we have, call it the binary variable  equal to 1 if the patient has a heart attach; 0 if he doesn't, that would be the probability of generating a single point with a patient having a heart attack, given the underlying probability that a person has a heart attack, right? In that case, the posterior is going to have two cases and . The question refers only to case right?

#5




Re: Q20
Quote:
It should be based on rater than out of. A source of confusion here is that is a probability, but then we have a probability distribution over . Let us just call the fraction of heart attacks in the population. Then the problem is addressing the probability distribution of that fraction  Is the fraction more likely to be 0.1 or 0.5 or 0.9 etc. The prior is that that fraction is equally likely to be anything (uniform probability). The problem then asks how this probability is modified if we get a sample of a single patient and they happen to have a heart attack. If I have not answered your question, please ask again perhaps in those terms.
__________________
Where everyone thinks alike, no one thinks very much 
#6




Re: Q20
Quote:

#7




Re: Q20
__________________
Where everyone thinks alike, no one thinks very much 
#8




Re: Q20
If I understand the problem correctly, P(X=1) is independent on P(h=f). Correct?

#9




Re: Q20
Quote:
The reason why you can ignore P(X=1) is because in Bayesian analysis you usually don't care about the absolute probability of the dataset since it is just a constant that all of your hypotheses are divided by, equally, so it doesn't affect which hypothesis is aposteriori most probable. 
#10




Re: Q20
These are some fine points here. We have to use with exact meaning of terms, using them loosely can create misunderstandings.
At a first reading I thought that Haowen's answer was not correct, and also ilya's remark was not correct too. On a second reading, Haowen's answer is correct, but I am not sure that it answers the initial question, since the initial question/remark by ilya was ambiguous. Let me explain. When we talk about independency we can talk about it in probability terms, where we have specific rules on random variables being independent, or we can talk about it in more loose/everyday terms when we want to express that something affects something else. Ilya's question is expressed loosely. It talks about independence of Probabilities not random events. It can be taken with several different meanings. 1) If you are really asking whether X=1 and h=f are independent events, then we can clearly say they are not. The choice of h clearly affects the probability of X=1. More specifically, the choice of h is the probability of X=1. 2) If you are asking whether the distribution of h=f affects P(X=1) for all possible h, (which can be taken as the more literal interpretation of what you are asking) then again: yes there is a connection and Haowen gives you the formula. 3) If you are asking in general "should we care about calculating the value for P(X=1)", then Haowen gives you the answer again. 4) If you are asking whether the event X=1 affects the probability of h=f, it depends whether you are really referring to the apriori or the aposteriori. It does not affect the apriori and it does affect the aposteriori (and Q20 asks how). 
Thread Tools  
Display Modes  

