LFD Book Forum (http://book.caltech.edu/bookforum/index.php)
-   Homework 5 (http://book.caltech.edu/bookforum/forumdisplay.php?f=134)
-   -   Hw5 Q8 E_out (http://book.caltech.edu/bookforum/showthread.php?t=4265)

 marek 05-05-2013 05:55 PM

Hw5 Q8 E_out

I am struggling to understand how to calculate in this question. I have two competing theories, which I will describe below. Any help is greatly appreciated.

Once the algorithm terminates, I have . I now generate a new set of data points . Using my original target function to generate the corresponding .

Case 1. Just use the same cross entropy error calculation but on this new data set.

Case 2. Directly calculate the expected output of our hypothesis function and compare to .

with probability

Ultimately this gives us the probability that our hypothesis aligns with Y:

In the lectures/book, we would multiply these probabilities to get the "likelihood" that the data was generated by this hypothesis. However, it seems that averaging over these should give the expected error in this sample.

It feels as though the first approach is the correct one, but I struggle because the second approach makes intuitive sense since that is how I historically I would have calculated . To make matters worse, the two approaches very closely approximate different answers in the question!

 yaser 05-05-2013 09:07 PM

Re: Hw5 Q8 E_out

Quote:
 Originally Posted by marek (Post 10706) I am struggling to understand how to calculate in this question. I have two competing theories, which I will describe below. Any help is greatly appreciated. Once the algorithm terminates, I have . I now generate a new set of data points . Using my original target function to generate the corresponding . Case 1. Just use the same cross entropy error calculation but on this new data set.
The above approach is correct. The problem specifies the cross entropy error measure, so , where the expectation is w.r.t. both . The above formula estimates that through a random sample.

 marek 05-05-2013 09:12 PM

Re: Hw5 Q8 E_out

Quote:
 Originally Posted by yaser (Post 10708) The above approach is correct. The problem specifies the cross entropy error measure, so , where the expectation is w.r.t. both . The above formula estimates that through a random sample.
I suspected as much. I'll try to figure out why my other approach is wrong tomorrow. I think I've burned out on it today and am probably not seeing something obvious. Thanks for your help!

 arcticblue 05-06-2013 02:37 AM

Re: Hw5 Q8 E_out

I am also a little unsure about exactly how this equation works:

Obviously the more negative is the closer E_out is to zero which is good. So is w supposed to be normalized? I presume so because otherwise I could just scale w and then E_out becomes very small. And if it is normalized then the values I'm getting for E_in and E_out are both much greater than any of the options. (Maybe it's meant to be like that, if so it's quite unnerving.)

 yaser 05-06-2013 10:14 AM

Re: Hw5 Q8 E_out

Quote:
 Originally Posted by arcticblue (Post 10720) I am also a little unsure about exactly how this equation works: Obviously the more negative is the closer E_out is to zero which is good. So is w supposed to be normalized? I presume so because otherwise I could just scale w and then E_out becomes very small. And if it is normalized then the values I'm getting for E_in and E_out are both much greater than any of the options. (Maybe it's meant to be like that, if so it's quite unnerving.)
No normalization. The value of is determined iteratively by the specific algorithm given in the lecture. If 'agrees' with all the training examples, then indeed the algorithm will try to scale it up to get the value of the logistic function closer to a hard threshold. When you evaluate the quoted formula on a test set, is frozen and no scaling or any other change in it is allowed.

 Michael Reach 05-06-2013 02:39 PM

Re: Hw5 Q8 E_out

Quote:
 If 'agrees' with all the training examples, then indeed the algorithm will try to scale it up to get the value of the logistic function closer to a hard threshold.
Thank you for this comment - I finally have some idea what I'm seeing in the homework. This point is worth stressing: the scale of w determines the sharpness of the threshold.

 arcticblue 05-06-2013 04:22 PM

Re: Hw5 Q8 E_out

Thank you for explaining that normalization is not required. Your explanation now makes a bit more sense why I see the weights continue to increase in value the more iterations that I run.

 hsolo 07-16-2013 05:27 AM

Re: Hw5 Q8 E_out

One very minor point which may help error prone folks such as myself:
In Linear methods we always have a d+1 dimensional weight vector by adding an extra pseudo-coordinate of 1 for each training and test point. I completely overlooked this and was struggling with the error never getting as small as the expected answer. Spent a long time rechecking code etc.

As soon as I fixed this, things fell into place. Likely I will never forget this :-)

 All times are GMT -7. The time now is 09:29 AM.