![]() |
Re: HW 2 Problem 6
Quote:
2-nd scenario: fit linear model only once. Repeat 1000 times: generate 100 out-of-sample points, test. Accumulate and average errors when done. Here I get remarkable variation in average error. I'd like to understand why these scenarios different. I believe they must not |
Re: HW 2 Problem 6
Quote:
|
Re: HW 2 Problem 6
Quote:
Now as I plotted both lines (target and hypothesis) per your advice I begin to think that this is maybe what we should expect. Linear regression not always fits well. Usually it looks good giving small in-sample error. But sometimes disagreement is visually large (>0.1 in-sample error). This is the root of variation in average error when I use same "bad" regression for all 1000 iterations. I hope this type of experiment isn't implied by the problem. Otherwise it has no certain answer - at least 2 answers match. So there is another question. Is >0.1 in-sample error and visually non-optimal fit still valid outcome of linear regression for linearly separable data? |
Re: HW 2 Problem 6
Quote:
|
Re: HW 2 Problem 6
Quote:
|
Re: HW 2 Problem 6
I also observe some discrepancy while computing Eout. When I hold the target function fixed, Eout is approximately equal to Ein. When I use different target functions for each experiment, Eout is significantly higher than Ein. Is this expected?
|
Re: HW 2 Problem 6
@MLearning
In my opinion E_out should be near E_in when target is not fixed(i.e averaging over 1000 iterations) . Unfortunately problem can be anywhere , but most probably in computing error. Could it be that when you are computing error for E_out for one iteration, the number of sample points for out of sample are 1000 , and if you forgot to change the number of samples points from 100 (from Q 5) to 1000 , then may be that is the cause of difference. (misclassified/sample_size) I am just guessing , since I make this kind of mistakes often. |
Re: HW 2 Problem 6
@dsvav,
Thank you for your comments. You were right, I did forget to change the sample number (N) to 1000. But that doesn't change the result. It is possible that Eout is not the same as Ein although that is what we want. Indeed, we are applying linear regression to a random data that it hasn't seen before; hence, the larger deviation between Eout and Ein. |
Re: HW 2 Problem 6
@MLearning
When I compute the difference between E_in and E_out I get the difference to be around 0.01. I still think difference should not be significant , does not this comes from Hoeffding Inequality ? Also since we are suppressing the "very bad event happening" and "very good event happening" by taking average over 1000 runs , so E_out should track E_in. This is my understanding , there is good chance that I am wrong :D By the way what is the difference you are getting ? |
Re: HW 2 Problem 6
Quote:
|
All times are GMT -7. The time now is 06:27 PM. |
Powered by vBulletin® Version 3.8.3
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.
The contents of this forum are to be used ONLY by readers of the Learning From Data book by Yaser S. Abu-Mostafa, Malik Magdon-Ismail, and Hsuan-Tien Lin, and participants in the Learning From Data MOOC by Yaser S. Abu-Mostafa. No part of these contents is to be communicated or made accessible to ANY other person or entity.