![]() |
Q4) h(x) = ax
This question is similar to that in the lectures i.e.,
in the lecture H1 equals h(x) = ax + b Is this question different to the lecture in the respect we shouldn't add "b" (i.e., X0 the bias/intercept) when applying? Or should I treat the same? My confusion is because in many papers etc a bias/intercept is assumed even if not specified i.e., h(x) = ax could be considered the same as h(x) = ax + b |
Re: Q4) h(x) = ax
Thanks for comfirmation, much appreciated :)
|
Re: Q4) h(x) = ax
Is there a best way to minimize the mean-squared error? I am doing gradient descent with a very low learning rate (0.00001) and my solution is diverging! not converging. Is it not feasible to do gradient descent with two points when approximating a sine?
Thanks |
Re: Q4) h(x) = ax
Never mind, I got my solution to converge, though I do not trust my answer. Oh well.
|
Re: Q4) h(x) = ax
Quote:
(since linear regression is basically analytical formula for minimizing mean square error). Also, you can confirm if your g_bar from simulation makes sense by calculate it directly. (calculate expectation of the hypothesis from each (x1,x2) over [-1,1] x [-1,1] ). This involves two integrals but you can plug in the expression to wolfram or mathematica. |
Re: Q4) h(x) = ax
I thought it would simply be (y1/x1 + y2/x2)/2 to find an a that minimizes the mean square error on two points, no?
|
Re: Q4) h(x) = ax
So, in this procedure we:
Pick two points; Find the best slope ![]() Do this N times and average all the ![]() Rather than: Pick two points; Calculate the squared error for those two points as a function of ![]() Do this N times, then find the ![]() Are we doing the first thing here or the second thing? Either way there's a simple analytic solution, but I'm not sure which procedure we're doing. |
Re: Q4) h(x) = ax
Quote:
![]() ![]() ![]() ![]() |
Re: Q4) h(x) = ax
|
All times are GMT -7. The time now is 04:20 PM. |
Powered by vBulletin® Version 3.8.3
Copyright ©2000 - 2021, Jelsoft Enterprises Ltd.
The contents of this forum are to be used ONLY by readers of the Learning From Data book by Yaser S. Abu-Mostafa, Malik Magdon-Ismail, and Hsuan-Tien Lin, and participants in the Learning From Data MOOC by Yaser S. Abu-Mostafa. No part of these contents is to be communicated or made accessible to ANY other person or entity.