#1




Important details missing
This book is missing a lot of very important details.
First of all, it doesn't even mention the necessity of residual analysis of the errors, and how regression is useless unless that fundamental element is checked. Second, on page 91 how does equation/function 3.8 come about with the information provided? And right below that, why does the function that is minimizes the ML have a (1/N)? As in, where did this (1/N) come from? We can go from a product to a sum of the logs, but we can't just add a 1/N and use the words "We can equivalently ..." to describe the transition. Can someone clarify these issues? 
#2




Re: Important details missing
No one seems to have answered this. I figured I would clear this up a bit so that possible future readers do not get the wrong impression.
Let me skip the first point for now. Regarding getting (3.8) out of the information on page 91, this comes about exactly as the authors pretty much spell out. When y=1, we want to get h(x) = theta(wt x) = theta(y wt x) since y =1. When y=1, we want to get 1h(x) = 1 theta(wt x) = theta( wt x) by the properties of the function theta, and finally we can write this as =theta(y wt x) since y =1. This is nice because we can now combine the two cases in one formula. Regarding the second point, the authors write that they can "equivalently minimize a more convenient quantity", taking a logarithm and putting a 1/N in front. If I want to minimize A, I might as well minimize 2A or 1/2 A, hence the 1/N is not a problem. It is just there for "convenience" (e.g. the typical scale of the resulting numbers), as the authors state. Regarding the first point I have nothing meaningful to say, as I know next to nothing about the subject, but based on the other two points I would take the criticism expressed here by TLMFQS with a rather large grain of salt. 
#3




Re: Important details missing
Quote:
__________________
When one teaches, two learn. 
Thread Tools  
Display Modes  

