Quote:
Originally Posted by munchkin
A likelihood analysis for logistic regression yields an expression with (1/probability) and so the terminology for entropy can be applied. This I understand. What I'm not clear on is just what the calculated number means as far as in-sample versus out of sample performance. For the training data the weights calculated are supposed to minimize Ein(w) in accord with the likelihood viewpoint (the training data is the most likely given the final hypothesis if these weights are used). OK. But does it make any sense to compare the calculated Ein for the final weights with the Eout calculated by applying those weights to the test data set?
|
It makes sense because of generalization. For any error measure,

tends to be close to

under conditions similar to those in the VC analysis. For non-binary error measures, the analysis is more involved and the variance of the error measure plays a role, but the same priniciple applies.