LFD Book Forum

LFD Book Forum (http://book.caltech.edu/bookforum/index.php)
-   Homework 7 (http://book.caltech.edu/bookforum/forumdisplay.php?f=136)
-   -   Clarification - Support Vectors decrease no. of features of W (http://book.caltech.edu/bookforum/showthread.php?t=4332)

a.sanyal902 06-05-2013 12:04 PM

Clarification - Support Vectors decrease no. of features of W
 
In Lecture 14, Professor mentions that because only the support vectors count towards W (the rest have alpha=0) which leads to a decrease in the number of features and thus, better generalization.

I'm not sure I got this point because I thought the VC dimension for W would be equal to d, the no. of dimensions of the space, regardless of the number of points being summed. Aren't we just summing the various "x" vectors, multiplied by alpha*y ? How does this decrease the number of features of W?

Thank You!

yaser 06-05-2013 01:14 PM

Re: Clarification - Support Vectors decrease no. of features of W
 
Quote:

Originally Posted by a.sanyal902 (Post 11027)
In Lecture 14, Professor mentions that because only the support vectors count towards W (the rest have alpha=0) which leads to a decrease in the number of features and thus, better generalization.

I'm not sure I got this point because I thought the VC dimension for W would be equal to d, the no. of dimensions of the space, regardless of the number of points being summed. Aren't we just summing the various "x" vectors, multiplied by alpha*y ? How does this decrease the number of features of W?

Thank You!

It decreases the effective number of parameters. If you have 5 support vectors for example, then the vector {\bf w} lives in a 5-dimensional subspace of the d-dimensional space it is formally defined over. There is a bit of liberty taken in this explanation since which vectors would be the support vectors is not known a priori.

Elroch 06-05-2013 01:35 PM

Re: Clarification - Support Vectors decrease no. of features of W
 
I spotted a less sophisticated way of thinking about it which seems helpful to me.

If you merely assume that general points which are associated with a particular target (say +1) are more likely to be near points in a sample D with that target than further away from them, then the bigger the margin, the lower the probability that a general point with target +1 will be wrongly classified (because the bigger the minimum distance from any point in D to a point that would be classified differently).

This ties in quite intuitively with the idea of distances from support vectors (or some sort of transformed distance if kernels are used) being the basis of the hypothesis.

a.sanyal902 06-06-2013 01:14 AM

Re: Clarification - Support Vectors decrease no. of features of W
 
Thank You Professor and Elroch for the answers! Clears things up.


All times are GMT -7. The time now is 01:03 PM.

Powered by vBulletin® Version 3.8.3
Copyright ©2000 - 2019, Jelsoft Enterprises Ltd.
The contents of this forum are to be used ONLY by readers of the Learning From Data book by Yaser S. Abu-Mostafa, Malik Magdon-Ismail, and Hsuan-Tien Lin, and participants in the Learning From Data MOOC by Yaser S. Abu-Mostafa. No part of these contents is to be communicated or made accessible to ANY other person or entity.