LFD Book Forum  

Go Back   LFD Book Forum > Course Discussions > Online LFD course > Homework 7

Reply
 
Thread Tools Display Modes
  #1  
Old 02-25-2013, 03:54 PM
ilya239 ilya239 is offline
Senior Member
 
Join Date: Jul 2012
Posts: 58
Question large margins and the growth function

I'm trying to understand why the large-margin requirement affects the growth function. For any size margin, we can find three points far enough from each other that they are shattered by perceptrons with at least that margin. How, then, is the growth function at n=3 less than 8?

Or, put differently: the growth function is a property of the hypothesis set. The large-margin requirement does not remove any hypotheses from the hypothesis set; it just prevents us from using particular hypotheses for particular training sets. This limitation is a property of the learning algorithm, but the VC analysis was independent of learning algorithm. If the hypothesis set has not changed, how can the growth function change?

Reply With Quote
  #2  
Old 02-25-2013, 04:11 PM
yaser's Avatar
yaser yaser is offline
Caltech
 
Join Date: Aug 2009
Location: Pasadena, California, USA
Posts: 1,477
Default Re: large margins and the growth function

Quote:
Originally Posted by ilya239 View Post
I'm trying to understand why the large-margin requirement affects the growth function. For any size margin, we can find three points far enough from each other that they are shattered by perceptrons with at least that margin.
The generalization result that relates to the margin assumes all the points lie within a limited-size region (so the value of the margin is meaningful relative to that).

Quote:
the growth function is a property of the hypothesis set. The large-margin requirement does not remove any hypotheses from the hypothesis set; it just prevents us from using particular hypotheses for particular training sets. This limitation is a property of the learning algorithm, but the VC analysis was independent of learning algorithm.
You are right. In this result, some liberty is taken in distinguishing the hypothesis set from the learning algorithm. The same liberty is also taken in the case of nearest-neighbor classifiers (a simple model that will be mentioned briefly in Lecture 16).

I would take the margin-based arguments for generalization as just motivational, and rely on the generalization results that relate to the number of support vectors.
__________________
Where everyone thinks alike, no one thinks very much
Reply With Quote
  #3  
Old 02-26-2013, 11:17 AM
ilya239 ilya239 is offline
Senior Member
 
Join Date: Jul 2012
Posts: 58
Default Re: large margins and the growth function

Quote:
Originally Posted by yaser View Post
The generalization result that relates to the margin assumes all the points lie within a limited-size region (so the value of the margin is meaningful relative to that).
Ah, thanks, makes sense. In real problems, parameters often do have only limited ranges.

Quote:
Originally Posted by yaser View Post
some liberty is taken in distinguishing the hypothesis set from the learning algorithm. The same liberty is also taken in the case of nearest-neighbor classifiers (a simple model that will be mentioned briefly in Lecture 16).
I see now -- it's like with heavy regularization, if the cost of complex hypotheses becomes so exorbitant that they're never chosen in practice, you could just as well say they've been removed from the hypothesis set. If they're rarely chosen, it's as if each hypothesis in the set has a "weight". I wonder if there exists a "fuzzy" version of VC analysis that formalizes this "middle ground" created by margins or regularizers.

Thanks for help!
Reply With Quote
Reply

Tags
growth function, theory, vc dimension

Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off

Forum Jump


All times are GMT -7. The time now is 12:55 PM.


Powered by vBulletin® Version 3.8.3
Copyright ©2000 - 2019, Jelsoft Enterprises Ltd.
The contents of this forum are to be used ONLY by readers of the Learning From Data book by Yaser S. Abu-Mostafa, Malik Magdon-Ismail, and Hsuan-Tien Lin, and participants in the Learning From Data MOOC by Yaser S. Abu-Mostafa. No part of these contents is to be communicated or made accessible to ANY other person or entity.