LFD Book Forum  

Go Back   LFD Book Forum > Course Discussions > Online LFD course > The Final

Reply
 
Thread Tools Display Modes
  #1  
Old 03-14-2013, 09:16 AM
melipone melipone is offline
Senior Member
 
Join Date: Jan 2013
Posts: 72
Default One-class SVMs

Might be off-topic but I'm not sure where it would go since there is no SVM chapter in the book.

I came across one-class SVMs where support vectors are found w/o class separation. How could that be? What is the hyperplane?
Reply With Quote
  #2  
Old 03-14-2013, 01:52 PM
htlin's Avatar
htlin htlin is offline
NTU
 
Join Date: Aug 2009
Location: Taipei, Taiwan
Posts: 601
Default Re: One-class SVMs

Quote:
Originally Posted by melipone View Post
Might be off-topic but I'm not sure where it would go since there is no SVM chapter in the book.

I came across one-class SVMs where support vectors are found w/o class separation. How could that be? What is the hyperplane?
There are two kinds of common one-class SVM formulations for separating outliers and normal examples without any labeling information. The two are equivalent when using some kernels. They are different in expressing what an "outlier" is.

Perhaps a formulation that's more intuitive is to use the "smallest" hypersphere to bound the normal examples, and then examples falling out of the hypersphere are considered outliers. So roughly, we minimize (the size of the hypersphere + the penalty for being outside the ball).

http://dl.acm.org/citation.cfm?id=960109

The formulation can then be kernelized using the Langrange dual, like the binary SVM discussed in class.

The more popular formulation nowadays consider the "normal" examples as those "far from the origin", and outliers as those close to the origin. In a sense, the observed examples are treated as belonging to the positive class, and the origin is treated as the representative of the negative class. The two classes are separated by a hyperplane. So roughly, we minimize (1 / the margin to the origin + the pentalty for being on the wrong side of the hyperplane). The actual formulation proposed and implemented in solvers like LIBSVM is slightly more sophisticated than that.

http://dl.acm.org/citation.cfm?id=1119749

The formulation can also be kernelized.

Hope this helps.
__________________
When one teaches, two learn.
Reply With Quote
Reply

Thread Tools
Display Modes

Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off

Forum Jump


All times are GMT -7. The time now is 10:11 PM.


Powered by vBulletin® Version 3.8.3
Copyright ©2000 - 2019, Jelsoft Enterprises Ltd.
The contents of this forum are to be used ONLY by readers of the Learning From Data book by Yaser S. Abu-Mostafa, Malik Magdon-Ismail, and Hsuan-Tien Lin, and participants in the Learning From Data MOOC by Yaser S. Abu-Mostafa. No part of these contents is to be communicated or made accessible to ANY other person or entity.