![]() |
#21
|
|||
|
|||
![]()
It is difficult to look at this homework and see how it applies to the real world. Being a non-US citizen, "symmetry" and "intensity", of zip codes means nothing? I assumed it probably meant nothing to anybody? So in reality given an unknown dataset and the question do a "0 versus all" you would probably apply weighting as you don't know anything about the data. I.e., consider a study where you are looking at heart-attack patients, it wouldn't be unreasonable to have 99% -1s and 1% 1s. This is a real issue for SVMs as these skewed classes will not focus in well on the 1s. So weighting is a good thing to use in the right circumstances.
Going back to the homework, the dataset probably has a similar distribution of 0s, 1s, 2s... 9s so in reality you probably wouldn't be asked by a client to "do 0 versus all", a real ask would be to "predict what classes each row belongs to". In which case the SVM should work well. But for the homework we are explicitly told to do this for binary which is fine enough - proof of concept and all that. |
#22
|
|||
|
|||
![]()
This data was introduced in lecture 3 as features of handwritten digits.
|
#23
|
|||
|
|||
![]()
Ah, that's interesting. All my points still hold though.
I do agree though that for answering this homework "-wi" should not be used (as I wrote in my first post on this thread (#16)) as I'm sure it would have been specifically mentioned otherwise. |
#24
|
|||
|
|||
![]()
Apart from:
0 versus all and 1 versus all all the other "versus all" SVM class models for the training set are populated solely of -1s. This to me says either: 1) I am doing something wrong. 2) the skewed class problem I mentioned earlier is apparent. Can anyone confirm they get the same results? I'd hate to fall at the first hurdle! |
#25
|
||||
|
||||
![]() Quote:
Also, FYI, the following paper defends that one-versus-all can still be useful when applied with care. http://jmlr.csail.mit.edu/papers/vol.../rifkin04a.pdf Hope this helps.
__________________
When one teaches, two learn. |
#26
|
|||
|
|||
![]()
Thanks very much for the link to that pdf, have been looking for something like that would describe the options in more detail... with pros and cons etc... Many thanks
![]() I did see earlier another option in LIBSVM for the "svm_type": 2 -- one-class SVM ...from my intial research this sounds like it maybe similar to what you mentioned with regards to "regression perspective" instead of classification? This would certainly be another angle to look at instead of re-weighting data so that the classifiers are balanced (the -wi option I was going on about)... lots of further research... ![]() ![]() |
#27
|
|||
|
|||
![]() Quote:
|
#28
|
||||
|
||||
![]() Quote:
For the regression perspective of one-versus-all, it may be worth referring to one of the earliest papers on one-versus-all, in which they use "regression decision tree" (called model tree) for one-versus-all. @article{EF98, title={{Using model trees for classification}}, author={Frank, E. and Wang, Y. and Inglis, S. and Holmes, G. and Witten, I.H. }, journal={Machine Learning}, volume={32}, number={1}, pages={63--76}, year={1998}, publisher={Springer} } Hope this helps.
__________________
When one teaches, two learn. |
#29
|
|||
|
|||
![]()
Thanks again, managed to find a copy here:
http://www.springerlink.com/content/...6/fulltext.pdf My reading list is getting pretty long ![]() |
#30
|
|||
|
|||
![]() Quote:
If you look at the training data, it has no +1/-1 label. I think you think to transform the digits in to binary labels data. The classifier expected to see +1/-1 labels in the training data; hence, very low cross-validation accuracy when you apply SVM on the raw data. |
![]() |
Thread Tools | |
Display Modes | |
|
|