![]() |
|
|
Thread Tools | Display Modes |
#1
|
|||
|
|||
![]()
New paper:
Rethinking statistical learning theory: learning using statistical invariants Vladimir Vapnik and Rauf Izmailov https://link.springer.com/article/10...994-018-5742-0 After doing the Edx course and reading the book and e-chapters, I was of course excited to find Vapnik's possibly important new improved approach to ML. The two main ideas, which I believe are new: 1) vSVM - the SVM but with V-matrix 2) Using Statistical Invariants to improve convergence without extra training samples. LUSI Learning Using Statistical Invariants. Here is the paper, which I have not read, because I can't get a copy that fits my budget (less that the $39 from Springer). Vapnik, V. & Izmailov, R. Mach Learn (2018). https://doi.org/10.1007/s10994-018-5742-0 However, Vapnik has given at least three related lectures in late 2018 (including slides with complicated math), one of which is on youtube here: https://www.youtube.com/watch?v=rNd7PDdhl4c Most intriguing to me, is his comment suggesting that these new techniques are more powerful than deep neural networks. I didn't think that was currently possible in general, or at least not in certain domains, ie. image recognition etc. I'm probably missing something. Can anyone, or the authors ![]() |
Tags |
invariant, svm, vapnik |
Thread Tools | |
Display Modes | |
|
|