New paper:

Rethinking statistical learning theory: learning using statistical invariants

Vladimir Vapnik and Rauf Izmailov

https://link.springer.com/article/10...994-018-5742-0
After doing the Edx course and reading the book and e-chapters, I was of course excited to find Vapnik's possibly important new improved approach to ML.

The two main ideas, which I believe are new:

1) vSVM - the SVM but with V-matrix

2) Using Statistical Invariants to improve convergence without extra training samples. LUSI Learning Using Statistical Invariants.

Here is the paper, which I have not read, because I can't get a copy that fits my budget (less that the $39 from Springer).

Vapnik, V. & Izmailov, R. Mach Learn (2018).

https://doi.org/10.1007/s10994-018-5742-0
However, Vapnik has given at least three related lectures in late 2018 (including slides with complicated math), one of which is on youtube here:

https://www.youtube.com/watch?v=rNd7PDdhl4c
Most intriguing to me, is his comment suggesting that these new techniques are more powerful than deep neural networks. I didn't think that was currently possible in general, or at least not in certain domains, ie. image recognition etc. I'm probably missing something.

Can anyone, or the authors

, comment in detail about how this paper fits into the framework of ideas presented in the book?