Learning without Concentration
We obtain sharp bounds on the convergence rate of Empirical Risk Minimization performed in a convex class and with respect to the squared loss, without any boundedness assumptions on class members or on the target. Rather than resorting to a concentration-based argument, the method relies on a ‘small-ball’ assumption and thus holds for heavy-tailed sampling and heavy-tailed targets. Moreover, the resulting estimates scale correctly with the ‘noise level’ of the problem. When applied to the...[Show more]
|Collections||ANU Research Publications|
|Source:||JMLR: Workshop and Conference Proceedings|
Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.