On the Connection between Lp- and Risk Consistency and its Implications on Regularized Kernel Methods
Hannes Köhler; 25(213):1−33, 2024.
Abstract
As a predictor's quality is often assessed by means of its risk, it is natural to regard risk consistency as a desirable property of learning methods, and many such methods have indeed been shown to be risk consistent. The first aim of this paper is to establish the close connection between risk consistency and $L_p$-consistency for a considerably wider class of loss functions than has been done before. The attempt to transfer this connection to shifted loss functions surprisingly reveals that this shift does not reduce the assumptions needed on the underlying probability measure to the same extent as it does for many other results. The results are applied to regularized kernel methods such as support vector machines.
[abs]
[pdf][bib]© JMLR 2024. (edit, beta) |