The loss rank principle for model selection
Date
2007
Authors
Hutter, Marcus
Journal Title
Journal ISSN
Volume Title
Publisher
Springer
Abstract
A key issue in statistics and machine learning is to automatically select the "right" model complexity, e.g. the number of neighbors to be averaged over in k nearest neighbor (kNN) regression or the polynomial degree in regression with polynomials. We suggest a novel principle (LoRP) for model selection in regression and classification. It is based on the loss rank, which counts how many other (fictitious) data would be fitted better. LoRP selects the model that has minimal loss rank. Unlike most penalized maximum likelihood variants (AIC,BIC,MDL), LoRP only depends on the regression functions and the loss function. It works without a stochastic noise model, and is directly applicable to any non-parametric regressor, like kNN.
Description
Keywords
Keywords: Computational complexity; Curve fitting; Learning systems; Polynomials; Regression analysis; Stochastic models; K nearest neighbor (kNN) regression; Loss rank principle; Model complexity; Model selection; Model checking
Citation
Collections
Source
Type
Book chapter
Book Title
Learning Theory
Entity type
Access Statement
License Rights
DOI
Restricted until
2037-12-31