Efficient hold-out for subset of regressors
Date
Authors
Pahikkala, Tapio
Suominen, Hanna
Boberg, Jorma
Salakoski, Tapio
Journal Title
Journal ISSN
Volume Title
Publisher
Access Statement
Abstract
Hold-out and cross-validation are among the most useful methods for model selection and performance assessment of machine learning algorithms. In this paper, we present a computationally efficient algorithm for calculating the hold-out performance for sparse regularized least-squares (RLS) in case the method is already trained with the whole training set. The computational complexity of performing the hold-out is O(|H|3 + |H|2n), where |H| is the size of the hold-out set and n is the number of basis vectors. The algorithm can thus be used to calculate various types of cross-validation estimates effectively. For example, when m is the number of training examples, the complexities of N-fold and leave-one-out cross-validations are O(m 3/N2 + (m2n)/N) and O(mn), respectively. Further, since sparse RLS can be trained in O(mn2) time for several regularization parameter values in parallel, the fast hold-out algorithm enables efficient selection of the optimal parameter value.
Description
Keywords
Citation
Collections
Source
Type
Book Title
Adaptive and Natural Computing Algorithms - 9th International Conference, ICANNGA 2009, Revised Selected Papers
Entity type
Publication