On the consistency of output code based learning algorithms for multiclass learning problems
| dc.contributor.author | Ramaswamy, Harish G. | en |
| dc.contributor.author | Babu, Balaji Srinivasan | en |
| dc.contributor.author | Agarwal, Shivani | en |
| dc.contributor.author | Williamson, Robert C. | en |
| dc.date.accessioned | 2026-01-01T08:41:11Z | |
| dc.date.available | 2026-01-01T08:41:11Z | |
| dc.date.issued | 2014 | en |
| dc.description.abstract | A popular approach to solving multiclass learning problems is to reduce them to a set of binary classification problems through some output code matrix: the widely used one-vs-all and all-pairs methods, and the error-correcting output code methods of Dietterich and Bakiri (1995), can all be viewed as special cases of this approach. In this paper, we consider the question of statistical consistency of such methods. We focus on settings where the binary problems are solved by minimizing a binary surrogate loss, and derive general conditions on the binary surrogate loss under which the one-vs-all and all-pairs code matrices yield consistent algorithms with respect to the multiclass 0-1 loss. We then consider general multiclass learning problems defined by a general multiclass loss, and derive conditions on the output code matrix and binary surrogates under which the resulting algorithm is consistent with respect to the target multiclass loss. We also consider probabilistic code matrices, where one reduces a multiclass problem to a set of class probability labeled binary problems, and show that these can yield benefits in the sense of requiring a smaller number of binary problems to achieve overall consistency. Our analysis makes interesting connections with the theory of proper composite losses (Buja et al., 2005; Reid and Williamson, 2010); these play a role in constructing the right 'decoding' for converting the predictions on the binary problems to the final multiclass prediction. To our knowledge, this is the first work that comprehensively studies consistency properties of output code based methods for multiclass learning. | en |
| dc.description.status | Peer-reviewed | en |
| dc.format.extent | 18 | en |
| dc.identifier.issn | 1532-4435 | en |
| dc.identifier.scopus | 84939635212 | en |
| dc.identifier.uri | https://hdl.handle.net/1885/733799006 | |
| dc.language.iso | en | en |
| dc.relation.ispartofseries | 27th Conference on Learning Theory, COLT 2014 | en |
| dc.rights | Publisher Copyright: © 2014 H.G. Ramaswamy, B.S. Babu, S. Agarwal & R.C. Williamson. | en |
| dc.source | Journal of Machine Learning Research | en |
| dc.subject | All-pairs | en |
| dc.subject | Consistency | en |
| dc.subject | Error-correcting output codes | en |
| dc.subject | Multiclass learning | en |
| dc.subject | One-versus-all | en |
| dc.subject | Output codes | en |
| dc.subject | Proper composite losses | en |
| dc.title | On the consistency of output code based learning algorithms for multiclass learning problems | en |
| dc.type | Conference paper | en |
| dspace.entity.type | Publication | en |
| local.bibliographicCitation.lastpage | 902 | en |
| local.bibliographicCitation.startpage | 885 | en |
| local.contributor.affiliation | Ramaswamy, Harish G.; Indian Institute of Science Bangalore | en |
| local.contributor.affiliation | Babu, Balaji Srinivasan; Indian Institute of Science Bangalore | en |
| local.contributor.affiliation | Agarwal, Shivani; Indian Institute of Science Bangalore | en |
| local.contributor.affiliation | Williamson, Robert C.; School of Computing, ANU College of Systems and Society, The Australian National University | en |
| local.identifier.ariespublication | u4056230xPUB471 | en |
| local.identifier.citationvolume | 35 | en |
| local.identifier.pure | 4eaefc0a-167d-46f1-a399-1bd55e25e2fe | en |
| local.identifier.url | https://www.scopus.com/pages/publications/84939635212 | en |
| local.type.status | Published | en |