On the consistency of output code based learning algorithms for multiclass learning problems

dc.contributor.authorRamaswamy, Harish G.en
dc.contributor.authorBabu, Balaji Srinivasanen
dc.contributor.authorAgarwal, Shivanien
dc.contributor.authorWilliamson, Robert C.en
dc.date.accessioned2026-01-01T08:41:11Z
dc.date.available2026-01-01T08:41:11Z
dc.date.issued2014en
dc.description.abstractA popular approach to solving multiclass learning problems is to reduce them to a set of binary classification problems through some output code matrix: the widely used one-vs-all and all-pairs methods, and the error-correcting output code methods of Dietterich and Bakiri (1995), can all be viewed as special cases of this approach. In this paper, we consider the question of statistical consistency of such methods. We focus on settings where the binary problems are solved by minimizing a binary surrogate loss, and derive general conditions on the binary surrogate loss under which the one-vs-all and all-pairs code matrices yield consistent algorithms with respect to the multiclass 0-1 loss. We then consider general multiclass learning problems defined by a general multiclass loss, and derive conditions on the output code matrix and binary surrogates under which the resulting algorithm is consistent with respect to the target multiclass loss. We also consider probabilistic code matrices, where one reduces a multiclass problem to a set of class probability labeled binary problems, and show that these can yield benefits in the sense of requiring a smaller number of binary problems to achieve overall consistency. Our analysis makes interesting connections with the theory of proper composite losses (Buja et al., 2005; Reid and Williamson, 2010); these play a role in constructing the right 'decoding' for converting the predictions on the binary problems to the final multiclass prediction. To our knowledge, this is the first work that comprehensively studies consistency properties of output code based methods for multiclass learning.en
dc.description.statusPeer-revieweden
dc.format.extent18en
dc.identifier.issn1532-4435en
dc.identifier.scopus84939635212en
dc.identifier.urihttps://hdl.handle.net/1885/733799006
dc.language.isoenen
dc.relation.ispartofseries27th Conference on Learning Theory, COLT 2014en
dc.rightsPublisher Copyright: © 2014 H.G. Ramaswamy, B.S. Babu, S. Agarwal & R.C. Williamson.en
dc.sourceJournal of Machine Learning Researchen
dc.subjectAll-pairsen
dc.subjectConsistencyen
dc.subjectError-correcting output codesen
dc.subjectMulticlass learningen
dc.subjectOne-versus-allen
dc.subjectOutput codesen
dc.subjectProper composite lossesen
dc.titleOn the consistency of output code based learning algorithms for multiclass learning problemsen
dc.typeConference paperen
dspace.entity.typePublicationen
local.bibliographicCitation.lastpage902en
local.bibliographicCitation.startpage885en
local.contributor.affiliationRamaswamy, Harish G.; Indian Institute of Science Bangaloreen
local.contributor.affiliationBabu, Balaji Srinivasan; Indian Institute of Science Bangaloreen
local.contributor.affiliationAgarwal, Shivani; Indian Institute of Science Bangaloreen
local.contributor.affiliationWilliamson, Robert C.; School of Computing, ANU College of Systems and Society, The Australian National Universityen
local.identifier.ariespublicationu4056230xPUB471en
local.identifier.citationvolume35en
local.identifier.pure4eaefc0a-167d-46f1-a399-1bd55e25e2feen
local.identifier.urlhttps://www.scopus.com/pages/publications/84939635212en
local.type.statusPublisheden

Downloads