Skip navigation
Skip navigation

Adaptivity in Online and Statistical Learning

Mhammedi, Zakaria

Description

Many modern machine learning algorithms, though successful, are still based on heuristics. In a typical application, such heuristics may manifest in the choice of a specific Neural Network structure, its number of parameters, or the learning rate during training. Relying on these heuristics is not ideal from a computational perspective (often involving multiple runs of the algorithm), and can also lead to over-fitting in some cases. This motivates the following question: for which machine...[Show more]

dc.contributor.authorMhammedi, Zakaria
dc.date.accessioned2021-08-08T09:49:33Z
dc.date.available2021-08-08T09:49:33Z
dc.identifier.otherb73316246
dc.identifier.urihttp://hdl.handle.net/1885/243257
dc.description.abstractMany modern machine learning algorithms, though successful, are still based on heuristics. In a typical application, such heuristics may manifest in the choice of a specific Neural Network structure, its number of parameters, or the learning rate during training. Relying on these heuristics is not ideal from a computational perspective (often involving multiple runs of the algorithm), and can also lead to over-fitting in some cases. This motivates the following question: for which machine learning tasks/settings do there exist efficient algorithms that automatically adapt to the best parameters? Characterizing the settings where this is the case and designing corresponding (parameter-free) algorithms within the online learning framework constitutes one of this thesis' primary goals. Towards this end, we develop algorithms for constrained and unconstrained online convex optimization that can automatically adapt to various parameters of interest such as the Lipschitz constant, the curvature of the sequence of losses, and the norm of the comparator. We also derive new performance lower-bounds characterizing the limits of adaptivity for algorithms in these settings. Part of systematizing the choice of machine learning methods also involves having ``certificates'' for the performance of algorithms. In the statistical learning setting, this translates to having (tight) generalization bounds. Adaptivity can manifest here through data-dependent bounds that become small whenever the problem is ``easy''. In this thesis, we provide such data-dependent bounds for the expected loss (the standard risk measure) and other risk measures. We also explore how such bounds can be used in the context of risk-monotonicity.
dc.language.isoen_AU
dc.titleAdaptivity in Online and Statistical Learning
dc.typeThesis (PhD)
local.contributor.supervisorGould, Stephen
local.contributor.supervisorcontactu4971180@anu.edu.au
dc.date.issued2021
local.identifier.doi10.25911/B2WE-2B20
local.identifier.proquestNo
local.thesisANUonly.author4c831dd0-7e19-47a1-b234-c01f18c14bf3
local.thesisANUonly.title000000015845_TC_1
local.thesisANUonly.keyf04f504f-4a0f-f821-1ab0-921ac30607e7
local.mintdoimint
CollectionsOpen Access Theses

Download

File Description SizeFormat Image
Mhammedi Thesis 2021.pdfThesis Material4.09 MBAdobe PDFThumbnail


Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.

Updated:  17 November 2022/ Responsible Officer:  University Librarian/ Page Contact:  Library Systems & Web Coordinator