Adaptivity in Online and Statistical Learning
Download (4.09 MB)
-
Altmetric Citations
Description
Many modern machine learning algorithms, though successful, are still based on heuristics. In a typical application, such heuristics may manifest in the choice of a specific Neural Network structure, its number of parameters, or the learning rate during training. Relying on these heuristics is not ideal from a computational perspective (often involving multiple runs of the algorithm), and can also lead to over-fitting in some cases. This motivates the following question: for which machine...[Show more]
dc.contributor.author | Mhammedi, Zakaria | |
---|---|---|
dc.date.accessioned | 2021-08-08T09:49:33Z | |
dc.date.available | 2021-08-08T09:49:33Z | |
dc.identifier.other | b73316246 | |
dc.identifier.uri | http://hdl.handle.net/1885/243257 | |
dc.description.abstract | Many modern machine learning algorithms, though successful, are still based on heuristics. In a typical application, such heuristics may manifest in the choice of a specific Neural Network structure, its number of parameters, or the learning rate during training. Relying on these heuristics is not ideal from a computational perspective (often involving multiple runs of the algorithm), and can also lead to over-fitting in some cases. This motivates the following question: for which machine learning tasks/settings do there exist efficient algorithms that automatically adapt to the best parameters? Characterizing the settings where this is the case and designing corresponding (parameter-free) algorithms within the online learning framework constitutes one of this thesis' primary goals. Towards this end, we develop algorithms for constrained and unconstrained online convex optimization that can automatically adapt to various parameters of interest such as the Lipschitz constant, the curvature of the sequence of losses, and the norm of the comparator. We also derive new performance lower-bounds characterizing the limits of adaptivity for algorithms in these settings. Part of systematizing the choice of machine learning methods also involves having ``certificates'' for the performance of algorithms. In the statistical learning setting, this translates to having (tight) generalization bounds. Adaptivity can manifest here through data-dependent bounds that become small whenever the problem is ``easy''. In this thesis, we provide such data-dependent bounds for the expected loss (the standard risk measure) and other risk measures. We also explore how such bounds can be used in the context of risk-monotonicity. | |
dc.language.iso | en_AU | |
dc.title | Adaptivity in Online and Statistical Learning | |
dc.type | Thesis (PhD) | |
local.contributor.supervisor | Gould, Stephen | |
local.contributor.supervisorcontact | u4971180@anu.edu.au | |
dc.date.issued | 2021 | |
local.identifier.doi | 10.25911/B2WE-2B20 | |
local.identifier.proquest | No | |
local.thesisANUonly.author | 4c831dd0-7e19-47a1-b234-c01f18c14bf3 | |
local.thesisANUonly.title | 000000015845_TC_1 | |
local.thesisANUonly.key | f04f504f-4a0f-f821-1ab0-921ac30607e7 | |
local.mintdoi | mint | |
Collections | Open Access Theses |
Download
File | Description | Size | Format | Image |
---|---|---|---|---|
Mhammedi Thesis 2021.pdf | Thesis Material | 4.09 MB | Adobe PDF | ![]() |
Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.
Updated: 17 November 2022/ Responsible Officer: University Librarian/ Page Contact: Library Systems & Web Coordinator