Skip navigation
Skip navigation

Essays on Robust Model Selection and Model Averaging for Linear Models

Chang, Le

Description

Model selection is central to all applied statistical work. Selecting the variables for use in a regression model is one important example of model selection. This thesis is a collection of essays on robust model selection procedures and model averaging for linear regression models. In the first essay, we propose robust Akaike information criteria (AIC) for MM-estimation and an adjusted robust scale based AIC for M and MM-estimation. Our proposed model...[Show more]

dc.contributor.authorChang, Le
dc.date.accessioned2018-01-11T05:08:11Z
dc.date.available2018-01-11T05:08:11Z
dc.identifier.otherb48528535
dc.identifier.urihttp://hdl.handle.net/1885/139176
dc.description.abstractModel selection is central to all applied statistical work. Selecting the variables for use in a regression model is one important example of model selection. This thesis is a collection of essays on robust model selection procedures and model averaging for linear regression models. In the first essay, we propose robust Akaike information criteria (AIC) for MM-estimation and an adjusted robust scale based AIC for M and MM-estimation. Our proposed model selection criteria can maintain their robust properties in the presence of a high proportion of outliers and the outliers in the covariates. We compare our proposed criteria with other robust model selection criteria discussed in previous literature. Our simulation studies demonstrate a significant outperformance of robust AIC based on MM-estimation in the presence of outliers in the covariates. The real data example also shows a better performance of robust AIC based on MM-estimation. The second essay focuses on robust versions of the ``Least Absolute Shrinkage and Selection Operator" (lasso). The adaptive lasso is a method for performing simultaneous parameter estimation and variable selection. The adaptive weights used in its penalty term mean that the adaptive lasso achieves the oracle property. In this essay, we propose an extension of the adaptive lasso named the Tukey-lasso. By using Tukey's biweight criterion, instead of squared loss, the Tukey-lasso is resistant to outliers in both the response and covariates. Importantly, we demonstrate that the Tukey-lasso also enjoys the oracle property. A fast accelerated proximal gradient (APG) algorithm is proposed and implemented for computing the Tukey-lasso. Our extensive simulations show that the Tukey-lasso, implemented with the APG algorithm, achieves very reliable results, including for high-dimensional data where p>n. In the presence of outliers, the Tukey-lasso is shown to offer substantial improvements in performance compared to the adaptive lasso and other robust implementations of the lasso. Real data examples further demonstrate the utility of the Tukey-lasso. In many statistical analyses, a single model is used for statistical inference, ignoring the process that leads to the model being selected. To account for this model uncertainty, many model averaging procedures have been proposed. In the last essay, we propose an extension of a bootstrap model averaging approach, called bootstrap lasso averaging (BLA). BLA utilizes the lasso for model selection. This is in contrast to other forms of bootstrap model averaging that use AIC or Bayesian information criteria (BIC). The use of the lasso improves the computation speed and allows BLA to be applied even when the number of variables p is larger than the sample size n. Extensive simulations confirm that BLA has outstanding finite sample performance, in terms of both variable and prediction accuracies, compared with traditional model selection and model averaging methods. Several real data examples further demonstrate an improved out-of-sample predictive performance of BLA.
dc.language.isoen
dc.subjectRobust model selection
dc.subjectRobust AIC
dc.subjectAdaptive lasso
dc.subjectRobust estimation
dc.subjectAccelerated proximal gradient algorithm
dc.subjectOracle property
dc.subjectModel averaging
dc.subjectBootstrap
dc.titleEssays on Robust Model Selection and Model Averaging for Linear Models
dc.typeThesis (PhD)
local.contributor.supervisorRoberts, Steven
local.contributor.supervisorcontactsteven.roberts@anu.edu.au
dcterms.valid2017
local.description.notesthe author deposited 11/01/18
local.type.degreeDoctor of Philosophy (PhD)
dc.date.issued2017
local.contributor.affiliationResearch School of Finance, Actuarial Studies & Statistics, College of Business and Economics, The Australian National University
local.identifier.doi10.25911/5d626f47ca582
local.mintdoimint
CollectionsOpen Access Theses

Download

File Description SizeFormat Image
Chang Thesis 2017.pdf1.04 MBAdobe PDFThumbnail


Items in Open Research are protected by copyright, with all rights reserved, unless otherwise indicated.

Updated:  17 November 2022/ Responsible Officer:  University Librarian/ Page Contact:  Library Systems & Web Coordinator