Open Research will be unavailable from 3am to 7am on Thursday 4th December 2025 AEDT due to scheduled maintenance.
 

Face Hallucination via Deep Neural Networks.

dc.contributor.authorYu, Xin
dc.date.accessioned2018-12-15T23:37:27Z
dc.date.available2018-12-15T23:37:27Z
dc.date.issued2019
dc.description.abstractWe firstly address aligned low-resolution (LR) face images (i.e. 16X16 pixels) by designing a discriminative generative network, named URDGN. URDGN is composed of two networks: a generative model and a discriminative model. We introduce a pixel-wise L2 regularization term to the generative model and exploit the feedback of the discriminative network to make the upsampled face images more similar to real ones. We present an end-to-end transformative discriminative neural network (TDN) devised for super-resolving unaligned tiny face images. TDN embeds spatial transformation layers to enforce local receptive fields to line-up with similar spatial supports. To upsample noisy unaligned LR face images, we propose decoder-encoder-decoder networks. A transformative discriminative decoder network is employed to upsample and denoise LR inputs simultaneously. Then we project the intermediate HR faces to aligned and noise-free LR faces by a transformative encoder network. Finally, high-quality hallucinated HR images are generated by our second decoder. Furthermore, we present an end-to-end multiscale transformative discriminative neural network (MTDN) to super-resolve unaligned LR face images of different resolutions in a unified framework. We propose a method that explicitly incorporates structural information of faces into the face super-resolution process by using a multi-task convolutional neural network (CNN). Our method not only uses low-level information (i.e. intensity similarity), but also middle-level information (i.e. face structure) to further explore spatial constraints of facial components from LR inputs images. We demonstrate that supplementing residual images or feature maps with additional facial attribute information can significantly reduce the ambiguity in face super-resolution. To explore this idea, we develop an attribute-embedded upsampling network. In this manner, our method is able to super-resolve LR faces by a large upscaling factor while reducing the uncertainty of one-to-many mappings remarkably. We further push the boundaries of hallucinating a tiny, non-frontal face image to understand how much of this is possible by leveraging the availability of large datasets and deep networks. To this end, we introduce a novel Transformative Adversarial Neural Network (TANN) to jointly frontalize very LR out-of-plane rotated face images (including profile views) and aggressively super-resolve them by 8X, regardless of their original poses and without using any 3D information. Besides recovering an HR face images from an LR version, this thesis also addresses the task of restoring realistic faces from stylized portrait images, which can also be regarded as face hallucination.
dc.identifier.otherb59285576
dc.identifier.urihttp://hdl.handle.net/1885/154708
dc.language.isoen_AU
dc.titleFace Hallucination via Deep Neural Networks.
dc.typeThesis (PhD)
local.contributor.affiliationANU College of Engineering and Computer Science, The Australian National University
local.contributor.supervisorHartley, Richard
local.identifier.doi10.25911/5d51433625f11
local.identifier.proquestYes
local.identifier.researcherIDhttps://orcid.org/0000-0002-0269-5649
local.mintdoimint
local.thesisANUonly.author8a796cad-fec1-4a50-a3f1-df79ca0975d4
local.thesisANUonly.keyb81361d6-0cd7-6071-903b-4e6d035b28ad
local.thesisANUonly.title000000015554_TS_1

Downloads

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Xin_PhDthesis.pdf
Size:
25.53 MB
Format:
Adobe Portable Document Format
Description:
Thesis Material