Show simple item record  

dc.contributor.advisorLim, Nick Jin Sean
dc.contributor.advisorFrank, Eibe
dc.contributor.authorBull, Daniel
dc.date.accessioned2021-12-16T20:29:45Z
dc.date.available2021-12-16T20:29:45Z
dc.date.issued2021
dc.identifier.urihttps://hdl.handle.net/10289/14695
dc.description.abstractCan deep neural networks super-resolve satellite imagery to a high perceptual quality? This thesis explores the juxtaposition between the pixel accuracy and perceptual qualities of super-resolved imagery by comparing and combining a discriminative and a generative network. Rather than solving a theoretical problem, we tackle a real-world low-resolution scenario: Sentinel-2 imagery is super-resolved and evaluated against high-resolution aerial photos as ground truth; this is in contrast to super-resolving previously down-sampled data, which is the methodology used in most other studies. An existing feed-forward network architecture designed for super-resolution, called DeepSUM, is used to super-resolve multiple low-resolution images by a factor of four to obtain a single high-resolution image. DeepSUM is trained using a range of loss functions, to assess the e ect on network accuracy. A novel loss function is created, called variation loss, to help better define edges and textures to create a sharper, perceptually better product. Using an SSIM loss function gives the best result in terms of pixel-based performance. Running DeepSUM alone creates a superior output compared to bicubically up-sampling the input data, but the output is blurry and not photo-realistic. A probabilistic model from the literature, ESRGAN (Enhanced SRGAN), a Generative Adversarial Network, is trained against both raw Sentinel-2 data and the output of DeepSUM. Using ERS-GAN for super-resolution, creates a perceptually better, more realistic looking output. However, the ESRGAN output is less accurate than the DeepSUM output, as measured using pixel-based metrics. Combining ESRGAN with DeepSUM is found to inherit some of the advantages of both approaches. In an end-to-end process, using ESRGAN with the output of DeepSUM trained using variation loss is found to super-resolve an image to better show boundaries, textures and detail.
dc.format.mimetypeapplication/pdf
dc.language.isoen
dc.publisherThe University of Waikato
dc.rightsAll items in Research Commons are provided for private study and research purposes and are protected by copyright with all rights reserved unless otherwise indicated.
dc.subjectSuper-resolution
dc.subject.lcshHigh resolution imaging
dc.subject.lcshRemote-sensing images
dc.titleSuper-resolution of satellite imagery
dc.typeThesis
thesis.degree.grantorThe University of Waikato
thesis.degree.levelMasters
thesis.degree.nameMaster of Science (Research) (MSc(Research))
dc.date.updated2021-11-17T09:25:37Z
pubs.place-of-publicationHamilton, New Zealanden_NZ


Files in this item

Thumbnail

This item appears in the following Collection(s)

Show simple item record