File Download
Supplementary

postgraduate thesis: Context and adversarial learning in low-level vision

TitleContext and adversarial learning in low-level vision
Authors
Advisors
Advisor(s):Yu, Y
Issue Date2020
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Li, H. [李灝峰]. (2020). Context and adversarial learning in low-level vision. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractIn recent years, with increasing network bandwidth and the popularity of portable cameras, massive video blogs and photographs have been produced and uploaded to webcam sites and social media. It is important to develop automated, intelligent and robust tools for digital image processing. How to harvest and exploit spatio-temporal contexts in visual data to achieve high-quality image processing is also an essential problem. On one hand, adversarial learning based attacks could threaten deep learning based models. On the other hand, generative adversarial networks can synthesize realistic image patches via adversarial learning. It is necessary to improve existing image processing tools with generative adversarial networks and against adversarial attacks. Thus, we propose three solutions based on context and adversarial learning for three low-level vision tasks: video salient object detection, robust salient object detection against adversarial attacks and image inpainting. For detecting salient objects in a video, we introduce a motion guided attention neural network that is integrated with a family of motion guided attention modules. The proposed motion guided attention modules model how motions computed from temporal contexts determine object saliency. The proposed neural network contains a motion branch and an appearance branch, which work by computing intermediate representations and a final saliency map for a pair of optical flow and color images. Our proposed algorithm significantly surpasses the state-of-the-art video salient object detection methods. For robust salient object detection resistant to adversarial samples, we develop a novel generic framework that boosts the robustness of arbitrary salient object detection models based on fully convolutional networks. The proposed salient object detection framework adopts a segment-wise shielding component to destroy adversarial perturbations in an input image. The framework includes a context-ware restoration component that uses spatial contextual correlations to refine saliency maps. Experimental results demonstrate that our proposed salient object detection framework performs better than other defenses against adversarial attacks. For semantic image completion, we propose a context-aware semantic inpainting algorithm and two novel evaluation metrics. The proposed algorithm consists of a fully convolutional generative network and a context-aware joint loss function. The generative network resorts to fully convolutional architecture without fully connected bottlenecks to maintain structural features. The joint loss function forces the proposed network to synthesize contents with the same semantic as the surrounding context. Both quantitative and qualitative comparisons suggest that our proposed algorithm obtains state-of-the-art performance. The proposed metrics can better rate an over-smoothed result and the overall semantic of an image.
DegreeDoctor of Philosophy
SubjectImage processing - Digital techniques
Dept/ProgramComputer Science
Persistent Identifierhttp://hdl.handle.net/10722/283128

 

DC FieldValueLanguage
dc.contributor.advisorYu, Y-
dc.contributor.authorLi, Haofeng-
dc.contributor.author李灝峰-
dc.date.accessioned2020-06-10T01:02:15Z-
dc.date.available2020-06-10T01:02:15Z-
dc.date.issued2020-
dc.identifier.citationLi, H. [李灝峰]. (2020). Context and adversarial learning in low-level vision. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/283128-
dc.description.abstractIn recent years, with increasing network bandwidth and the popularity of portable cameras, massive video blogs and photographs have been produced and uploaded to webcam sites and social media. It is important to develop automated, intelligent and robust tools for digital image processing. How to harvest and exploit spatio-temporal contexts in visual data to achieve high-quality image processing is also an essential problem. On one hand, adversarial learning based attacks could threaten deep learning based models. On the other hand, generative adversarial networks can synthesize realistic image patches via adversarial learning. It is necessary to improve existing image processing tools with generative adversarial networks and against adversarial attacks. Thus, we propose three solutions based on context and adversarial learning for three low-level vision tasks: video salient object detection, robust salient object detection against adversarial attacks and image inpainting. For detecting salient objects in a video, we introduce a motion guided attention neural network that is integrated with a family of motion guided attention modules. The proposed motion guided attention modules model how motions computed from temporal contexts determine object saliency. The proposed neural network contains a motion branch and an appearance branch, which work by computing intermediate representations and a final saliency map for a pair of optical flow and color images. Our proposed algorithm significantly surpasses the state-of-the-art video salient object detection methods. For robust salient object detection resistant to adversarial samples, we develop a novel generic framework that boosts the robustness of arbitrary salient object detection models based on fully convolutional networks. The proposed salient object detection framework adopts a segment-wise shielding component to destroy adversarial perturbations in an input image. The framework includes a context-ware restoration component that uses spatial contextual correlations to refine saliency maps. Experimental results demonstrate that our proposed salient object detection framework performs better than other defenses against adversarial attacks. For semantic image completion, we propose a context-aware semantic inpainting algorithm and two novel evaluation metrics. The proposed algorithm consists of a fully convolutional generative network and a context-aware joint loss function. The generative network resorts to fully convolutional architecture without fully connected bottlenecks to maintain structural features. The joint loss function forces the proposed network to synthesize contents with the same semantic as the surrounding context. Both quantitative and qualitative comparisons suggest that our proposed algorithm obtains state-of-the-art performance. The proposed metrics can better rate an over-smoothed result and the overall semantic of an image.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshImage processing - Digital techniques-
dc.titleContext and adversarial learning in low-level vision-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineComputer Science-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2020-
dc.identifier.mmsid991044242097003414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats