File Download
Supplementary
-
Citations:
- Appears in Collections:
postgraduate thesis: Towards robust image recognition via deep generative classifiers
Title | Towards robust image recognition via deep generative classifiers |
---|---|
Authors | |
Advisors | Advisor(s):Yiu, SM |
Issue Date | 2020 |
Publisher | The University of Hong Kong (Pokfulam, Hong Kong) |
Citation | Wang, X. [王昕]. (2020). Towards robust image recognition via deep generative classifiers. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. |
Abstract | Recent years have witnessed the success of deep neural network models in image recognition. Yet at the same time, they are also surprisingly vulnerable to
malicious inputs, e.g. adversarial examples, corrupted examples, out-of-distribution~(OOD) samples. Previous work usually tries to address one of them and
design specific solutions that are not applicable to other ones. Another important but largely neglected fact is that all the previous work focuses only on
the discriminative classifiers. This can be explained by the fact that the progresses of image recognition during the past several years is completely due to
the discriminative models. Though generative models are believed to be more robust, and demonstrate great success in the relistic synthesis of images, audio
etc, they perform poorly in terms of classification tasks.
In this thesis, we first explore why fully likelihood-based generative models fail in image classification. Second, we propose an end-to-end generative classifier
Supervised Deep Infomax~(SDIM). SDIM models
the generative process on the representations, rather than the raw image pixels. SDIM is able to achieve same level accuracy as discriminative
ones. With the explicit class conditionals in our hands, we could reject illegal inputs by setting thresholds. Our experiments on
adversarial examples and OOD samples show promissing results. Third, instead of training SDIM-based generative classifiers from scratch, we propose SDIM-\emph{logit},
which takes the logits of any discriminative classifier as inputs, and transformes it into a generative one. The training of SDIM-\emph{logit} is very cheap compared to
the full training. Based on increasingly powerful well-trained discriminative classifiers, we see improved results on various malicious inputs detection.
|
Degree | Doctor of Philosophy |
Subject | Optical pattern recognition Computer vision Image processing |
Dept/Program | Computer Science |
Persistent Identifier | http://hdl.handle.net/10722/301040 |
DC Field | Value | Language |
---|---|---|
dc.contributor.advisor | Yiu, SM | - |
dc.contributor.author | Wang, Xin | - |
dc.contributor.author | 王昕 | - |
dc.date.accessioned | 2021-07-16T14:38:41Z | - |
dc.date.available | 2021-07-16T14:38:41Z | - |
dc.date.issued | 2020 | - |
dc.identifier.citation | Wang, X. [王昕]. (2020). Towards robust image recognition via deep generative classifiers. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR. | - |
dc.identifier.uri | http://hdl.handle.net/10722/301040 | - |
dc.description.abstract | Recent years have witnessed the success of deep neural network models in image recognition. Yet at the same time, they are also surprisingly vulnerable to malicious inputs, e.g. adversarial examples, corrupted examples, out-of-distribution~(OOD) samples. Previous work usually tries to address one of them and design specific solutions that are not applicable to other ones. Another important but largely neglected fact is that all the previous work focuses only on the discriminative classifiers. This can be explained by the fact that the progresses of image recognition during the past several years is completely due to the discriminative models. Though generative models are believed to be more robust, and demonstrate great success in the relistic synthesis of images, audio etc, they perform poorly in terms of classification tasks. In this thesis, we first explore why fully likelihood-based generative models fail in image classification. Second, we propose an end-to-end generative classifier Supervised Deep Infomax~(SDIM). SDIM models the generative process on the representations, rather than the raw image pixels. SDIM is able to achieve same level accuracy as discriminative ones. With the explicit class conditionals in our hands, we could reject illegal inputs by setting thresholds. Our experiments on adversarial examples and OOD samples show promissing results. Third, instead of training SDIM-based generative classifiers from scratch, we propose SDIM-\emph{logit}, which takes the logits of any discriminative classifier as inputs, and transformes it into a generative one. The training of SDIM-\emph{logit} is very cheap compared to the full training. Based on increasingly powerful well-trained discriminative classifiers, we see improved results on various malicious inputs detection. | - |
dc.language | eng | - |
dc.publisher | The University of Hong Kong (Pokfulam, Hong Kong) | - |
dc.relation.ispartof | HKU Theses Online (HKUTO) | - |
dc.rights | The author retains all proprietary rights, (such as patent rights) and the right to use in future works. | - |
dc.rights | This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License. | - |
dc.subject.lcsh | Optical pattern recognition | - |
dc.subject.lcsh | Computer vision | - |
dc.subject.lcsh | Image processing | - |
dc.title | Towards robust image recognition via deep generative classifiers | - |
dc.type | PG_Thesis | - |
dc.description.thesisname | Doctor of Philosophy | - |
dc.description.thesislevel | Doctoral | - |
dc.description.thesisdiscipline | Computer Science | - |
dc.description.nature | published_or_final_version | - |
dc.date.hkucongregation | 2021 | - |
dc.identifier.mmsid | 991044390191303414 | - |