File Download

There are no files associated with this item.

  Links for fulltext
     (May Require Subscription)
Supplementary

Conference Paper: Disentangled representation learning for controllable image synthesis: An information-theoretic perspective

TitleDisentangled representation learning for controllable image synthesis: An information-theoretic perspective
Authors
Issue Date2020
Citation
Proceedings - International Conference on Pattern Recognition, 2020, p. 10042-10049 How to Cite?
AbstractIn this paper, we look into the problem of disentangled representation learning and controllable image synthesis in a deep generative model. We develop an encoder-decoder architecture for a variant of the Variational Auto-Encoder (VAE) with two latent codes z1 and z2. Our framework uses z2 to capture specified factors of variation while z1 captures the complementary factors of variation. To this end, we analyze the learning problem from the perspective of multivariate mutual information, derive optimizable lower bounds of the conditional mutual information in the image synthesis processes and incorporate them into the training objective. We validate our method empirically on the Color MNIST dataset and the CelebA dataset by showing controllable image syntheses. Our proposed paradigm is simple yet effective and is applicable to many situations, including those where there is not an explicit factorization of features available, or where the features are non-categorical.
Persistent Identifierhttp://hdl.handle.net/10722/327775
ISSN
2023 SCImago Journal Rankings: 0.584
ISI Accession Number ID

 

DC FieldValueLanguage
dc.contributor.authorTang, Shichang-
dc.contributor.authorZhou, Xu-
dc.contributor.authorHe, Xuming-
dc.contributor.authorMa, Yi-
dc.date.accessioned2023-05-08T02:26:43Z-
dc.date.available2023-05-08T02:26:43Z-
dc.date.issued2020-
dc.identifier.citationProceedings - International Conference on Pattern Recognition, 2020, p. 10042-10049-
dc.identifier.issn1051-4651-
dc.identifier.urihttp://hdl.handle.net/10722/327775-
dc.description.abstractIn this paper, we look into the problem of disentangled representation learning and controllable image synthesis in a deep generative model. We develop an encoder-decoder architecture for a variant of the Variational Auto-Encoder (VAE) with two latent codes z1 and z2. Our framework uses z2 to capture specified factors of variation while z1 captures the complementary factors of variation. To this end, we analyze the learning problem from the perspective of multivariate mutual information, derive optimizable lower bounds of the conditional mutual information in the image synthesis processes and incorporate them into the training objective. We validate our method empirically on the Color MNIST dataset and the CelebA dataset by showing controllable image syntheses. Our proposed paradigm is simple yet effective and is applicable to many situations, including those where there is not an explicit factorization of features available, or where the features are non-categorical.-
dc.languageeng-
dc.relation.ispartofProceedings - International Conference on Pattern Recognition-
dc.titleDisentangled representation learning for controllable image synthesis: An information-theoretic perspective-
dc.typeConference_Paper-
dc.description.naturelink_to_subscribed_fulltext-
dc.identifier.doi10.1109/ICPR48806.2021.9411925-
dc.identifier.scopuseid_2-s2.0-85110472024-
dc.identifier.spage10042-
dc.identifier.epage10049-
dc.identifier.isiWOS:000681331402073-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats