Author
Abrol, V
Sharma, P
Patra, A
Journal title
IEEE Transactions on Multimedia
DOI
10.1109/TMM.2020.3008053
Volume
23
Last updated
2021-08-20T12:19:31.293+01:00
Page
2153-2161
Abstract
In this paper we propose a conditional generative modelling (CGM) approach for unsupervised disentangled representation learning using variational autoencoder (VAE). CGM employs a multimodal/categorical conditional prior distribution in the latent space to learn global uncertainty in data by modelling the variations at local level. Thus, the proposed framework enforces the model to independently estimate the inherent patterns within each category, which improves the interpretability of the latent representations learned by the VAE model. The evidence lower bound objective for training the generative model is maximized using a mutual information criterion between the global latent categorical variable and the encoded inputs. Further, the approach has a built-in mechanism for bounding the information flow between the encoder and the decoder which addresses the problems of posterior collapse in conventional VAE models. Experiments on a variety of datasets demonstrate that our objective can learn disentangled representations and the proposed approach achieves competitive results on various task such as generative modelling, image classification
and image denoising.
Symplectic ID
1116905
Publication type
Journal Article
Publication date
8 July 2020
Please contact us with feedback and comments about this page. Created on 07 Jul 2020 - 17:30.