Towards Latent Space Optimality for Auto-Encoder Based Generative Models

12/10/2019
by   Arnab Kumar Mondal, et al.
0

The field of neural generative models is dominated by the highly successful Generative Adversarial Networks (GANs) despite their challenges, such as training instability and mode collapse. Auto-Encoders (AE) with regularized latent space provides an alternative framework for generative models, albeit their performance levels have not reached that of GANs. In this work, we identify one of the causes for the under-performance of AE-based models and propose a remedial measure. Specifically, we hypothesize that the dimensionality of the AE model's latent space has a critical effect on the quality of the generated data. Under the assumption that nature generates data by sampling from a "true" generative latent space followed by a deterministic non-linearity, we show that the optimal performance is obtained when the dimensionality of the latent space of the AE-model matches with that of the "true" generative latent space. Further, we propose an algorithm called the Latent Masked Generative Auto-Encoder (LMGAE), in which the dimensionality of the model's latent space is brought closer to that of the "true" generative latent space, via a novel procedure to mask the spurious latent dimensions. We demonstrate through experiments on synthetic and several real-world datasets that the proposed formulation yields generation quality that is better than the state-of-the-art AE-based generative models and is comparable to that of GANs.

READ FULL TEXT
research
05/24/2019

Generative Latent Flow: A Framework for Non-adversarial Image Generation

Generative Adversarial Networks (GANs) have been shown to outperform non...
research
06/10/2020

To Regularize or Not To Regularize? The Bias Variance Trade-off in Regularized AEs

Regularized Auto-Encoders (AE) form a rich class of methods within the l...
research
02/09/2021

Using Deep LSD to build operators in GANs latent space with meaning in real space

Generative models rely on the key idea that data can be represented in t...
research
03/12/2018

Learning the Base Distribution in Implicit Generative Models

Popular generative model learning methods such as Generative Adversarial...
research
11/06/2017

Optimal transport maps for distribution preserving operations on latent spaces of Generative Models

Generative models such as Variational Auto Encoders (VAEs) and Generativ...
research
07/17/2023

Complexity Matters: Rethinking the Latent Space for Generative Modeling

In generative modeling, numerous successful approaches leverage a low-di...
research
04/06/2021

OodGAN: Generative Adversarial Network for Out-of-Domain Data Generation

Detecting an Out-of-Domain (OOD) utterance is crucial for a robust dialo...

Please sign up or login with your details

Forgot password? Click here to reset