We introduce a modern retake on the paper "Autoencoding beyond pixels using a learned similarity matrix" by Larsen et al. (paper). By introducing a beta-VAE architecture we improve on the latent space disentangling and present a new beta-VAE-GAN hybrid.
Recent research in probabilistic generative models based on deep neural networks has led to image generation systems of a quality previously unseen. We re-explore an algorithm first introduced by that combines Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). Its aim is to combine the strength of the two approaches by informing the VAE's loss with the GAN's discriminator, thus creating a feature-wise loss. This helps find disentangledlatent representations, which often better capture features of the data capable of generalizing beyond training samples. These result in improved sample generation. We explore the reproducibility challenges of the algorithm, compare it with VAE and GAN and augment it with
- Install dependencies from requirements.txt
- Download data from here
- "Autoencoding beyond pixels using a learned similarity matrix" by Larsen et al. (paper)
- " beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework." (paper)
- Imported boilerplate from here