Skip to content

Latest commit

 

History

History
16 lines (12 loc) · 1.96 KB

README.md

File metadata and controls

16 lines (12 loc) · 1.96 KB

MLP Semester 2 - Group Coursework Group 079

We introduce a modern retake on the paper "Autoencoding beyond pixels using a learned similarity matrix" by Larsen et al. (paper). By introducing a beta-VAE architecture we improve on the latent space disentangling and present a new beta-VAE-GAN hybrid.

Abstract

Recent research in probabilistic generative models based on deep neural networks has led to image generation systems of a quality previously unseen. We re-explore an algorithm first introduced by that combines Generative Adversarial Networks (GANs) and Variational Autoencoders (VAEs). Its aim is to combine the strength of the two approaches by informing the VAE's loss with the GAN's discriminator, thus creating a feature-wise loss. This helps find disentangledlatent representations, which often better capture features of the data capable of generalizing beyond training samples. These result in improved sample generation. We explore the reproducibility challenges of the algorithm, compare it with VAE and GAN and augment it with $\beta$-VAE, an extension that has been shown to improve the disentanglement of latent representation in VAE. This choice involves the tuning of one key hyperparameter: we avoid the expensive heuristics proposed in the literature and show improved results on our baselines with a simple Bayesian optimization procedure with a 10 % decrease in Frechet Inception Distance score.

Setup

  • Install dependencies from requirements.txt
  • Download data from here

References

  • "Autoencoding beyond pixels using a learned similarity matrix" by Larsen et al. (paper)
  • " beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework." (paper)

Change Log

  • Imported boilerplate from here