Skip to content

Latest commit

 

History

History
41 lines (31 loc) · 1.73 KB

README.md

File metadata and controls

41 lines (31 loc) · 1.73 KB

Variational Bayesian GAN for MNIST Generation

In this repository, we implement our proposed Varational Bayesian GAN (VBGAN) and VBGAN with Wasserstein metric based on auto-encoder based generator and train the model with adversarial learning. The latter extend the Jensen Shannon divergence between data distribution and generating distribution to geometric distance to make the model more robust and genearate better result. To quantify the result of our proposed, we also train our model in semi-supervised learning and the accuracy results show that our model did learning some information from adversarial learning process.

Bayesian GAN credit to https://github.com/vasiloglou/mltrain-nips-2017/blob/master/ben_athiwaratkun/pytorch-bayesgan/Bayesian%20GAN%20in%20PyTorch.ipynb

Bayes by Backbrop credit to https://gist.github.com/vvanirudh/9e30b2f908e801da1bd789f4ce3e7aac

  • Our Model Architecture

VBGAN

VBGAN with Wasserstein metric

Setting

  • Framework:
    • Pytorch 0.4.0
  • Hardware:
    • CPU: Intel Core i7-2600 @3.40 GHz
    • RAM: 20 GB DDR4-2400
    • GPU: GeForce GTX 980

Result of sampling

VBGAN
VBGAN_w

Test accuracy

Test accuracy of our proposed and Bayesian GAN