Implemented Variational Autoencoder generative model in Keras for image generation and its latent space visualization on MNIST and CIFAR10 datasets
Variational AutoEncoder - Keras implementation on mnist and cifar10 datasets
code is highly inspired from keras examples of vae : ,
(source files contains some code duplication)
network architecture |
---|
![]() |
ld_<latent_dim>_id_<intermediate_dim>_e_<epochs>_<vae/encoder/decoder>.h5
where <latent_dim>
is number of latent dimensions, <intermediate_dim>
is number of neurons in hidden layer and <epochs>
is number of training epochsld_<latent_dim>_id_<intermediate_dim>_e_<epochs>_history.pkl
mnist_params.py
mnist_2d_latent_space_and_generate.py
but it is for 3d latent spacemnist_params.py
latent space | uniform sampling |
---|---|
![]() |
![]() |
uniform sampling | random sampling |
---|---|
![]() |
![]() |
encoder |
---|
![]() |
decoder |
---|
![]() |
implementation structure is same as mnist files
25 epochs | 50 epochs | 75 epochs |
---|---|---|
![]() |
![]() |
![]() |
600 epochs |
---|
![]() |
caltech101_<sz>_train.py
and caltech101_<sz>_generate.py
(where sz
is the size of input image - here the training was done for two sizes - 92*92 and 128*128) are same as cifar10 dataset filesdataset
directory,