Skip to content

dspoka/sa-vae

 
 

Repository files navigation

Semi-Amortized Variational Autoencoders

Code for the paper:
Semi-Amortized Variational Autoencoders
Yoon Kim, Sam Wiseman, Andrew Miller, David Sontag, Alexander Rush

Dependencies

The code was tested in python 3.6 and pytorch 0.2. We also require the h5py package.

Data

The raw datasets can be downloaded from here.

Text experiments use the Yahoo dataset from Yang et al. 2017, which is itself derived from Zhang et al. 2015.

Image experiments use the OMNIGLOT dataset Lake et al. 2015 with preprocessing from Burda et al. 2015.

Please cite the original papers when using the data.

Text

After downloading the data, run

python preprocess_text.py --trainfile data/yahoo/train.txt --valfile data/yahoo/val.txt
--testfile data/yahoo/test.txt --outputfile data/yahoo/yahoo

This will create the *.hdf5 files (data tensors) to be used by the model, as well as the *.dict file which contains the word-to-integer mapping for each word.

The basic model command is

python train_text.py --train_file data/yahoo/yahoo-train.hdf5 --val_file data/yahoo/yahoo-val.hdf5
--gpu 1 --checkpoint_path model-path

where model-path is the path to save the best model and the *.hdf5 files are obtained from running preprocess_text.py. You can specify which GPU to use by changing the input to the --gpu command.

To train the various models, add the following:

  • Autoregressive (i.e. language model): --model autoreg
  • VAE: --model vae
  • SVI: --model svi --svi_steps 20 --train_n2n 0
  • VAE+SVI: --model savae --svi_steps 20 --train_n2n 0 --train_kl 0
  • VAE+SVI+KL: --model savae --svi_steps 20 --train_n2n 0 --train_kl 1
  • SA-VAE: --model savae --svi_steps 20 --train_n2n 1

Number of SVI steps can be changed with the --svi_steps command.

To evaluate, run

python train_text.py --train_from model-path --test_file data/yahoo/yahoo-test.hdf5 --test 1 --gpu 1

Make sure the append the relevant model configuration at test time too.

Images

After downloading the data, run

python preprocess_img.py --raw_file data/omniglot/chardata.mat --output data/omniglot/omniglot.pt

To train, the basic command is

python train_img.py --data_file data/omniglot/omniglot.pt --gpu 1 --checkpoint_path model-path

To train the various models, add the following:

  • Autoregressive (i.e. Gated PixelCNN): --model autoreg
  • VAE: --model vae
  • SVI: --model svi --svi_steps 20
  • VAE+SVI: --model savae --svi_steps 20 --train_n2n 0 --train_kl 0
  • VAE+SVI+KL: --model savae --svi_steps 20 --train_n2n 0 --train_kl 1
  • SA-VAE: --model savae --svi_steps 20 --train_n2n 1

To evaluate, run

python train_img.py --train_from model-path --test 1 --gpu 1

Make sure the append the relevant model configuration at test time too.

Acknowledgements

Some of our code is based on VAE with a VampPrior.

License

MIT

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%