Skip to content

houlu369/Loss-aware-weight-quantization

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Loss-aware-weight-quantization

Implementation of ICLR 2018 paper "Loss-aware Weight Quantization of Deep Networks", tested with GTX TITAN X, python 2.7, theano 0.9.0 and lasagne 0.2.dev1.

This repository is divided in two subrepositories:

  • FNN: enables the reproduction of the FNN results(on MNIST, CIFAR-10, CIFAR-100, SVHN) reported in the article

  • RNN: enables the reproduction of the RNN results(on War and Peace, Linux Kernel, PTB) reported in the article

Requirements This software is implemented on top of the implementation of BinaryConnect and has all the same requirements.

Example training command on War and Peace dataset:

  • training using approximate ternarization method LATa
python warpeace.py --method="LATa" --lr_start=0.002  --len=100
  • training using 3-bit linear quantization method LAQ_linear
python warpeace.py --method="LAQ_linear" --lr_start=0.002  --len=100

If you find loss-aware weight quantization useful in your research, please consider citing the the paper

@InProceedings{hou2017loss,
	title={Loss-aware Binarization of Deep Networks},
	author={Hou, Lu and Yao, Quanming and Kwok, James T.},
	booktitle={International Conference on Learning Representations},
	year={2017}
}

@InProceedings{hou2018loss,
	title={Loss-aware Weight Quantization of Deep Networks},
	author={Hou, Lu and Kwok, James T.},
	booktitle={International Conference on Learning Representations},
	year={2018}
}

@InProceedings{hou2019analysis,
	title={Analysis of Quantized Models},
	author={Hou, Lu and Zhang, Ruiliang and Kwok, James T.},
	booktitle={International Conference on Learning Representations},
	year={2019}
}

About

Implementation of ICLR 2018 paper "Loss-aware Weight Quantization of Deep Networks"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published