Skip to content

Deep Compression follows a pipeline procedure involving Pruning, Quantization and Huffman Coding to compress deep neural network models sizes by 35x-40x. We implemented (pseudo) pruning similar to the first stage of the pipeline on LeNet-5 and decreased it's download bandwidth by 65% (after (tensorlfow) quantization & gzip compression).

Notifications You must be signed in to change notification settings

honorpeter/Caffe-DeepCompression

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

32 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Caffe-DeepCompression

Read the readmes in the respective folders

About

Deep Compression follows a pipeline procedure involving Pruning, Quantization and Huffman Coding to compress deep neural network models sizes by 35x-40x. We implemented (pseudo) pruning similar to the first stage of the pipeline on LeNet-5 and decreased it's download bandwidth by 65% (after (tensorlfow) quantization & gzip compression).

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%