Skip to content

Exploration of hardware efficient efficient training algorithms

Notifications You must be signed in to change notification settings

ShihuiYin/HWEfficientTraining

Repository files navigation

FP8 training

The FP8 training scheme is forked from QPytorch

IBM8.pytorch

QPyTorch simulation for the NeurIPS 2018 paper, Training Deep Neural Networks with 8-bit Floating Point Numbers. Note that we simulate the numerical behavior of using the proposed 8-bit and 16-bit floating point number but not the chunk-based accumulation. Accumulation is still done in single precision. Also, due to the absence of an official reference, the hyperparameters used in this repo are different from the paper.

Citation

./example.sh

Citation

If you find this simulation useful, please cite the original paper,

@incollection{NIPS2018_7994,
title = {Training Deep Neural Networks with 8-bit Floating Point Numbers},
author = {Wang, Naigang and Choi, Jungwook and Brand, Daniel and Chen, Chia-Yu and Gopalakrishnan, Kailash},
booktitle = {Advances in Neural Information Processing Systems 31},
editor = {S. Bengio and H. Wallach and H. Larochelle and K. Grauman and N. Cesa-Bianchi and R. Garnett},
pages = {7675--7684},
year = {2018},
publisher = {Curran Associates, Inc.},
url = {http://papers.nips.cc/paper/7994-training-deep-neural-networks-with-8-bit-floating-point-numbers.pdf}
}

About

Exploration of hardware efficient efficient training algorithms

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published