Back to Search Start Over

Exploring the Potential of Low-bit Training of Convolutional Neural Networks

Authors :
Zhong, Kai
Ning, Xuefei
Dai, Guohao
Zhu, Zhenhua
Zhao, Tianchen
Zeng, Shulin
Wang, Yu
Yang, Huazhong
Zhong, Kai
Ning, Xuefei
Dai, Guohao
Zhu, Zhenhua
Zhao, Tianchen
Zeng, Shulin
Wang, Yu
Yang, Huazhong
Publication Year :
2020

Abstract

In this work, we propose a low-bit training framework for convolutional neural networks, which is built around a novel multi-level scaling (MLS) tensor format. Our framework focuses on reducing the energy consumption of convolution operations by quantizing all the convolution operands to low bit-width format. Specifically, we propose the MLS tensor format, in which the element-wise bit-width can be largely reduced. Then, we describe the dynamic quantization and the low-bit tensor convolution arithmetic to leverage the MLS tensor format efficiently. Experiments show that our framework achieves a superior trade-off between the accuracy and the bit-width than previous low-bit training frameworks. For training a variety of models on CIFAR-10, using 1-bit mantissa and 2-bit exponent is adequate to keep the accuracy loss within $1\%$. And on larger datasets like ImageNet, using 4-bit mantissa and 2-bit exponent is adequate to keep the accuracy loss within $1\%$. Through the energy consumption simulation of the computing units, we can estimate that training a variety of models with our framework could achieve $8.3\sim10.2\times$ and $1.9\sim2.3\times$ higher energy efficiency than training with full-precision and 8-bit floating-point arithmetic, respectively.<br />Comment: 13 pages, 7 figures

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1228412045
Document Type :
Electronic Resource