Back to Search
Start Over
A neural network pruning and quantization algorithm for hardware deployment.
- Source :
-
Computer Engineering & Science / Jisuanji Gongcheng yu Kexue . Sep2024, Vol. 46 Issue 9, p1547-1553. 7p. - Publication Year :
- 2024
-
Abstract
- Abstract:Due to their superior performance, deep neural networks have been widely applied in fields such as image recognition and object detection. However, they contain a large number of parameters and require immense computational power, posing challenges for deployment on mobile edge devices that require low latency and low power consumption. To address this issue, a compression algorithm that replaces multiplication operations with bit-shifting and addition is proposed. This algorithm compresses neural network parameters to low bit-widths through pruning and quantization. This algorithm reduces the hardware deployment difficulty under limited multiplication resources, meets the requirements of low latency and low power consumption on mobile edge devices, and improves operational efficiency. Experiments conducted on classical neural networks with the ImageNet dataset revealed that when the neural network parameters were compressed to 4 bits, the accuracy remained essentially unchanged compared to the full-precision neural network. Furthermore, for ResNet18, ResNet50, and GoogleNet, the Top-1/Top-5 accuracies even improved by 0.38%/0.22%, 0.35%/0.21%, and 1.14%/0.57%, respectively. When testing the eighth convolutional layer of VGG16 deployed on Zynq7035, the results showed that the compressed network reduced the inference time by 51.1% and power consumption by 46.7%, while using 43% fewer DSP resources. [ABSTRACT FROM AUTHOR]
Details
- Language :
- Chinese
- ISSN :
- 1007130X
- Volume :
- 46
- Issue :
- 9
- Database :
- Academic Search Index
- Journal :
- Computer Engineering & Science / Jisuanji Gongcheng yu Kexue
- Publication Type :
- Academic Journal
- Accession number :
- 180188476
- Full Text :
- https://doi.org/10.3969/j.issn.1007-130X.2024.09.004