Back to Search Start Over

A Compressed Data Partition and Loop Scheduling Scheme for Neural Networks

Authors :
Dejian Li
Rongqiang Fang
Jing Wang
Dongyan Zhao
Ting Chong
Zengmin Ren
Jun Ma
Source :
IEEE Access, Vol 10, Pp 95219-95228 (2022)
Publication Year :
2022
Publisher :
IEEE, 2022.

Abstract

Neural networks (NNs) have been widely adopted in various application domains. Deeper NNs greatly enhance the output accuracy, but complex NNs with more parameters incur intensive memory accesses, and the data usually need to be partitioned since it may exceed the on-chip storage. However, there is no research considering the partition and scheduling co-design of the NNs. In this paper, we propose a sparse NN data partition and loop scheduling scheme. We establish the compression efficiency model of the matrix sparse algorithm and design a partition selection method based on sparsity characteristics analyzed by the compression efficiency model. Further, we design a loop scheduling scheme based on the proper partition size. The experiment results show that the average memory access of each layer can be compressed to 68% of the original, and the throughput of the AlexNet, VGG and VGG19 is increased to an average of 1.66 times.

Details

Language :
English
ISSN :
21693536
Volume :
10
Database :
Directory of Open Access Journals
Journal :
IEEE Access
Publication Type :
Academic Journal
Accession number :
edsdoj.42a23a8f535546dbb1bb5781d7433cf8
Document Type :
article
Full Text :
https://doi.org/10.1109/ACCESS.2022.3204038