Back to Search Start Over

Diluted binary neural network.

Authors :
Lin, Yuhan
Niu, Lingfeng
Xiao, Yang
Zhou, Ruizhi
Source :
Pattern Recognition. Aug2023, Vol. 140, pN.PAG-N.PAG. 1p.
Publication Year :
2023

Abstract

• We provide a new insight on the reason of BNNs' performance degradation. • A novel DBNN model is proposed to handle the accuracy drop of BNNs. • A sparsity-binarization scheme for weight is given to avoid mandatory representation. • A stable binarization strategy for activation is developed with layer normalization. • A customized proximal gradient method is designed to derive diluted binary weights. Binary neural networks (BNNs) are promising on resource-constrained devices because they reduce memory consumption and accelerate inference effectively. However, they are still potential on performance improvement. Prior studies attribute performance degradation of BNNs to limited representation ability and gradient mismatch. In this paper, we find that it also results from the mandatory representation of small full-precision auxiliary weights to large values. To tackle with this issue, we propose an approach dubbed as Diluted Binary Neural Network (DBNN). Besides avoiding mandatory representation effectively, the proposed DBNN also alleviates sign flip problem to a large extent. For activations, we jointly minimize quantization error and maximize information entropy to develop the binarization scheme. Compared with existing sparsity-binarization approaches, DBNN trains network from scratch without other procedures and achieves larger sparsity. Experiments on several datasets with various networks demonstrate the superiority of our approach. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00313203
Volume :
140
Database :
Academic Search Index
Journal :
Pattern Recognition
Publication Type :
Academic Journal
Accession number :
163267101
Full Text :
https://doi.org/10.1016/j.patcog.2023.109556