Back to Search Start Over

DSIP: A Scalable Inference Accelerator for Convolutional Neural Networks.

Authors :
Jo, Jihyuck
Cha, Soyoung
Rho, Dayoung
Park, In-Cheol
Source :
IEEE Journal of Solid-State Circuits; Feb2018, Vol. 53 Issue 2, p605-618, 14p
Publication Year :
2018

Abstract

This paper presents a scalable inference accelerator called a deep-learning specific instruction-set processor (DSIP) to support various convolutional neural networks (CNNs). For CNNs requiring a large amount of computations and memory accesses, a programmable inference system called master–slave instruction set architecture (ISA) is newly proposed to achieve high flexibility, processing speed, and energy efficiency. The master is responsible for sending and receiving feature maps in order to deal with neural networks in a scalable way, and the slave performs CNN operations, such as multiply accumulate, max pooling, and activation functions, on the features received from the master. The master–slave ISA maximizes computation speed by overlapping the off-chip data transmission and the CNN operations, and reduces power consumption by performing the convolution incrementally to reuse input and partial-sum data as maximally as possible. An inference system can be configured by connecting multiple DSIPs in a form of either 1-D or 2-D chain structure in order to enhance computation speed further. To evaluate the proposed accelerator, a prototype chip is implemented and evaluated for AlexNet. Compared to the state-of-the-art accelerator, the DSIP-based system enhances the energy efficiency by 2.17 $\times $ . [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00189200
Volume :
53
Issue :
2
Database :
Complementary Index
Journal :
IEEE Journal of Solid-State Circuits
Publication Type :
Academic Journal
Accession number :
127632728
Full Text :
https://doi.org/10.1109/JSSC.2017.2764045