Back to Search
Start Over
33.1 A 74 TMACS/W CMOS-RRAM Neurosynaptic Core with Dynamically Reconfigurable Dataflow and In-situ Transposable Weights for Probabilistic Graphical Models
- Source :
- ISSCC
- Publication Year :
- 2020
- Publisher :
- IEEE, 2020.
-
Abstract
- Many powerful neural network (NN) models such as probabilistic graphical models (PGMs) and recurrent neural networks (RNNs) require flexibility in dataflow and weight access patterns as shown in Fig. 33.1.1 Typically, Compute-In-Memory (CIM) designs do not implement such dataflows or do so by replicating circuits at the memory periphery such as ADCs/neurons along both the rows and columns of the memory array, leading to an overhead in operation. This paper describes a CIM architecture implemented in a 130nm CMOS/RRAM process, that delivers the highest reported computational energy-efficiency of 74 tera-multiply-accumulates per second per watt (TMACS/W) for RRAM-based CIM architectures while simultaneously offering dataflow reconfigurability to address the limitations of previous designs. This is made possible through two key features: 1) a runtime reconfigurable dataflow with in-situ access to RRAM array and its transpose for efficient access to NN weights and 2) a voltage sensing stochastic integrate-and-fire analog neuron (I&F) that is reused for correlated double sampling (CDS), stochastic voltage integration, and threshold comparison.
- Subjects :
- 010302 applied physics
Artificial neural network
Computer science
Dataflow
020208 electrical & electronic engineering
Probabilistic logic
Reconfigurability
02 engineering and technology
01 natural sciences
Recurrent neural network
CMOS
Computer architecture
0103 physical sciences
0202 electrical engineering, electronic engineering, information engineering
Overhead (computing)
Graphical model
Subjects
Details
- Database :
- OpenAIRE
- Journal :
- 2020 IEEE International Solid- State Circuits Conference - (ISSCC)
- Accession number :
- edsair.doi...........5be3ee3ce6db6a424306d82c68020229