1. Roofline-Model-Based Design Space Exploration for Dataflow Techniques of CNN Accelerators
- Author
-
Sungkyung Park, Chan Park, and Chester Park
- Subjects
General Computer Science ,Dataflow ,Design space exploration ,Computer science ,processing element (PE) ,Accelerator ,02 engineering and technology ,Parallel computing ,convolutional neural networks (CNNs) ,roofline ,Model-based design ,0202 electrical engineering, electronic engineering, information engineering ,General Materials Science ,Throughput (business) ,Loop unrolling ,Input/output ,General Engineering ,Memory bandwidth ,simulation ,020202 computer hardware & architecture ,Memory management ,Loop interchange ,020201 artificial intelligence & image processing ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,dataflow techniques ,lcsh:TK1-9971 - Abstract
To effectively compute convolutional layers, a complex design space must exist (e.g., the dataflow techniques associated with the layer parameters, loop transformation techniques, and hardware parameters). For efficient design space exploration (DSE) of various dataflow techniques, namely, the weight-stationary ( WS ), output-stationary ( OS ), row-stationary ( RS ), and no local reuse ( NLR ) techniques, the processing element (PE) structure and computational pattern of each dataflow technique are analyzed. Various performance metrics are calculated, namely, the throughput (in giga-operations per second, GOPS), computation-to-communication ratio ( CCR ), on-chip memory usage, and off-chip memory bandwidth, as closed-form expressions of the layer and hardware parameters. In addition, loop interchange and loop unrolling techniques with a double-buffer architecture are assumed. Many roofline model-based simulations are performed to explore relevant dataflow techniques for a wide variety of convolutional layers of typical neural networks. Through simulation, this paper provides insights into the trends in accelerator performance as the layer parameters change. For convolutional layers with large input and output feature map ( ifmap and ofmap ) widths and heights, the GOPS of the NLR dataflow technique tends to be higher than that of the techniques. For convolutional layers with low weight and ofmap widths and heights, the RS dataflow technique achieves optimal GOPS and on-chip memory usage. In the case of convolutional layers with small weight widths and heights, the GOPS of the WS dataflow technique tends to be high. In the case of convolutional layers with small ofmap widths and heights, the OS dataflow technique achieves optimal GOPS and on-chip memory usage.
- Published
- 2020