Back to Search Start Over

CU.POKer: Placing DNNs on WSE With Optimal Kernel Sizing and Efficient Protocol Optimization

Authors :
Xiaopeng Zhang
Jingsong Chen
Fangzhou Wang
Lixin Liu
Evangeline F. Y. Young
Bentian Jiang
Jinwei Liu
Source :
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems. 41:1888-1901
Publication Year :
2022
Publisher :
Institute of Electrical and Electronics Engineers (IEEE), 2022.

Abstract

The tremendous growth in deep learning (DL) applications has created an exponential demand for computing power, which leads to the rise of AI-specific hardware. Targeted towards accelerating computation-intensive deep learning applications, AI hardware, including but not limited to GPGPU, TPU, ASICs, etc., have been adopted ubiquitously. As a result, domainspecific CAD tools play more and more important roles and have been deeply involved in both the design and compilation stages of modern AI hardware. Recently, ISPD 2020 contest introduced a special challenge targeting at the physical mapping of neural network workloads onto the largest commercial deep learning accelerator, CS-1 Wafer-Scale Engine (WSE). In this paper, we proposed CU.POKer, a high-performance engine fullycustomized for WSE’s DNN workload placement challenge. A provably optimal placeable kernel candidate searching scheme and a data-flow-aware placement tool are developed accordingly to ensure the state-of-the-art quality on the real industrial benchmarks. Experimental results on ISPD 2020 contest evaluation suites demonstrated the superiority of our proposed framework over not only the state-of-the-art (SOTA) placer but also the conventional heuristics used in general floorplanning.

Details

ISSN :
19374151 and 02780070
Volume :
41
Database :
OpenAIRE
Journal :
IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems
Accession number :
edsair.doi...........a34ab6d03d9dce221a1515ad24eefdb8
Full Text :
https://doi.org/10.1109/tcad.2021.3096458