1. Impact of Phase‐Change Memory Flicker Noise and Weight Drift on Analog Hardware Inference for Large‐Scale Deep Learning Networks
- Author
-
Jin-Ping Han, Malte J. Rasch, Zuoguang Liu, Paul Solomon, Kevin Brew, Kangguo Cheng, Injo Ok, Victor Chan, Michael Longstreet, Wanki Kim, Robert L. Bruce, Cheng-Wei Cheng, Nicole Saulnier, and Vijay Narayanan
- Subjects
analog RPU inference ,deep-learning networks ,flicker noise ,hardware awareness training ,PCM ,weight drift ,Computer engineering. Computer hardware ,TK7885-7895 ,Control engineering systems. Automatic machinery (General) ,TJ212-225 - Abstract
The analog AI core concept is appealing for deep‐learning (DL) because it combines computation and memory functions into a single device. Yet, significant challenges such as noise and weight drift will impact large‐scale analog in‐memory computing. Here, effects of flicker noise and drift on large DL systems are explored using a new flicker‐noise model with memory, which preserves temporal correlations, including a flicker noise figure of merit (FOM) Ar to quantify impacts on system performance. Flicker noise is characterized for Ge2Sb2Te5 (GST) based phase‐change memory (PCM) cells with a discovery of read‐noise asymmetry tied to shape asymmetry of mushroom cells. This experimental read polarity dependence is consistent with Pirovano's trap activation and defect annihilation model in an asymmetric GST cell. The impact of flicker noise and resistance drift of analog PCM synaptic devices on deep‐learning hardware is assessed for six large‐scale deep neural networks (DNNs) used for image classification, finding that the inference top‐1 accuracy degraded with the accumulated device flicker noise and drift as ∝Ar×twait, and ∝twait−ν, respectively, where ν is the drift coefficient. These negative impacts could be mitigated with a new hardware‐aware (HWA) (pre)‐training of the DNNs, which is applied before programming to the analog arrays.
- Published
- 2022
- Full Text
- View/download PDF