Back to Search Start Over

Resistive Memory-based Neural Differential Equation Solver for Score-based Diffusion Model

Authors :
Yang, Jichang
Chen, Hegan
Chen, Jia
Wang, Songqi
Wang, Shaocong
Yu, Yifei
Chen, Xi
Wang, Bo
Zhang, Xinyuan
Cui, Binbin
Li, Yi
Lin, Ning
Xu, Meng
Xu, Xiaoxin
Qi, Xiaojuan
Wang, Zhongrui
Zhang, Xumeng
Shang, Dashan
Wang, Han
Liu, Qi
Cheng, Kwang-Ting
Liu, Ming
Yang, Jichang
Chen, Hegan
Chen, Jia
Wang, Songqi
Wang, Shaocong
Yu, Yifei
Chen, Xi
Wang, Bo
Zhang, Xinyuan
Cui, Binbin
Li, Yi
Lin, Ning
Xu, Meng
Xu, Xiaoxin
Qi, Xiaojuan
Wang, Zhongrui
Zhang, Xumeng
Shang, Dashan
Wang, Han
Liu, Qi
Cheng, Kwang-Ting
Liu, Ming
Publication Year :
2024

Abstract

Human brains image complicated scenes when reading a novel. Replicating this imagination is one of the ultimate goals of AI-Generated Content (AIGC). However, current AIGC methods, such as score-based diffusion, are still deficient in terms of rapidity and efficiency. This deficiency is rooted in the difference between the brain and digital computers. Digital computers have physically separated storage and processing units, resulting in frequent data transfers during iterative calculations, incurring large time and energy overheads. This issue is further intensified by the conversion of inherently continuous and analog generation dynamics, which can be formulated by neural differential equations, into discrete and digital operations. Inspired by the brain, we propose a time-continuous and analog in-memory neural differential equation solver for score-based diffusion, employing emerging resistive memory. The integration of storage and computation within resistive memory synapses surmount the von Neumann bottleneck, benefiting the generative speed and energy efficiency. The closed-loop feedback integrator is time-continuous, analog, and compact, physically implementing an infinite-depth neural network. Moreover, the software-hardware co-design is intrinsically robust to analog noise. We experimentally validate our solution with 180 nm resistive memory in-memory computing macros. Demonstrating equivalent generative quality to the software baseline, our system achieved remarkable enhancements in generative speed for both unconditional and conditional generation tasks, by factors of 64.8 and 156.5, respectively. Moreover, it accomplished reductions in energy consumption by factors of 5.2 and 4.1. Our approach heralds a new horizon for hardware solutions in edge computing for generative AI applications.

Details

Database :
OAIster
Publication Type :
Electronic Resource
Accession number :
edsoai.on1438544560
Document Type :
Electronic Resource