Back to Search Start Over

An analog-AI chip for energy-efficient speech recognition and transcription

Authors :
Ambrogio, S.
Narayanan, P.
Okazaki, A.
Fasoli, A.
Mackin, C.
Hosokawa, K.
Nomura, A.
Yasuda, T.
Chen, A.
Friz, A.
Ishii, M.
Luquin, J.
Kohda, Y.
Saulnier, N.
Brew, K.
Choi, S.
Ok, I.
Philip, T.
Chan, V.
Silvestre, C.
Ahsan, I.
Narayanan, V.
Tsai, H.
Burr, G. W.
Source :
Nature; August 2023, Vol. 620 Issue: 7975 p768-775, 8p
Publication Year :
2023

Abstract

Models of artificial intelligence (AI) that have billions of parameters can achieve high accuracy across a range of tasks1,2, but they exacerbate the poor energy efficiency of conventional general-purpose processors, such as graphics processing units or central processing units. Analog in-memory computing (analog-AI)3–7can provide better energy efficiency by performing matrix–vector multiplications in parallel on ‘memory tiles’. However, analog-AI has yet to demonstrate software-equivalent (SWeq) accuracy on models that require many such tiles and efficient communication of neural-network activations between the tiles. Here we present an analog-AI chip that combines 35 million phase-change memory devices across 34 tiles, massively parallel inter-tile communication and analog, low-power peripheral circuitry that can achieve up to 12.4 tera-operations per second per watt (TOPS/W) chip-sustained performance. We demonstrate fully end-to-end SWeqaccuracy for a small keyword-spotting network and near-SWeqaccuracy on the much larger MLPerf8recurrent neural-network transducer (RNNT), with more than 45 million weights mapped onto more than 140 million phase-change memory devices across five chips.

Details

Language :
English
ISSN :
00280836 and 14764687
Volume :
620
Issue :
7975
Database :
Supplemental Index
Journal :
Nature
Publication Type :
Periodical
Accession number :
ejs63891565
Full Text :
https://doi.org/10.1038/s41586-023-06337-5