All-in-One Analog AI Hardware: On-Chip Training and Inference with Conductive-Metal-Oxide/HfOx ReRAM Devices

By Donato Francesco Falcone 1, Victoria Clerico 1, Wooseok Choi 1, Tommaso Stecconi 1Folkert Horst 1, Laura Bégon-Lours  1, Matteo Galetta 1, Antonio La Porta 1, Nikhil Garg 2,3Fabien Alibart 2,3, Bert Jan Offrein 1 and Valeria Bragaglia 1
1 IBM Research, Switzerland
2 Université de Sherbrooke, Canada
3 Université de Lille, France

Abstract 

Analog in-memory computing is an emerging paradigm designed to efficiently accelerate deep neural network workloads. Recent advancements have focused on either inference or training acceleration. However, a unified analog in-memory technology platform–capable of on-chip training, weight retention, and long-term inference acceleration–has yet to be reported. This work presents an all-in-one analog AI accelerator, combining these capabilities to enable energy-efficient, continuously adaptable AI systems. The platform leverages an array of analog filamentary conductive-metal-oxide (CMO)/HfOx resistive switching memory cells (ReRAM) integrated into the back-end-of-line (BEOL). The array demonstrates reliable resistive switching with voltage amplitudes below 1.5 V, compatible with advanced technology nodes. The array's multi-bit capability (over 32 stable states) and low programming noise (down to 10 nS) enable a nearly ideal weight transfer process, more than an order of magnitude better than other memristive technologies. Inference performance is validated through matrix-vector multiplication simulations on a 64 × 64 array, achieving a root-mean-square error improvement by a factor of 20 at 1 s and 3 at 10 years after programming, compared to state-of-the-art. Training accuracy closely matching the software equivalent is achieved across different datasets. The CMO/HfOx ReRAM technology lays the foundation for efficient analog systems accelerating both inference and training in deep neural networks.

To read the full article, click here

×
Semiconductor IP