All-in-One Analog AI Hardware: On-Chip Training and Inference with Conductive-Metal-Oxide/HfOx ReRAM Devices
By Donato Francesco Falcone 1, Victoria Clerico 1, Wooseok Choi 1, Tommaso Stecconi 1, Folkert Horst 1, Laura Bégon-Lours 1, Matteo Galetta 1, Antonio La Porta 1, Nikhil Garg 2,3, Fabien Alibart 2,3, Bert Jan Offrein 1 and Valeria Bragaglia 1
1 IBM Research, Switzerland
2 Université de Sherbrooke, Canada
3 Université de Lille, France
Abstract
Analog in-memory computing is an emerging paradigm designed to efficiently accelerate deep neural network workloads. Recent advancements have focused on either inference or training acceleration. However, a unified analog in-memory technology platform–capable of on-chip training, weight retention, and long-term inference acceleration–has yet to be reported. This work presents an all-in-one analog AI accelerator, combining these capabilities to enable energy-efficient, continuously adaptable AI systems. The platform leverages an array of analog filamentary conductive-metal-oxide (CMO)/HfOx resistive switching memory cells (ReRAM) integrated into the back-end-of-line (BEOL). The array demonstrates reliable resistive switching with voltage amplitudes below 1.5 V, compatible with advanced technology nodes. The array's multi-bit capability (over 32 stable states) and low programming noise (down to 10 nS) enable a nearly ideal weight transfer process, more than an order of magnitude better than other memristive technologies. Inference performance is validated through matrix-vector multiplication simulations on a 64 × 64 array, achieving a root-mean-square error improvement by a factor of 20 at 1 s and 3 at 10 years after programming, compared to state-of-the-art. Training accuracy closely matching the software equivalent is achieved across different datasets. The CMO/HfOx ReRAM technology lays the foundation for efficient analog systems accelerating both inference and training in deep neural networks.
To read the full article, click here
Related Semiconductor IP
- AI accelerator
- AI Accelerator
- AI Accelerator Specifically for CNN
- Deeply Embedded AI Accelerator for Microcontrollers and End-Point IoT Devices
- Performance Efficiency Leading AI Accelerator for Mobile and Edge Devices
Related White Papers
- Generative AI for Analog Integrated Circuit Design: Methodologies and Applications
- Programmable accelerators: hardware performance with software flexibility
- H.264 Baseline Decoder With ADI Blackfin DSP and Hardware Accelerators
- Reducing Power in Embedded Systems by Adding Hardware Accelerators
Latest White Papers
- All-in-One Analog AI Hardware: On-Chip Training and Inference with Conductive-Metal-Oxide/HfOx ReRAM Devices
- From I2C to I3C: Evolution of Two-Wire Communication in Embedded Systems
- QiMeng: Fully Automated Hardware and Software Design for Processor Chip
- RISC-V source class riscv_asm_program_gen, the brain behind assembly instruction generator
- Concealable physical unclonable functions using vertical NAND flash memory