AI-driven SRAM demand needs integrated repair and security
By Meng-Yi Wu, eMemory
embedded.com (July 15, 2024)
Increasing popularity of AI applications and DPU architecture has led to growing demand for higher SRAM densities, in turn placing challenges on SRAM yield and reliability.
Along with the rise of the internet of things (IoT), mobile devices, and edge computing, the boom in AI-enhanced features has enabled the addition of even greater functionality in applications such as intelligent sensing, in-vehicle driver assistance (ADAS), and voice recognition, all of which require the use of increasingly larger training models.
However, as the progress of CPU performance slows, new ideas to reduce the I/O and data loading on CPUs are becoming more popular. These include such solutions as DPU or PIM (process in memory) architectures, as well as the introduction of hierarchical data processing. However, as more CPUs are required for hierarchical processing, there is a corresponding need for more SRAM caches to serve these high-speed CPUs. Thus, the increasing popularity of AI applications and DPU architecture implementations has led to growing demand for higher SRAM densities
To read the full article, click here
Related Semiconductor IP
- Process/Voltage/Temperature Sensor with Self-calibration (Supply voltage 1.2V) - TSMC 3nm N3P
- USB 20Gbps Device Controller
- SM4 Cipher Engine
- Ultra-High-Speed Time-Interleaved 7-bit 64GSPS ADC on 3nm
- Fault Tolerant DDR2/DDR3/DDR4 Memory controller
Related White Papers
- Secure Your Security Key in On-Chip SRAM: Techniques to avoid Data Remanance Attacks
- Security in transit
- Security needs more than checklist compliance
- Basics of SRAM PUF and how to deploy it for IoT security
Latest White Papers
- Fault Injection in On-Chip Interconnects: A Comparative Study of Wishbone, AXI-Lite, and AXI
- eFPGA – Hidden Engine of Tomorrow’s High-Frequency Trading Systems
- aTENNuate: Optimized Real-time Speech Enhancement with Deep SSMs on RawAudio
- Combating the Memory Walls: Optimization Pathways for Long-Context Agentic LLM Inference
- Hardware Acceleration of Kolmogorov-Arnold Network (KAN) in Large-Scale Systems