AI-driven SRAM demand needs integrated repair and security
By Meng-Yi Wu, eMemory
embedded.com (July 15, 2024)
Increasing popularity of AI applications and DPU architecture has led to growing demand for higher SRAM densities, in turn placing challenges on SRAM yield and reliability.
Along with the rise of the internet of things (IoT), mobile devices, and edge computing, the boom in AI-enhanced features has enabled the addition of even greater functionality in applications such as intelligent sensing, in-vehicle driver assistance (ADAS), and voice recognition, all of which require the use of increasingly larger training models.
However, as the progress of CPU performance slows, new ideas to reduce the I/O and data loading on CPUs are becoming more popular. These include such solutions as DPU or PIM (process in memory) architectures, as well as the introduction of hierarchical data processing. However, as more CPUs are required for hierarchical processing, there is a corresponding need for more SRAM caches to serve these high-speed CPUs. Thus, the increasing popularity of AI applications and DPU architecture implementations has led to growing demand for higher SRAM densities
To read the full article, click here
Related Semiconductor IP
- UCIe D2D Adapter & PHY Integrated IP
- Low Dropout (LDO) Regulator
- 16-Bit xSPI PSRAM PHY
- MIPI CSI-2 CSE2 Security Module
- ASIL B Compliant MIPI CSI-2 CSE2 Security Module
Related Articles
- Secure Your Security Key in On-Chip SRAM: Techniques to avoid Data Remanance Attacks
- Security in transit
- Security needs more than checklist compliance
- Basics of SRAM PUF and how to deploy it for IoT security
Latest Articles
- RISC-V Functional Safety for Autonomous Automotive Systems: An Analytical Framework and Research Roadmap for ML-Assisted Certification
- Emulation-based System-on-Chip Security Verification: Challenges and Opportunities
- A 129FPS Full HD Real-Time Accelerator for 3D Gaussian Splatting
- SkipOPU: An FPGA-based Overlay Processor for Large Language Models with Dynamically Allocated Computation
- TensorPool: A 3D-Stacked 8.4TFLOPS/4.3W Many-Core Domain-Specific Processor for AI-Native Radio Access Networks