Breaking the HBM Bit Cost Barrier: Domain-Specific ECC for AI Inference Infrastructure
By Rui Xie †, Asad Ul Haq †, Yunhua Fang †, Linsen Ma †, Sanchari Sen ∗, Swagath Venkataramani ∗, Liu Liu †, Tong Zhang †
† Rensselaer Polytechnic Institute, NY, USA,
∗ IBM, NY, USA
Abstract
High-Bandwidth Memory (HBM) delivers exceptional bandwidth and energy efficiency for AI workloads, but its high cost per bit, driven in part by stringent on-die reliability requirements, poses a growing barrier to scalable deployment. This work explores a systemlevel approach to cost reduction by eliminating on-die ECC and shifting all fault management to the memory controller. We introduce a domainspecific ECC framework combining large-codeword Reed–Solomon (RS) correction with lightweight fine-grained CRC detection, differential parity updates to mitigate write amplification, and tunable protection based on data importance. Our evaluation using LLM inference workloads shows that, even under raw HBM bit error rates up to 10−3, the system retains over 78% of throughput and 97% of model accuracy compared with systems equipped with ideal error-free HBM. By treating reliability as a tunable system parameter rather than a fixed hardware constraint, our design opens a new path toward low-cost, high-performance HBM deployment in AI infrastructure.
Index Terms — HBM, Error-Correcting Code, CRC, Reliability
To read the full article, click here
Related Semiconductor IP
- HBM Memory Controller
- HBM 4 Verification IP
- Verification IP for HBM
- TSMC CLN16FFGL+ HBM PHY IP
- Simulation VIP for HBM
Related White Papers
- Partitioning to optimize AI inference for multi-core platforms
- The Growing Importance of AI Inference and the Implications for Memory Technology
- Top 5 Reasons why CPU is the Best Processor for AI Inference
- Why Software is Critical for AI Inference Accelerators
Latest White Papers
- OmniSim: Simulating Hardware with C Speed and RTL Accuracy for High-Level Synthesis Designs
- Balancing Power and Performance With Task Dependencies in Multi-Core Systems
- LLM Inference with Codebook-based Q4X Quantization using the Llama.cpp Framework on RISC-V Vector CPUs
- PCIe 5.0: The universal high-speed interconnect for High Bandwidth and Low Latency Applications Design Challenges & Solutions
- Basilisk: A 34 mm2 End-to-End Open-Source 64-bit Linux-Capable RISC-V SoC in 130nm BiCMOS