Delivering Terabyte-Scale Bandwidth with HBM3-Ready Memory Subsystem
An exponential rise in data volume, and the meteoric rise of advanced workloads like AI/ML training, requires constant innovation in all aspects of computing. Memory bandwidth is a critical enabler of unleashing the power of processors and accelerators, and the High Bandwidth Memory (HBM) standard has evolved rapidly to deliver the performance required by the most demanding applications.
For current generation HBM2E, Rambus introduced the industry’s fastest memory subsystem capable of 4 gigabits per second (Gbps) operation. With a 1024-bit wide interface, 4 Gbps signaling delivers 512 gigabytes per second (GB/s) of bandwidth. In accelerator architectures with 4-6 HBM2E DRAM devices (each device being a 3D stack of DRAM chips), there’s the capability for 2-3 Terabytes per second (TB/s) of memory bandwidth. That’s enormous, but the appetite for bandwidth is insatiable, so the wheel of innovation needs to keep spinning.
To read the full article, click here
Related Semiconductor IP
- HBM3 PHY V2 - TSMC N3E
- HBM3 PHY V2 (Hard) - TSMC N3P
- HBM3 PHY - TSMC N6
- HBM3 PHY - TSMC N5 1.2V
- HBM3 PHY (Hard 1) - TSMC N6
Related Blogs
- Arasan Chip Systems Inc. Empowers Camera and Display Semiconductor Manufacturers with multiple -Supported Process and Integrated Subsystems, Delivering Cost and Time Savings
- Solve SoC Bottlenecks with Smart Local Memory in AI/ML Subsystems
- Qualcomm JEDEC Mobile Keynote: Memory Bandwidth and Thermal Limits
- It's about the mobile GPU memory bandwidth per watt, folks
Latest Blogs
- FiRa 3.0 Use Cases: Expanding the Future of UWB Technology
- Cadence Announces Industry's First Verification IP for Embedded USB2v2 (eUSB2v2)
- The Industry’s First USB4 Device IP Certification Will Speed Innovation and Edge AI Enablement
- Understanding Extended Metadata in CXL 3.1: What It Means for Your Systems
- 2025 Outlook with Mahesh Tirupattur of Analog Bits