Pie: Pooling CPU Memory for LLM Inference
By Yi Xu, Ziming Mao, Xiangxi Mo, Shu Liu, Ion Stoica (UC Berkeley)
The rapid growth of LLMs has revolutionized natural language processing and AI analysis, but their increasing size and memory demands present significant challenges. A common solution is to spill over to CPU memory; however, traditional GPU-CPU memory swapping often results in higher latency and lower throughput.
This paper introduces Pie, an LLM inference framework that addresses these challenges with performance-transparent swapping and adaptive expansion. By leveraging predictable memory access patterns and the high bandwidth of modern hardware like the NVIDIA GH200 Grace Hopper Superchip, Pie enables concurrent data swapping without affecting foreground computation, expanding effective memory without added latency. Adaptive expansion dynamically adjusts CPU memory allocation based on real-time information, optimizing memory usage and performance under varying conditions.
Pie maintains low computation latency, high throughput, and high elasticity. Our experimental evaluation demonstrates that Pie achieves optimal swapping policy during cache warmup and effectively balances increased memory capacity with negligible impact on computation. With its extended capacity, Pie outperforms vLLM by up to 1.9X in throughput and 2X in latency. Additionally, Pie can reduce GPU memory usage by up to 1.67X while maintaining the same performance. Compared to FlexGen, an offline profiling-based swapping solution, Pie achieves magnitudes lower latency and 9.4X higher throughput.
To read the full article, click here
Related Semiconductor IP
- Configurable CPU tailored precisely to your needs
- Multi-core capable 64-bit RISC-V CPU with vector extensions
- Multi-core capable 32-bit RISC-V CPU with vector extensions
- Ultra-low power consumption out-of-order commercial-grade 64-bit RISC-V CPU IP
- CPU IP Following the RVA23 Profile, supporting RVV1.0 and all extensions of Vector Crypto
Related Articles
- Combating the Memory Walls: Optimization Pathways for Long-Context Agentic LLM Inference
- SPAD: Specialized Prefill and Decode Hardware for Disaggregated LLM Inference
- LLM Inference with Codebook-based Q4X Quantization using the Llama.cpp Framework on RISC-V Vector CPUs
- The Growing Importance of AI Inference and the Implications for Memory Technology
Latest Articles
- RISC-V Functional Safety for Autonomous Automotive Systems: An Analytical Framework and Research Roadmap for ML-Assisted Certification
- Emulation-based System-on-Chip Security Verification: Challenges and Opportunities
- A 129FPS Full HD Real-Time Accelerator for 3D Gaussian Splatting
- SkipOPU: An FPGA-based Overlay Processor for Large Language Models with Dynamically Allocated Computation
- TensorPool: A 3D-Stacked 8.4TFLOPS/4.3W Many-Core Domain-Specific Processor for AI-Native Radio Access Networks