Why ASIC Design Makes Sense for LLM-On-Device
A look at architectural and design considerations when designing ASICs for LLM-on-device.
By Steve Xu, Co-Founder and Chief Architect, XgenSilicon
EETimes | July 14, 2025
Multimodality LLMs can enable powerful real-time vision and audio applications if chip power and cost meet the constraints of edge devices. By adopting an ASIC approach, it’s possible to achieve a hardware-efficient implementation through custom design, resulting in lower power and cost compared to using off-the-shelf components, such as GPUs, NPUs, and application processors.
An ASIC design is a systematic approach to address power efficiency bottlenecks, which may be different from model to model and per deployment constraint.
For example, the power of Snapdragon AR1+ Gen 1 running a 1B vision model is 1 watt. An ASIC implementation of the same model can reduce it to 0.1 watt with design tradeoffs between silicon die area and power consumption by shifting the design from NPU + DDR architecture to ASIC + on-chip memory architecture. For smart glasses with a 500 mAh battery, this translates the active time of vision from 0.5 hours to 5 hours.
In this article, we’ll illustrate architectural and design considerations to be taken into account when planning and designing ASICs for LLM-on-device.
To read the full article, click here
Related Semiconductor IP
- Simulation VIP for Ethernet UEC
- Bluetooth® Low Energy 6.2 PHY IP with Channel Sounding
- Simulation VIP for UALink
- General use, integer-N 4GHz Hybrid Phase Locked Loop on TSMC 28HPC
- JPEG XL Encoder
Related News
- Why the Microsemi-Actel deal makes complete sense
- Analysis: Why ARM-AMD makes sense
- Aion Silicon Expands Barcelona Design Center to Meet Surging Demand for ASIC and SoC Solutions
- Cadence-Mentor deal makes sense, says analyst
Latest News
- Mixel MIPI IP Integrated into Automotive Radar Processors Supporting Safety-critical Applications
- GlobalFoundries and Navitas Semiconductor Partner to Accelerate U.S. GaN Technology and Manufacturing for AI Datacenters and Critical Power Applications
- VLSI EXPERT selects Innatera Spiking Neural Processors to build industry-led neuromorphic talent pool
- SkyWater Technology and Silicon Quantum Computing Team to Advance Hybrid Quantum-Classical Computing
- Dnotitia Revolutionizes AI Storage at SC25: New VDPU Accelerator Delivers Up to 9x Performance Boost