Why Software is Critical for AI Inference Accelerators
By Geoff Tate, Flex Logix (April 23, 2020)
Inference accelerators represent an incredible market opportunity not only to chip and IP companies, but also to the customers who desperately need them. As inference accelerators come to market, a common comment we hear is: “Why is my inference chip not performing like it was designed to?”
Oftentimes, the simple answer is the software.
Software is key
All inference accelerators today are programmable because customers believe their model will evolve over time. This programmability will allow them to take advantage of enhancements in the future, something that would not be possible with hard-wired accelerators. However, customers want this programmability in a way where they can get the most throughput for a certain cost, and for a certain amount of power. This means they have to use the hardware very efficiently. The only way to do this is to design the software in parallel with the hardware to make sure they work together very well to achieve the maximum throughput.
To read the full article, click here
Related Semiconductor IP
- eFPGA
- eFPGA on GlobalFoundries GF12LP
- Heterogeneous eFPGA architecture with LUTs, DSPs, and BRAMs on GlobalFoundries GF12LP
- eFPGA Soft IP
- Radiation-Hardened eFPGA
Related Articles
- Top 5 Reasons why CPU is the Best Processor for AI Inference
- Why Interlaken is a great choice for architecting chip to chip communications in AI chips
- Why the Memory Subsystem is Critical in Inferencing Chips
- AI Edge Inference is Totally Different to Data Center
Latest Articles
- Analog Foundation Models
- Modeling and Optimizing Performance Bottlenecks for Neuromorphic Accelerators
- RISC-V Based TinyML Accelerator for Depthwise Separable Convolutions in Edge AI
- Exclude Smart in Functional Coverage
- A 0.32 mm² 100 Mb/s 223 mW ASIC in 22FDX for Joint Jammer Mitigation, Channel Estimation, and SIMO Data Detection