Why Software is Critical for AI Inference Accelerators
By Geoff Tate, Flex Logix (April 23, 2020)
Inference accelerators represent an incredible market opportunity not only to chip and IP companies, but also to the customers who desperately need them. As inference accelerators come to market, a common comment we hear is: “Why is my inference chip not performing like it was designed to?”
Oftentimes, the simple answer is the software.
Software is key
All inference accelerators today are programmable because customers believe their model will evolve over time. This programmability will allow them to take advantage of enhancements in the future, something that would not be possible with hard-wired accelerators. However, customers want this programmability in a way where they can get the most throughput for a certain cost, and for a certain amount of power. This means they have to use the hardware very efficiently. The only way to do this is to design the software in parallel with the hardware to make sure they work together very well to achieve the maximum throughput.
To read the full article, click here
Related Semiconductor IP
- eFPGA
- eFPGA Hard IP Generator
- Radiation-Hardened eFPGA
- eFPGA IP as a synthesizable RTL core
- eFPGA IP - 100% third party standard cells
Related White Papers
- Top 5 Reasons why CPU is the Best Processor for AI Inference
- Why Interlaken is a great choice for architecting chip to chip communications in AI chips
- Why the Memory Subsystem is Critical in Inferencing Chips
- AI Edge Inference is Totally Different to Data Center
Latest White Papers
- Reimagining AI Infrastructure: The Power of Converged Back-end Networks
- 40G UCIe IP Advantages for AI Applications
- Recent progress in spin-orbit torque magnetic random-access memory
- What is JESD204C? A quick glance at the standard
- Open-Source Design of Heterogeneous SoCs for AI Acceleration: the PULP Platform Experience