eFPGA Saved Us Millions of Dollars. It Can Do the Same for You
By Andy Jaros, Flex Logix
For those of you who follow Flex Logix®, you already know that we have an IP business, EFLX® eFGPA, and an edge inferencing co-processor chip and board business, InferX®. InferX came about because we had many customers ask if they can run AI/ML algorithms in EFLX. The answer was and still is, of course you can – EFLX is an FPGA fabric similar to what FPGA chips use. Our co-founder, Cheng Wang, took some time and studied the challenges of AI processing in more detail and came up with a highly efficient edge inferencing processor leveraging Flex Logix proprietary eFPGA technology. When performance, power and area results were shared with our board of directors, they thought it so compelling, they told us to build a chip. Hence, InferX X1 was born.
The X1 was specified to be a lean, high performance edge accelerator for AI inference processing incorporating Flex Logix’ proprietary tensor processor, PCIe, DDR, memory and a NoC. When it came time to architect the chip, there was an internal debate about adding EFLX to the X1 chip, mainly because it takes up area and our use case was pretty basic: support a GPIO interface and help with chip debug. Not a strong reason to add one square millimeter in 16nm. We proceeded anyway to demonstrate “Eating our own dog food”, by connecting the eFPGA to both the NoC bus and GPIO to maximize flexibility. Fast forward to chip bring-up.
To read the full article, click here
Related Semiconductor IP
- eFPGA
- eFPGA Hard IP Generator
- Radiation-Hardened eFPGA
- eFPGA IP as a synthesizable RTL core
- eFPGA IP - 100% third party standard cells
Related White Papers
- Last-Time Buy Notifications For Your ASICs? How To Make the Most of It
- Are you optimizing the benefits of cloud computing for faster reliability verification?
- How Low Can You Go? Pushing the Limits of Transistors - Deep Low Voltage Enablement of Embedded Memories and Logic Libraries to Achieve Extreme Low Power
- CANsec: Security for the Third Generation of the CAN Bus
Latest White Papers
- Reimagining AI Infrastructure: The Power of Converged Back-end Networks
- 40G UCIe IP Advantages for AI Applications
- Recent progress in spin-orbit torque magnetic random-access memory
- What is JESD204C? A quick glance at the standard
- Open-Source Design of Heterogeneous SoCs for AI Acceleration: the PULP Platform Experience