Neural Network Efficiency with Embedded FPGA's
The traditional metrics for evaluating IP are performance, power, and area, commonly abbreviated as PPA. Viewed independently, PPA measures can be difficult to assess. As an example, design constraints that are purely based on performance, without concern for the associated power dissipation and circuit area, are increasingly rare. There is a related set of characteristics of importance, especially given the increasing integration of SoC circuitry associated with deep neural networks (DNN) – namely, the implementation energy and area efficiency, usually represented as a performance per watt measure and a performance per area measure.
The DNN implementation options commonly considered are: a software-programmed (general purpose) microprocessor core, a programmed graphics processing unit (GPU), a field-programmable gate array, and a hard-wired logic block. In 2002, Broderson and Zhang from UC-Berkeley published a Technical Report that described the efficiency of different options, targeting digital signal processing algorithms. [1]
To read the full article, click here
Related Semiconductor IP
- Securyzr™ neo Core Platform
- 112G Multi-SerDes
- SHA3 Cryptographic Hash Cores
- ISO/IEC 7816 Verification IP
- 50MHz to 800MHz Integer-N RC Phase-Locked Loop on SMIC 55nm LL
Related Blogs
- Embedded Vision: The Road Ahead for Neural Networks and Five Likely Surprises
- FPGAs take on convolutional neural networks
- Synopsys Fields Processor Core for Neural Network Computer Vision Applications
- Reconfigurable redefined with embedded FPGA core IP
Latest Blogs
- Why Anti-tamper Sensors Matter: Agile Analog and Rambus Deliver Comprehensive Security Solution
- Automotive Ethernet with Comcores – Safety, Quality and ASIL certification of IP
- A Comparison on Different AMBA 5 CHI Verification IPs
- Cadence Recognized as TSMC OIP Partner of the Year at 2025 OIP Ecosystem Forum
- Accelerating Development Cycles and Scalable, High-Performance On-Device AI with New Arm Lumex CSS Platform