Solve SoC Bottlenecks with Smart Local Memory in AI/ML Subsystems
In today’s disaggregated electronics supply chain the (1) application software developer, (2) the ML model developer, (3) the device maker, (4) the SoC design team and (5) the NPU IP vendor often work for as many as five different companies. It can be difficult or impossible for the SoC team to know or predict actual AI/ML workloads and full system behaviors as many as two or three years in advance of the actual deployment. But then how can that SoC team make good choices provisioning compute engines and adequate memory resources for the unknown future without defaulting to “Max TOPS / Min Area”?
There has to be a smarter way to eliminate bottlenecks while determining the optimum local memory for AI/ML subsystems.
To read the full article, click here
Related Semiconductor IP
- CXL 3 Controller IP
- PCIe GEN6 PHY IP
- FPGA Proven PCIe Gen6 Controller IP
- Real-Time Microcontroller - Ultra-low latency control loops for real-time computing
- AI inference engine for real-time edge intelligence
Related Blogs
- High Speed Memory in Smart Phones: MIPI UniPro v1.8 for JEDEC UFS v3.0
- Faster Embedded Smartphone & Tablet Memory Is On The Way
- The Challenges of Video on Smart Mobile Devices
- SSD Interfaces and Performance Effects
Latest Blogs
- Arm Compute Platform at the Heart of Malaysia’s Silicon Vision
- IEEE 802.1ASdm-2024 Becomes an IEEE Standard – Advancing Time-Sensitive Networking
- Introducing the MIPS Atlas Portfolio for Physical AI
- Real-Time Intelligence for Physical AI at the Edge
- Moving the World with MIPS M8500 Real-Time Compute Solutions