NVIDIA Previews Open-source Processor Core for Deep Neural Network Inference
With the proliferation of deep learning, NVIDIA has realized its longstanding aspirations to make general-purpose graphics processing units (GPGPUs) a mainstream technology. The company's GPUs are commonly used to accelerate neural network training, and are also being adopted for neural network inference acceleration in self-driving cars, robots and other high-end autonomous platforms. NVIDIA also sees plenty of opportunities for inference acceleration in IoT and other "edge" platforms, although it doesn't intend to supply them with chips. Instead, it's decided to open-source the NVDLA deep learning processor core found in its "Xavier" SoC introduced last fall.
To read the full article, click here
Related Semiconductor IP
- Deep Buffering Memory 1G Ethernet Switch
- Deep Learning-based Video Super Resolution Accelerator IP
- High performance-efficient deep learning accelerator for edge and end-point inference
- Distributed Thermal Sensor (DTS) Deep NWELL, TSMC N3E
- Temperature Sensor Deep NWELL, TSMC N3E
Related Blogs
- Synopsys Fields Processor Core for Neural Network Computer Vision Applications
- Imagination Technologies' Upgraded GPUs, New Neural Network Core Provide Deep Learning Processing Options
- Processor Wars: NVIDIA reveals a phantom fifth ARM Cortex-A9 processor core in Kal-El mobile processor IC. Guess why it's there?
- Tensilica Vision P6 Processor Core Adopts Deep Learning-Focused Enhancements
Latest Blogs
- A Bench-to-In-Field Telemetry Platform for Datacenter Power Management
- IDS-Verify™: From Specification to Sign-Off – Automated CSR, Hardware Software Interface and CPU-Peripheral Interface Verification
- RISC-V and GPU Synergy in Practice: A Path Towards High-Performance SoCs from SpacemiT K3
- EDA AI Agents: Intelligent Automation in Semiconductor & PCB Design
- Why Security Can't Exist Without Trust