NeuReality Boosts AI Acelerator Utilization With NAPU
By Sally Ward-Foxton, EETimes (April 4, 2024)
Startup NeuReality wants to replace the host CPU in data center AI inference systems with dedicated silicon that can cut total cost of ownership and power consumption. The Israeli startup developed a class of chip it calls the network addressable processing unit (NAPU), which includes hardware implementations for typical CPU functions like the hypervisor. NeuReality’s aim is to increase AI accelerator utilization by removing bottlenecks caused by today’s host CPUs.
NeuReality CEO Moshe Tanach told EE Times its NAPU enables 100% utilization of AI accelerators.
To read the full article, click here
Related Semiconductor IP
- Bluetooth Low Energy 6.0 Digital IP
- Ultra-low power high dynamic range image sensor
- Flash Memory LDPC Decoder IP Core
- SLM Signal Integrity Monitor
- Digital PUF IP
Related News
- Israeli AI Chip Startup Raises Seed Funding
- NeuReality unveils novel AI-centric platform to empower the growth of real-life AI applications
- IBM and NeuReality team up to build the next generation of AI inference platforms
- Samsung Ventures invests in Israeli AI systems and semiconductor company NeuReality
Latest News
- SiMa.ai Raises $85M to Scale Physical AI, Bringing Total Funding to $355M
- Armv9 and CSS Royalties Drive Growth in $1bn Arm Q1 Earnings
- Creonic Releases DVB-S2X Demodulator Version 6.0 with Increased Bitwidth and Annex M Support
- Arm Q1 FYE26 Revenue Exceeds $1 Billion for Second Consecutive Quarter
- 1‑VIA Expands Globally with New India R&D Office in Pune to Accelerate Innovation in Data Center Connectivity