NeuReality Boosts AI Acelerator Utilization With NAPU
By Sally Ward-Foxton, EETimes (April 4, 2024)
Startup NeuReality wants to replace the host CPU in data center AI inference systems with dedicated silicon that can cut total cost of ownership and power consumption. The Israeli startup developed a class of chip it calls the network addressable processing unit (NAPU), which includes hardware implementations for typical CPU functions like the hypervisor. NeuReality’s aim is to increase AI accelerator utilization by removing bottlenecks caused by today’s host CPUs.
NeuReality CEO Moshe Tanach told EE Times its NAPU enables 100% utilization of AI accelerators.
To read the full article, click here
Related Semiconductor IP
- DeWarp IP
- 6-bit, 12 GSPS Flash ADC - GlobalFoundries 22nm
- LunaNet AFS LDPC Encoder and Decoder IP Core
- ReRAM NVM in DB HiTek 130nm BCD
- UFS 5.0 Host Controller IP
Related News
- Israeli AI Chip Startup Raises Seed Funding
- NeuReality unveils novel AI-centric platform to empower the growth of real-life AI applications
- IBM and NeuReality team up to build the next generation of AI inference platforms
- Samsung Ventures invests in Israeli AI systems and semiconductor company NeuReality
Latest News
- GUC Monthly Sales Report – March 2026
- Qualitas Semiconductor Licenses 2nm Process-Based MIPI C/D-PHY IP to U.S. Edge AI SoC Company
- Global Semiconductor Sales Increase Substantially in February
- Hardware Root of Trust Essential for AI Chip Integrity
- AI Compute Demand Drives 44% YoY Growth for Top 10 Global Fabless IC Firms in 2025