Vendor: Ceva, Inc. Category: NPU

NPU IP for Embedded ML

The Ceva-NeuPro-Nano is a efficient and self-sufficient Edge NPU designed for Embedded ML applications.

Overview

The Ceva-NeuPro-Nano is a highly efficient and self-sufficient Edge NPU designed for Embedded ML applications.

This Edge NPU, which is the smallest of Ceva’s NeuPro NPU product family, delivers the optimal balance of ultra-low power and high performance in a small area to efficiently execute Embedded ML workloads across AIoT product categories, including Hearables, Wearables, Home Audio, Smart Home, Smart Factory, and more​. Ranging from 10 GOPS up to 200 GOPS per core, Ceva-NeuPro-Nano is designed to enable always-on audio, voice, vision, and sensing use cases in battery-operated devices across a wide array of end markets. Ceva-NeuPro-Nano makes the possibilities enabled by Embedded ML into realities for low cost, energy efficient AIoT devices.

Ceva-NeuPro-Nano is a stand-alone neural processing unit (NPU), not an AI accelerator, and therefore does not require a host CPU/DSP to operate.​ The IP core includes all the processing elements of a standalone NPU, including code execution and memory management. The Ceva-NeuPro-Nano Embedded ML NPU architecture is fully programmable and efficiently executes neural networks, feature extraction, control code and DSP code. It also supports the most advanced machine-learning data types and operators including native transformer computation, sparsity acceleration, and fast quantization to efficiently execute a wide range of neural networks, delivering a highly optimized solution with excellent performance. 

Adding to the solution’s power and efficiency, both NeuPro-Nano cores provide hardware-based decompression of weights, reducing weight memory footprint by up to 80 percent. The Ceva-NPN64 adds hardware sparsity acceleration that can double effective performance. The NeuPro-Nano is fully supported in Ceva-NeuPro Studio, which facilitates importing, compiling, and debugging models from open frameworks such as LiteRT for Microcontrollers and µTVM.

Key features

  • Fully programmable to efficiently execute Neural Networks, feature extraction, signal processing, audio and control code
  • Scalable performance to meet a wide range of use cases, with MAC configurations up to 64 int8 (native 128 of 4×8) MACs per cycle
  • Two NPU configurations to address a wide variety of use cases are available:
    • Ceva-NPN32 with 32 4×8, 32 8×8, 16 16×8, 8 16×16, 4 32×32 MAC operations per cycle
    • Ceva-NPN64 with 128 4×8, 64 8×8, 32 16×8, 16 16×16, 4 32×32 MAC operations per cycle and 2x performance acceleration using 50% weight sparsity (Sparsity Acceleration)
  • Future proof architecture that supports the most advanced ML data types and operators, including 4-bit to 32-bit integer support and native transformer computation
  • Ultimate ML performance for all use cases, with Sparsity Acceleration, acceleration of non-linear activation types, and fast quantization – up to 5 times acceleration of internal re-quantizing tasks
  • Powerful microcontroller and DSP capabilities with a Coremark/MHz score of 6.0
  • Ultra-low memory requirements achieved with Ceva-NetSqueeze™, yielding up to 80% memory footprint reduction through direct processing of compressed model weights without the need for an intermediate decompression stage. NetSqueeze solves a key bottleneck inhibiting the broad adoption of AIoT processors today
  • Ultra-low energy achieved through innovative energy optimization​​s, including dynamic voltage and frequency scaling support tunable for the use-case​, and dramatic energy and bandwidth reduction by distilling computations using weight-sparsity acceleration
  • Complete, simple to use Ceva-NeuPro-Studio AI SDK, optimized to work seamlessly with leading, open-source AI inference frameworks, such as LiteRT for Microcontrollers and µTVM
  • Model Zoo of pre-trained and optimized machine learning models covering Embedded ML audio, voice, vision and sensing use cases
  • Comprehensive portfolio of optimized runtime libraries and off-the-shelf application-specific software

Block Diagram

Benefits

  • The Ceva-NeuPro-Nano NPU family is specially designed to bring the power of AI to the Internet of Things (IoT), through efficient deployment of Embedded ML models on low-power, resource-constrained devices. Ceva-NeuPro-Nano NPUs’ optimized, self-sufficient architecture enables them to deliver superior power efficiency, with a smaller silicon footprint, and optimal performance for Embedded ML workloads, compared to the existing processor solutions, which utilize a combination of CPU or DSP with a separate AI accelerator.
  • The Ceva-NeuPro Studio, which leverages open source AI framework and provides an easy-to-use software development environment, extensive pre-optimized models in the Ceva Model Zoo, and a wide range of runtime libraries, speeds product development for chip designers, OEM’s and software developers.

Applications

  • Consumer IoT
  • Automotive
  • Industrial Automation

Files

Note: some files may require an NDA depending on provider policy.

Specifications

Identity

Part Number
Ceva-NeuPro-Nano
Vendor
Ceva, Inc.

Provider

Ceva, Inc.
HQ: USA
The Smart Edge runs on Ceva! Ceva is the leader in innovative silicon and software IP solutions that enable smart edge products to connect, sense, and infer data more reliably and efficiently. At Ceva, we are passionate about the smart edge. Providing the technology and market expertise our customers need to be successful is what we do best, and we’ve been doing it for over 30 years. With the industry’s only portfolio of comprehensive communications and scalable edge AI IP, Ceva powers the connectivity, sensing, and inference in today’s most advanced smart edge products across consumer IoT, mobile, automotive, infrastructure, industrial, and personal computing. More than 17 billion of the world’s most innovative smart edge products from smartphones to drones to cellular base stations and more are powered by Ceva. We create innovative technologies that help our customers turn great ideas into extraordinary products. We license our portfolio of wireless communications and scalable edge AI IP to our customers, breaking down barriers to entry and enabling them to bring new cutting-edge products to market faster, more reliably, efficiently, and economically. Ceva is a trusted partner to over 400 of the leading semiconductor and OEM companies including Actions, Artosyn, ASR, Atmosic, Autotalks, Beken, Bestechnic, Brite, Broadcom, Celeno, Ceragon, Cirrus Logic, Dialog Semiconductor, DSP Group, Espressif, FujiFilm, GCT Semi, iCatch, InPlay, Intel, Itron, Leadcore, LG Electronics, Mediatek, Microchip, Nextchip, Nokia, Novatek, NXP, ON Semiconductor, Optek, Oticon, Panasonic, RDA, Renesas, Rockchip, Rohm, Samsung, Sanechips, Sharp, Siflower, SigmaStar, Socionext, Sony, Socionext, Sonova, STMicroelectronics, Toshiba, Unisoc, Vatics, Yamaha and ZTE all leverage Ceva’s industry-leading IP. These companies incorporate our IP into application-specific integrated circuits (“ASICs”) and application-specific standard products (“ASSPs”) that they manufacture, market and sell to consumer electronics companies. Headquartered in Rockville, Maryland, Ceva has over 400 employees worldwide, with design centers in Israel, Ireland, France, United Kingdom, United States, Serbia, and sales and support offices located in Europe, the U.S. and throughout Asia. Ceva is a sustainable and environmentally conscious company, adhering to our Code of Business Conduct and Ethics. As such, we emphasize and focus on environmental preservation, recycling, the welfare of our employees and privacy – which we promote on a corporate level. At Ceva, we are committed to social responsibility, values of preservation and consciousness towards these purposes.

Learn more about NPU IP core

Heterogeneous NPU Data Movement Tax: Intel's Own Slides Tell the Story

At Quadric, we have long argued that heterogeneous NPU designs — those that stitch together multiple specialized fixed-function engines — carry an unavoidable hidden cost: data has to move. A lot. And data movement burns power, adds latency, and creates silicon-area overhead that scales with every new generation of AI models. Now, Intel has made that case for us.

The Upcoming NPU Shakeout

The IP industry is no stranger to boom and bust cycles, and it looks to be at the crest of another wave.

Frequently asked questions about NPU IP cores

What is NPU IP for Embedded ML?

NPU IP for Embedded ML is a NPU IP core from Ceva, Inc. listed on Semi IP Hub.

How should engineers evaluate this NPU?

Engineers should review the overview, key features, supported foundries and nodes, maturity, deliverables, and provider information before shortlisting this NPU IP.

Can this semiconductor IP be compared with similar products?

Yes. Buyers can compare this product with similar semiconductor IP cores or IP families based on category, provider, process options, and structured technical specifications.

×
Semiconductor IP