Semidynamics Announces Cervell™ All-in-One RISC-V NPU Delivering Scalable AI Compute for Edge and Datacenter Applications
New fully programmable Neural Processing Unit (NPU) combines CPU, Vector, and Tensor processing to deliver up to 256 TOPS for LLMs, Deep Learning, and Recommendation Systems.
Barcelona, Spain – 6 May 2025 – Semidynamics, the only provider of fully customizable RISC-V processor IP, announces Cervell™, a scalable and fully programmable Neural Processing Unit (NPU) built on RISC-V. Cervell combines CPU, vector, and tensor capabilities in a single, unified all-in-one architecture, unlocking zero-latency AI compute across applications from edge AI to datacenter-scale LLMs.
Delivering up to 256 TOPS (Tera Operations Per Second) at 2GHz, Cervell scales from C8 to C64 configurations, allowing designers to tune performance to application needs — from 8 TOPS INT8 at 1GHz in compact edge deployments to 256 TOPS INT4 in high-end AI inference.
Says Roger Espasa, CEO of Semidynamics:
“Cervell is designed for a new era of AI compute — where off-the-shelf solutions aren’t enough. As an NPU, it delivers the scalable performance needed for everything from edge inference to large language models. But what really sets it apart is how it’s built: fully programmable, with no lock-in thanks to the open RISC-V ISA, and deeply customizable down to the instruction level. Combined with our Gazillion Misses™ memory subsystem, Cervell removes traditional data bottlenecks and gives chip designers a powerful foundation to build differentiated, high-performance AI solutions.”
Why NPUs Matter
AI is rapidly becoming a core differentiator across industries — but traditional compute architectures weren’t built for its demands. NPUs are purpose-designed to accelerate the types of operations AI relies on most, enabling faster insights, lower latency, and greater energy efficiency. For companies deploying large models or scaling edge intelligence, NPUs are the key to unlocking performance without compromise.
Cervell NPUs are purpose-built to accelerate matrix-heavy operations, enabling higher throughput, lower power consumption, and real-time response. By integrating NPU capabilities with standard CPU and vector processing in a unified architecture, designers can eliminate latency and maximize performance across diverse AI tasks, from recommendation systems to deep learning pipelines.
Unlocking High-Bandwidth AI Performance
Cervell is tightly integrated with Gazillion Misses™, Semidynamics’ breakthrough memory management subsystem. This enables:
- Up to 128 simultaneous memory requests, eliminating latency stalls
- Over 60 bytes/cycle of sustained data streaming
- Massively parallel access to off-chip memory, essential for large model inference and sparse data processing
The result is an NPU architecture that maintains full pipeline saturation, even in bandwidth-heavy applications like recommendation systems and deep learning.
Built to Customer Specifications
Like all Semidynamics cores, Cervell is fully customizable and customers may:
- Add scalar or vector instructions
- Configure scratchpad memories and custom I/O FIFOs
- Define memory interfaces and synchronization schemes
- Request bespoke features to suit your application
As demand grows for differentiated AI hardware, chip designers are increasingly looking for ways to embed proprietary features directly into their processor cores. While many IP providers offer limited configurability from fixed option sets, Semidynamics takes a different approach — enabling deep customization at the RTL level, including the insertion of customer-defined instructions. This allows companies to integrate their unique “secret sauce” directly into the solution protecting their ASIC investment from imitation and ensuring the design is fully optimized for power, performance, and area. With a flexible development model that includes early FPGA drops and parallel verification, Semidynamics helps customers accelerate time-to-market while reducing project risk.
This flexibility, combined with RISC-V openness, ensures customers are never locked in — and always in control.
Cervell At-a-Glance
Configuration | INT8 @ 1GHz | INT4 @ 1GHz | INT8 @ 2GHz | INT4 @ 2GHz |
---|---|---|---|---|
C8 | 8 TOPS | 16 TOPS | 16 TOPS | 32 TOPS |
C16 | 16 TOPS | 32 TOPS | 32 TOPS | 64 TOPS |
C32 | 32 TOPS | 64 TOPS | 64 TOPS | 128 TOPS |
C64 | 64 TOPS | 128 TOPS | 128 TOPS | 256 TOPS |
Related Semiconductor IP
- All-In-One RISC-V NPU
- NPU
- Image Processing NPU IP
- Optional extension of NPX6 NPU tensor operations to include floating-point support with BF16 or BF16+FP16
- NPU IP for Data Center and Automotive
Related News
- ESWIN Computing Pairs SiFive CPU, Imagination GPU and In House NPU in Latest RISC-V Edge Computing SoC
- Semidynamics on major recruitment drive for RISC-V software engineers
- UPMEM selects Semidynamics RISC-V AI IP for Large Language Model Application
- Semidynamics: The landscape for RISC-V and AI compute
Latest News
- Numem Appoints Former Intel Executives to Leadership Team
- Agile Analog appoints CEO to drive growth
- Semidynamics Announces Cervell™ All-in-One RISC-V NPU Delivering Scalable AI Compute for Edge and Datacenter Applications
- The European Chips Act has no clothes
- Enabled on makeChip and powered by Racyics, the SpiNNaker2 chip forms the core of the newly launched SpiNNcould supercomputer!