Semidynamics Announces Cervell™ All-in-One RISC-V NPU Delivering Scalable AI Compute for Edge and Datacenter Applications

New fully programmable Neural Processing Unit (NPU) combines CPU, Vector, and Tensor processing to deliver up to 256 TOPS for LLMs, Deep Learning, and Recommendation Systems.

Barcelona, Spain – 6 May 2025 – Semidynamics, the only provider of fully customizable RISC-V processor IP, announces Cervell™, a scalable and fully programmable Neural Processing Unit (NPU) built on RISC-V. Cervell combines CPU, vector, and tensor capabilities in a single, unified all-in-one architecture, unlocking zero-latency AI compute across applications from edge AI to datacenter-scale LLMs.

Delivering up to 256 TOPS (Tera Operations Per Second) at 2GHz, Cervell scales from C8 to C64 configurations, allowing designers to tune performance to application needs — from 8 TOPS INT8 at 1GHz in compact edge deployments to 256 TOPS INT4 in high-end AI inference.

Says Roger Espasa, CEO of Semidynamics:
“Cervell is designed for a new era of AI compute — where off-the-shelf solutions aren’t enough. As an NPU, it delivers the scalable performance needed for everything from edge inference to large language models. But what really sets it apart is how it’s built: fully programmable, with no lock-in thanks to the open RISC-V ISA, and deeply customizable down to the instruction level. Combined with our Gazillion Misses™ memory subsystem, Cervell removes traditional data bottlenecks and gives chip designers a powerful foundation to build differentiated, high-performance AI solutions.”

Why NPUs Matter

AI is rapidly becoming a core differentiator across industries — but traditional compute architectures weren’t built for its demands. NPUs are purpose-designed to accelerate the types of operations AI relies on most, enabling faster insights, lower latency, and greater energy efficiency. For companies deploying large models or scaling edge intelligence, NPUs are the key to unlocking performance without compromise.

Cervell NPUs are purpose-built to accelerate matrix-heavy operations, enabling higher throughput, lower power consumption, and real-time response. By integrating NPU capabilities with standard CPU and vector processing in a unified architecture, designers can eliminate latency and maximize performance across diverse AI tasks, from recommendation systems to deep learning pipelines.

Unlocking High-Bandwidth AI Performance

Cervell is tightly integrated with Gazillion Misses™, Semidynamics’ breakthrough memory management subsystem. This enables:

  • Up to 128 simultaneous memory requests, eliminating latency stalls
  • Over 60 bytes/cycle of sustained data streaming
  • Massively parallel access to off-chip memory, essential for large model inference and sparse data processing

The result is an NPU architecture that maintains full pipeline saturation, even in bandwidth-heavy applications like recommendation systems and deep learning.

Built to Customer Specifications

Like all Semidynamics cores, Cervell is fully customizable and customers may:

  • Add scalar or vector instructions
  • Configure scratchpad memories and custom I/O FIFOs
  • Define memory interfaces and synchronization schemes
  • Request bespoke features to suit your application

As demand grows for differentiated AI hardware, chip designers are increasingly looking for ways to embed proprietary features directly into their processor cores. While many IP providers offer limited configurability from fixed option sets, Semidynamics takes a different approach — enabling deep customization at the RTL level, including the insertion of customer-defined instructions. This allows companies to integrate their unique “secret sauce” directly into the solution protecting their ASIC investment from imitation and ensuring the design is fully optimized for power, performance, and area. With a flexible development model that includes early FPGA drops and parallel verification, Semidynamics helps customers accelerate time-to-market while reducing project risk.

This flexibility, combined with RISC-V openness, ensures customers are never locked in — and always in control.

Cervell At-a-Glance

Configuration INT8 @ 1GHz INT4 @ 1GHz INT8 @ 2GHz INT4 @ 2GHz
C8 8 TOPS 16 TOPS 16 TOPS 32 TOPS
C16 16 TOPS 32 TOPS 32 TOPS 64 TOPS
C32 32 TOPS 64 TOPS 64 TOPS 128 TOPS
C64 64 TOPS 128 TOPS 128 TOPS 256 TOPS
×
Semiconductor IP