Vendor: Expedera Category: NPU

Neural engine IP - Tiny and Mighty

Small, low-power dedicated AI engines are essential for home appliances, security cameras, and always-on smartphone features.

Overview

Small, low-power dedicated AI engines are essential for home appliances, security cameras, and always-on smartphone features. Customized for specific use cases, Origin™ E1 delivers targeted low-power performance and requires little to no external memory.

Perfect-Fit Solutions

The Origin E1 NPUs are individually customized to various neural networks commonly deployed in edge devices, including home appliances, smartphones, and security cameras. For products like these that require dedicated AI processing that minimizes power consumption, silicon area, and system cost, E1 cores offer the lowest power consumption and area in a 1 TOPS engine.

Power-Sipping, Always-Sensing AI

Always-sensing cameras continuously sample and analyze visual data to identify specific triggers relevant to the user experience. They enable a seamless, more natural user experience. However, always-sensing data requires specialized AI processing due to the quantity and complexity of data generated. OEMs are turning to specialized AI engines like Expedera’s LittleNPU. The LittleNPU is optimized to process the low-power, high-quality neural networks used by leading OEMs in always-sensing applications. It runs at low power—often as low as 10-20mW—and keeps all camera data securely within the LittleNPU subsystem to preserve user privacy.

Innovative Architecture

The Origin E1 neural engines use Expedera’s unique packet-based architecture, which enables parallel execution across multiple layers, achieving better resource utilization and deterministic performance. This innovative approach significantly increases performance while lowering power, area, and latency.

Specifications

Compute Capacity 0.5K INT8 MACs
Multi-tasking Run Simultaneous Jobs
Power Efficiency 18 TOPS/W effective; no pruning, sparsity or compression required (though supported)
Example Networks Supported MobileNet, EfficientNet, NanoDet, PicoDet, Inception V3, RNN-T, MobileNet SSD, BERT, FSR CNN, CPN, CenterNet, Unet, YOLO V3, ShuffleNet2, others
Layer Support Standard NN functions, including Conv, Deconv, FC, Activations, Reshape, Concat, Elementwise, Pooling, Softmax, others.
Data types INT4/INT8/INT10/INT12/INT16 Activations/Weights
Quantization Channel-wise Quantization (TFLite Specification)
Software toolchain supports Expedera, customer-supplied, or third-party quantization
Latency Deterministic performance guarantees, no back pressure
Frameworks TensorFlow, TFlite, ONNX, others supported

Key features

  • Choose the Features You Need: Customization brings many advantages, including increased performance, lower latency, reduced power consumption, and eliminating dark silicon waste. Expedera works with customers to understand their use case(s), PPA goals, and deployment needs during their design stage. Using this information, we configure Origin IP to create a customized solution that perfectly fits the application.
  • Market-Leading 18 TOPS/W: Sustained power efficiency is key to successful AI deployments. Continually cited as one of the most power-efficient architectures in the market, Origin NPU IP achieves a market-leading, sustained 18 TOPS/W.
  • Efficient Resource Utilization: Origin IP scales from GOPS to 128 TOPS in a single core. The architecture eliminates the memory sharing, security, and area penalty issues faced by lower-performing, tiled AI accelerator engines. Origin NPUs achieve sustained utilization averaging 80%—compared to the 20-40% industry norm—avoiding dark silicon waste.
  • Full TVM-Based Software Stack: Origin uses a TVM-based full software stack. TVM is widely trusted and used by OEMs worldwide. This easy-to-use software allows the importing of trained networks and provides various quantization options, automatic completion, compilation, estimator and profiling tools. It also supports multi-job APIs.
  • Successfully Deployed in 10M Devices: Quality is key to any successful product. Origin IP has successfully deployed in over 10 million consumer devices, with designs in multiple leading-edge nodes.

Block Diagram

Benefits

  • 1 TOPS performance 
  • Support for standard, custom, and proprietary neural networks 
  • Performance efficiencies up to 18 TOPS/Watt
  • Full software stack provided, including compiler, estimator, scheduler, and quantizer
  • Runs CNN, RNN, DNN, LSTM, and other network types
  • Delivered as Soft IP (RTL) or GDS

Applications

  • Home Appliances
  • Smartphone
  • Consumer Electronics
  • Handheld devices

Files

Note: some files may require an NDA depending on provider policy.

Specifications

Identity

Part Number
Origin E1
Vendor
Expedera

Provider

Expedera
HQ: USA
Expedera provides scalable neural engine semiconductor IP that enables major improvements in performance, power, and latency while reducing cost and complexity in AI inference applications. Third-party silicon validated and shipped in more than 10M customer devices, Expedera’s solutions produce superior performance and are scalable to a wide range of applications from edge nodes and smartphones to automotive. Expedera’s Origin™ Neural Processing Unit IP solutions are easily integrated, readily scalable, and customized to unique use cases and application requirements. The company is headquartered in Santa Clara, California, with engineering and sales offices around the globe.

Learn more about NPU IP core

Heterogeneous NPU Data Movement Tax: Intel's Own Slides Tell the Story

At Quadric, we have long argued that heterogeneous NPU designs — those that stitch together multiple specialized fixed-function engines — carry an unavoidable hidden cost: data has to move. A lot. And data movement burns power, adds latency, and creates silicon-area overhead that scales with every new generation of AI models. Now, Intel has made that case for us.

The Upcoming NPU Shakeout

The IP industry is no stranger to boom and bust cycles, and it looks to be at the crest of another wave.

Frequently asked questions about NPU IP cores

What is Neural engine IP - Tiny and Mighty?

Neural engine IP - Tiny and Mighty is a NPU IP core from Expedera listed on Semi IP Hub.

How should engineers evaluate this NPU?

Engineers should review the overview, key features, supported foundries and nodes, maturity, deliverables, and provider information before shortlisting this NPU IP.

Can this semiconductor IP be compared with similar products?

Yes. Buyers can compare this product with similar semiconductor IP cores or IP families based on category, provider, process options, and structured technical specifications.

×
Semiconductor IP