Sagence AI Emerges from Stealth Tackling Economic Viability of Inference Hardware for Generative AI
High performance AI inference compute at astounding 100X lower MAC power and 20X lower cost
Santa Clara, CA — November 19, 2024 — Sagence AI™ today emerged from stealth unveiling a groundbreaking advanced analog in-memory compute architecture that directly addresses the untenable power/performance/price and environmental sustainability conundrum facing AI inferencing. Driven by its industry-first architectural innovations using analog technology, Sagence AI makes possible multiple orders of magnitude improvement in energy efficiency and cost reductions, while sustaining performance equivalent to high performance GPU/CPU based systems.
Compared to industry’s leading volume GPU processing the Llama2-70B large language model with performance normalized to 666K tokens/sec, Sagence technology performs with 10X lower power, 20X lower price, and 20X smaller rack space. Using a modular chiplet architecture for maximum integration, Sagence technology makes possible a highly efficient inference machine that scales from data center generative AI to edge computer visions applications across multiple industries. This previously unimaginable balance of high performance and low power at affordable cost addresses the growing ROI problem for generative AI applications at scale, as AI compute in the data center shifts from training models to deployment of models to inference tasks.
“A fundamental advancement in AI inference hardware is vital to the future of AI. Use of large language models (LLMs) and Generative AI drives demand for rapid and massive change at the nucleus of computing, requiring an unprecedented combination of highest performance at lowest power and economics that match costs to the value created,” said Vishal Sarin, CEO & Founder, Sagence AI. “The legacy computing devices today that are capable of extreme high performance AI inference cost too much to be economically viable and consume too much energy to be environmentally sustainable. Our mission is to break those performance and economic limitations in an environmentally responsible way.”
“The demands of the new generation of AI models have resulted in accelerators with massive on-package memory and consequently extremely high-power consumption. Between 2018 and today, the most powerful GPUs have gone from 300W to 1200W, while top-tier server CPUs have caught up to the power consumption levels of NVIDIA’s A100 GPU from 2020,” said Alexander Harrowell, Principal Analyst, Advanced Computing, Omdia. “This has knock-on effects for data center cooling, electrical distribution, AI applications’ unit economics, and much else. One way out of the bind is to rediscover analog computing, which offers much lower power consumption, very low latency, and permits working with mature process nodes.”
On the Frontier of Analog In-memory Compute
Sagence AI leads the industry on the frontier of in-memory compute innovation. Sagence technology is the first to do deep subthreshold compute inside multi-level memory cells, an unprecedented combination that opens doors to the orders of magnitude improvements necessary to deliver inference at scale. As digital technology reaches limits in ability to scale power and cost, Sagence innovated a new analog path forward leveraging the inherent benefits of analog in energy efficiency and costs to make possible mass adoption of AI that is both economically viable and environmentally sustainable.
In-memory Computing Aligned to AI Inference
In-memory computing aligns closely with the essential elements of efficiency in AI inference applications. Merging storage and compute inside memory cells eliminates single-purpose memory storage and complex scheduled multiply-accumulate circuits that run the vector-matrix multiplication integral to AI computing. The resulting chips and systems are much simpler, lower cost, lower power and with vastly more compute capability.
Sagence views the AI inference challenge not as a general-purpose computing problem, but a mathematically intensive data processing problem. Managing the massive amount of arithmetic processing needed to “run” a neural network on CPU/GPU digital machines requires extremely complicated hardware reuse and hardware scheduling. The natural hardware solution is not a general-purpose computing machine, rather an architecture that more closely mirrors how biological neural networks operate.
Reduced Software Complexity
The statically scheduled deep subthreshold in-memory compute architecture employed by Sagence chips is much simpler and eliminates the variabilities and complexities of the dynamic scheduling required of CPUs and GPUs. Dynamic scheduling places extreme demands on the SDK to generate the runtime code and contributes to cost and power inefficiencies. The Sagence AI design flow imports a trained neural network using standards-based interfaces like PyTorch, ONNX and TensorFlow, and automatically converts it into Sagence format. The Sagence system receives the neural network long after GPU software created it, negating further need of the GPU software.
About Sagence AI
Sagence AI, formerly known as Analog Inference, innovates fundamental semiconductor technology to fully unlock the untapped potential of Analog in-memory compute methods to meet the unprecedented performance, power and cost requirements of pervasive AI computing. The Sagence technology team, comprised of multiple 30-year career experts in compute, memory, and AI technologies, developed the final frontier of true in-memory compute to solve the economic viability and sustainability issues facing AI inference deployments. Sagence AI has US $58M in funding from strategic and venture investors. The company was seeded by marque silicon valley investors – Vinod Khosla (Khosla Ventures), Andy Bechtolsheim (founder SUN Microsystems), and Atiq Raza (former President COO, AMD). Series A and B investors include TDK Ventures, P7 Ventures (Aramco Ventures), Blue Ivy Ventures, Cambium Capital Partners, and New Science Ventures. Sagence is headquartered in Santa Clara, CA.
Related Semiconductor IP
- Ultra Low Power AI core
- High-performance AI dataflow processor with scalable vector compute capabilities
- NPU IP for Embedded AI
- NPU / AI accelerator with emphasis in LLM
- AI accelerator
Related News
- Analog Compute is Key to The Next Era of AI Innovation
- Andes Technology and TetraMem Collaborate to Build Groundbreaking AI Accelerator Chip with Analog In-Memory Computing
- Google Cloud Delivers Customized Silicon Powered by Arm Neoverse for General-Purpose Compute and AI Inference Workloads
- Alphawave Semi Tapes Out Industry-First, Multi-Protocol I/O Connectivity Chiplet for High-Performance Compute and AI Infrastructure
Latest News
- Cadence Unveils Arm-Based System Chiplet
- Frontgrade Gaisler Unveils GR716B, a New Standard in Space-Grade Microcontrollers
- Blueshift Memory launches BlueFive processor, accelerating computation by up to 50 times and saving up to 65% energy
- What RISC-V Means for the Future of Chip Development
- CXL Fabless Startup Panmnesia Secures Over $60M in Series A Funding, Aiming to Lead the CXL Switch Silicon Chip and CXL IP