Untether Unveils 2-PFLOPS AI Chip, Edge Roadmap
By Sally Ward-Foxton, EETimes (August 23, 2022)
At Hot Chips this week, Untether unveiled its second-gen architecture for AI inference, the first chip using this architecture, as well as plans to expand to edge and endpoint accelerators.
Untether’s new architecture, internally codenamed Boqueria, addresses trends for very large neural networks, including transformer networks in natural language processing and beyond, endpoint applications that require power efficiency, and applications that require performance and power efficiency combined with prediction accuracy.
The first chip to use the Boqueria architecture, SpeedAI, is a data center inference accelerator capable of 2 PFLOPS of FP8 performance running at peak power consumption (66 W), or 30 TFLOPS/W based on a more usual 30-35 W power envelope. (Untether’s first generation chip, RunAI, could handle 500 TOPS of INT8.)
To read the full article, click here
Related Semiconductor IP
- NPU IP Core for Mobile
- MSP7-32 MACsec IP core for FPGA or ASIC
- UHF RFID tag IP with 3.6kBit EEPROM and -18dBm sensitivity
- NPU IP Core for Edge
- Specialized Video Processing NPU IP
Related News
- Gelsinger Invests In British AI Chip Startup Fractile
- OpenAI Reportedly to Finalize In-House AI Chip Design Soon, Set for TSMC’s 3nm Production
- Q.ANT and IMS CHIPS Launch Production of High-Performance AI Chips, Establish Blueprint for Strengthening Chip Sovereignty
- sureCore PowerMiser IP enables KU Leuven chip for AI applications to achieve dynamic power saving of greater than 40%
Latest News
- Numem Addresses AI’s Dirty Secret: Memory Is the Real Bottleneck
- proteanTecs Gains Momentum as the Trusted Choice for Lifecycle Monitoring of AI Hardware
- TSMC May 2025 Revenue Report
- Momentum Builds on RISC-V European Adoption
- FortifyIQ Expands its Cybersecurity Market Reach with a New Business Organization