A New DSP Approach to Accelerate 5G and AI Design Development
By Lauro Rizzatti
January 14, 2019
VSORA, a startup from Paris, recently emerged from stealth mode with a new approach to accelerate 5G broadband design. Khaled Maalej, its founder and CEO, and I recently talked about VSORA’s multicore digital signal processing (DSP) intellectual property (IP) and development flow for 5G and AI applications.
Serial entrepreneur Maalej and VSORA’s founders come from DiBcom, a Parisian startup (now part of Parrot) that designed chipsets for low-power mobile TV and radio reception. Over several years at DiBcom, they noticed that the development of high-bandwidth baseband systems was not evolving and improving. They determined that this was due to the inadequacy of the heavily sequential development flow that forced a time-consuming feedback loop between algorithmic designers and DSP hardware developers.
As Maalej continued, he explained that algorithmic engineers would devise an algorithm in the MATLAB environment. Once they trusted that their creation provided satisfactory results, they would freeze the algorithm and pass it over to the DSP developers for implementation. That group would spend a couple of months implementing a thoroughly verified register transfer level (RTL) design of the algorithm.
To read the full article, click here
Related News
- Cadence Unleashes ChipStack AI Super Agent, Pioneering a New Frontier in Chip Design and Verification
- Seligman Ventures Leads Cognichip’s $60M Series A to Back Physics-Informed AI for Chip Design, Intel CEO Lip-Bu Tan and Seligman Ventures’ Umesh Padval Join the Board
- OpenTitan Introduces New Membership Tiers and Deliverables to Accelerate Deployment
- Lightmatter and Cadence Collaborate to Accelerate Optical Interconnect for AI Infrastructure
Latest News
- SiMa.ai Secures Strategic Investment from Micron to Scale High-Performance, Power-Efficient Physical AI
- Codasip announces strategic pivot and divestiture
- UMC Reports Sales for March 2026
- Semidynamics Secures a Strategic Investment to Advance Memory-Centric AI Inference Chips
- Ultra Accelerator Link™ (UALink™) Consortium Publishes Four Specifications Defining In-Network Compute, Chiplets, Manageability and 200G Performance