Cadence Collaboration with Kudan and Visionary.ai Enables Rapid Deployment of VSLAM and AI ISP-Based Solutions
Are you confused while navigating new environments, especially in less optimum light conditions?
I do, and the problem worsens when lighting conditions are lousy and GPS connectivity is poor! Then, I try to locate my destination by looking at some landmarks.
Likewise, robots and self-driving vehicles use simultaneous localization and mapping (SLAM) and digital imaging to navigate unknown environments, even in the worst lighting conditions. SLAM and digital imaging work as an embedded vision for robots and self-driving vehicles as they make maps (especially while working indoors) of unknown environments while navigating through them. Such advancements incorporate different kinds of sensors and computer vision-based algorithms for SLAM, 3D object detection, tracking, and trajectory estimation. These computational imaging algorithms are very complex, involve extensive computational resources, and may lead to increased latency and power requirements. Digital Imaging is another major trend plagued by challenges such as low light, high/wide dynamic range (HDR/WDR) environments, and fast-moving objects. These are the most significant barriers to clear and crisp imaging. To best address these challenges, Cadence has signed up with Kudan and Visionary.ai for its Tensilica software partner ecosystem, bringing SLAM and AI image signal processor (ISP) to its processor cores. This partnership helps achieve the best performance in various segments such as advanced automotive, mobile, consumer and IoT, and drones. The proof is in the pudding; Cadence strengthens Tensilica Vision and AI software partner ecosystem for advanced automotive, mobile, consumer, and IoT Applications. While Tensilica Vision Q7 DSP helped achieve a nearly 15% speedup of Kudan's proprietary SLAM implementation pipeline compared to CPU-based implementation, Tensilica NNA110 accelerator helps customers implement a camera pipeline with a resolution of more than full HD at over 30fps.
How do Robots/Self-Driving Vehicles Navigate? And Why do We Need AI-Based Image Signal Processors (ISP)?
Related Semiconductor IP
- AES GCM IP Core
- High Speed Ethernet Quad 10G to 100G PCS
- High Speed Ethernet Gen-2 Quad 100G PCS IP
- High Speed Ethernet 4/2/1-Lane 100G PCS
- High Speed Ethernet 2/4/8-Lane 200G/400G PCS
Related Blogs
- AI Requires Tailored DRAM Solutions: Part 1
- AI Requires Tailored DRAM Solutions: Part 2
- AI Requires Tailored DRAM Solutions: Part 3
- AI Requires Tailored DRAM Solutions: Part 4
Latest Blogs
- Why Choose Hard IP for Embedded FPGA in Aerospace and Defense Applications
- Migrating the CPU IP Development from MIPS to RISC-V Instruction Set Architecture
- Quintauris: Accelerating RISC-V Innovation for next-gen Hardware
- Say Goodbye to Limits and Hello to Freedom of Scalability in the MIPS P8700
- Why is Hard IP a Better Solution for Embedded FPGA (eFPGA) Technology?