How 5G is Driving AI at the Edge
Introduction
The ongoing transition from 4G to 5G is driving major infrastructure upgrades that include the integration of AI and machine learning capabilities at the edge. This is due to several major factors, the most important of which is the relentless growth in the amount of the world’s digital data. According to a recent Forbes article, approximately 2.5 quintillion bytes of data are created each day. By 2020, DOMO estimates that for every person on earth, 1.7 MB of data will be created every second.
Beyond the incredible rate of global data growth, carriers see 5G as a lucrative opportunity to generate new revenue streams and bolster the average revenue per user (ARPU). Neural networks and machine learning will continue playing prominent roles in supporting a range of low-latency, bandwidth-intensive applications at the edge including augmented reality, virtual reality, the IoT and Industry 4.0. As such, companies like Nvidia, Google, Intel and ARM are all shipping AI-optimized edge computing platforms.
To read the full article, click here
Related Semiconductor IP
Related Blogs
- How Is AI Driving the Next Innovation Wave for Electronic Design?
- The Future of Driving: How Advanced DSP is Shaping Car Infotainment Systems
- Real-Time Intelligence for Physical AI at the Edge
- How SiFive is Driving AI and Datacenter Innovation
Latest Blogs
- The Growing Importance of PVT Monitoring for Silicon Lifecycle Management
- Unlock early software development for custom RISC-V designs with faster simulation
- HBM4 Boosts Memory Performance for AI Training
- Using AI to Accelerate Chip Design: Dynamic, Adaptive Flows
- Locking When Emulating Xtensa LX Multi-Core on a Xilinx FPGA