How 5G is Driving AI at the Edge
Introduction
The ongoing transition from 4G to 5G is driving major infrastructure upgrades that include the integration of AI and machine learning capabilities at the edge. This is due to several major factors, the most important of which is the relentless growth in the amount of the world’s digital data. According to a recent Forbes article, approximately 2.5 quintillion bytes of data are created each day. By 2020, DOMO estimates that for every person on earth, 1.7 MB of data will be created every second.
Beyond the incredible rate of global data growth, carriers see 5G as a lucrative opportunity to generate new revenue streams and bolster the average revenue per user (ARPU). Neural networks and machine learning will continue playing prominent roles in supporting a range of low-latency, bandwidth-intensive applications at the edge including augmented reality, virtual reality, the IoT and Industry 4.0. As such, companies like Nvidia, Google, Intel and ARM are all shipping AI-optimized edge computing platforms.
To read the full article, click here
Related Semiconductor IP
Related Blogs
- Real-Time Intelligence for Physical AI at the Edge
- How SiFive is Driving AI and Datacenter Innovation
- How Physical AI Is Redefining the Automotive Industry
- Enabling AI Innovation at The Far Edge
Latest Blogs
- Enabling End-to-End EDA Flow on Arm-Based Compute for Infrastructure Flexibility
- Real PPA improvements from analog IC migration
- Design specification: The cornerstone of an ASIC collaboration
- The importance of ADCs in low-power electrocardiography ASICs
- VESA Adaptive-Sync V2 Operation in DisplayPort VIP