How 5G is Driving AI at the Edge
Introduction
The ongoing transition from 4G to 5G is driving major infrastructure upgrades that include the integration of AI and machine learning capabilities at the edge. This is due to several major factors, the most important of which is the relentless growth in the amount of the world’s digital data. According to a recent Forbes article, approximately 2.5 quintillion bytes of data are created each day. By 2020, DOMO estimates that for every person on earth, 1.7 MB of data will be created every second.
Beyond the incredible rate of global data growth, carriers see 5G as a lucrative opportunity to generate new revenue streams and bolster the average revenue per user (ARPU). Neural networks and machine learning will continue playing prominent roles in supporting a range of low-latency, bandwidth-intensive applications at the edge including augmented reality, virtual reality, the IoT and Industry 4.0. As such, companies like Nvidia, Google, Intel and ARM are all shipping AI-optimized edge computing platforms.
To read the full article, click here
Related Semiconductor IP
Related Blogs
- How SiFive is Driving AI and Datacenter Innovation
- Enabling AI Innovation at The Far Edge
- X100 - Securing the System - RISC-V AI at the Edge
- Rethinking Edge AI Interconnects: Why Multi-Protocol Is the New Standard
Latest Blogs
- Synopsys Advances Die‑to‑Die Connectivity with 64G UCIe IP Tape‑Out
- The 5 Biggest Challenges in Modern SoC Design (And How to Solve Them)
- Can Your NPU Run DOOM? Chimera Can.
- Importance Of Hardware Security Verification In Pre-Silicon Design
- Arteris × XuanTie: The “Data Highway” for High-Performance RISC-V SoCs