Reviewing different Neural Network Models for Multi-Agent games on Arm using Unity
During the Game Developer Conference (GDC) in March 2023, we showcased our multi-agent demo called Candy Clash, a mobile game containing 100 intelligent agents. In the demo, the agents are developed using Unity’s ML-Agents Toolkit which allows us to train them using reinforcement learning (RL). To find out more about the demo and its development, see our previous blog series. Previously, the agents had a simple Multi-Layer Perceptron (MLP) Neural Network (NN) model. This blog explores the impact of using other types of neural networks models on the gaming experience and performance.
To read the full article, click here
Related Semiconductor IP
- Special Purpose Low (Statistical) offset Operation Amplifier
- Rail to Rail Input and Output Operational Amplifier
- Special Purpose Low offset Operational Amplifier
- Special Purpose Low offset Operational Amplifier
- High Current, Low offset fast Operation Amplifier
Related Blogs
- Benefit of pruning and clustering a neural network for before deploying on Arm Ethos-U NPU
- Neural Network Model quantization on mobile
- Optimizing AI models for Arm Ethos-U NPUs using the NVIDIA TAO Toolkit
- Develop Software for the Cortex-M Security Extensions Using Arm DS and Arm GNU Toolchain
Latest Blogs
- Rivos and Canonical partner to deliver scalable RISC-V solutions in Data Centers and enable an enterprise-grade Ubuntu experience across Rivos platforms
- ReRAM-Powered Edge AI: A Game-Changer for Energy Efficiency, Cost, and Security
- Ceva-XC21 and Ceva-XC23 DSPs: Advancing Wireless and Edge AI Processing
- Cadence Silicon Success of UCIe IP on Samsung Foundry’s 5nm Automotive Process
- Empowering your Embedded AI with 22FDX+