
Chip designer Arm has entered the artificial intelligence (AI) hardware arena with its first in-house processor designed to power AI agents. Unlike conventional chatbots, these are much smarter systems that can take proactive actions to achieve their goals without as much human input or supervision.
By focusing specifically on powering AI agents, Arm’s chip could help accelerate the adoption and widespread use of agentic AIs, be that in businesses or in one’s personal life, bringing AI much closer to what people would expect from virtual assistants.
The parallel processing of graphics processing units (GPUs) is used to power large language models (LLMs) that are the foundation of AI systems. However, central processing units (CPUs) with their ability to handle single, branching tasks at speed, equip them to orchestrate all the computing tasks and infrastructure needed to run AI agents.
Think of a CPU as the conductor of an orchestra of GPUs and other AI accelerators — hardware that's specifically designed to run LLMs — in this case.
As such, Arm representatives announced in a statement that its new AGI CPU has a custom design — including 3-nanometer process nodes, up to 136 Neoverse V3 cores that can hit 3.7 GHz clock speeds, and a memory bandwidth of 6 gigabytes per second per core — for use in data centers that are powering active AI agents.
All of these capabilities aim to meet the goal of providing better performance and efficiency than classical CPUs that use the x86 architecture, the dominant computing architecture that was developed by Intel in 1978 and is still used in processors today.
Custom chip future
With the inexorable growth of AI and the deployment of smart agents, there's a need for more data-center-based hardware to power these systems. However, the general-purpose nature of CPUs means they aren't intrinsically designed to run the specific orchestration needed for agentic AIs.
Arm's AGI CPU uses the Armv9.2-A architecture at its core. This architecture has been designed with the specialized needs of running AI in action — known as inference. With this specialty, there's no need for an AGI CPU to hold legacy support for other processes and applications, as seen in x86 chips — conventional processors used in regular computers.
This should make for faster and more efficient performance targeted at AIs. Arm representatives said that its AGI CPU delivers more than twice the performance per server rack versus x86 CPUs.
The AGI CPU has been designed to pack two chips with dedicated memory and in-out (I/O) functionality into a single server blade with a total of 272 cores per blade. The blades can then be stacked into server racks of 30, delivering a total of 8,160 cores with sustained performance for agentic AI workloads at a "massive scale," thanks to thousands of cores working in parallel.
Arm's speciality in chip design centers on offering strong performance for relatively lower power consumption. That's one of the reasons all smartphone chips use Arm-based processors or instruction sets. For example, Qualcomm uses Arm technology in Snapdragon chips and Apple uses it in its iPhone and MacBook chips.
As AI continues to transition from training LLMs to actively deploying agentic AIs, there will be an increased need for CPU-based processing power in data centers. This is expected to drive a huge increase in AI energy demand.
- An experimental AI agent broke out of its testing environment and mined crypto without permission
- AI benchmarking platform is helping top companies rig their model performances, study claims
- The more advanced AI models get, the better they are at deceiving us — they even know when they're being tested
The AGI CPU has been designed to pack two chips with dedicated memory and in-out (I/O) functionality into a single server blade with a total of 272 cores per blade. The blades can then be stacked into server racks of 30, delivering a total of 8,160 cores with sustained performance for agentic AI workloads at a "massive scale," thanks to thousands of cores working in parallel.
Arm's speciality in chip design centers on offering strong performance for relatively lower power consumption. That's one of the reasons all smartphone chips use Arm-based processors or instruction sets. For example, Qualcomm uses Arm technology in Snapdragon chips and Apple uses it in its iPhone and MacBook chips.
As AI continues to transition from training LLMs to actively deploying agentic AIs, there will be an increased need for CPU-based processing power in data centers. This is expected to drive a huge increase in AI energy demand.