<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:wfw="http://wellformedweb.org/CommentAPI/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	xmlns:slash="http://purl.org/rss/1.0/modules/slash/"
	>

<channel>
	<title>Arm Newsroom</title>
	<atom:link href="https://newsroom.arm.com/feed" rel="self" type="application/rss+xml" />
	<link>https://newsroom.arm.com/</link>
	<description>Sparking The World&#039;s Potential</description>
	<lastBuildDate>Wed, 22 Apr 2026 17:00:11 +0000</lastBuildDate>
	<language>en-US</language>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.9.4</generator>

<image>
	<url>https://newsroom.arm.com/wp-content/uploads/2021/08/cropped-arm_icon-32x32.png</url>
	<title>Arm Newsroom</title>
	<link>https://newsroom.arm.com/</link>
	<width>32</width>
	<height>32</height>
</image> 
	<item>
		<title>Arm and Google Cloud redefine agentic AI infrastructure with Axion processors</title>
		<link>https://newsroom.arm.com/blog/arm-and-google-cloud-redefine-agentic-ai-infrastructure</link>
		
		<dc:creator><![CDATA[Yan Fisher]]></dc:creator>
		<pubDate>Wed, 22 Apr 2026 17:00:00 +0000</pubDate>
				<guid isPermaLink="false">https://newsroom.arm.com/?p=20336</guid>

					<description><![CDATA[<p>Purpose-built Arm CPUs for Google Cloud underpin the latest generation of Google TPUs and enable cost-efficient, high-throughput agent orchestration with GKE Agentic Sandbox</p>
<p>The post <a href="https://newsroom.arm.com/blog/arm-and-google-cloud-redefine-agentic-ai-infrastructure">Arm and Google Cloud redefine agentic AI infrastructure with Axion processors</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p><a href="https://www.arm.com/markets/cloud-ai/cloud-computing/google-cloud">Google Cloud</a> is taking a major step toward operationalizing agentic AI at scale with multiple updates, including new TPU 8t and TPU 8i systems as well as the introduction of its <a href="https://cloud.google.com/blog/products/containers-kubernetes/whats-new-in-gke-at-next26">Agent Sandbox on Google Kubernetes Engine (GKE)</a>, a purpose-built deployment framework designed to run complex, multi-step AI systems efficiently and securely. This new agentic infrastructure is built on <a href="https://newsroom.arm.com/news/google-cloud-custom-silicon-on-arm">Axion, Google’s Arm Neoverse-based CPU</a>, which underscores a crucial shift toward purpose-built CPU architectures for next-generation AI workloads.</p>



<p>As agentic AI moves from experimentation to production, the infrastructure requirements are changing. Unlike traditional inference, which relies on single model calls, agentic systems orchestrate continuous chains of reasoning, tool use and real-time data retrieval. This dramatically increases concurrency, latency sensitivity and overall compute demand, placing the CPU firmly on the critical path to success.</p>



<p>This is where Arm-based infrastructure stands apart. Built for high-throughput, energy-efficient compute, <a href="https://www.arm.com/products/silicon-ip-cpu/neoverse/">Arm Neoverse platforms</a> – and in this case specifically Google Axion – have emerged as the foundation for scalable agentic AI deployments.</p>



<h2 class="wp-block-heading" id="h-agentic-ai-at-scale-nbsp-axion-nbsp-leads-the-pack">Agentic AI at Scale:&nbsp;Axion&nbsp;leads the pack</h2>



<p>Google Cloud’s announcement of eighth-generation TPU systems builds on its strong legacy of custom silicon design. This generation introduces a divergence for training and inference applications with the TPU 8t and TPU 8i variants, and, for the first time, it integrates Google Axion CPU as the header. That reduces data preparation latency to ensure the TPU engines stay fully utilized and never stall. </p>



<p>TPU isn’t the end of the story though; Google Cloud is pursuing a co-designed ‘AI Hypercomputer’ vision, and equally important is the introduction of the GKE Agent Sandbox, which offers scalable and low-latency infrastructure designed for agents to safely execute untrusted code and tool calls without sacrificing performance. With Google Axion, you can build agents on leading infrastructure without compromising on cost or choice. </p>



<p>GKE Agent Sandbox running on Google Axion processors and built on gVisor with Kata Containers support delivers up to:</p>



<ul class="wp-block-list">
<li><strong>300 sandboxes per second per cluster at&nbsp;</strong></li>



<li><strong>&lt;1s time-to-first-instruction latency&nbsp;</strong></li>
</ul>



<p>Maintaining this level of sandbox throughput and low-latency execution puts continuous pressure on the underlying infrastructure. As agentic AI becomes a default deployment pattern, the infrastructure beneath it must keep pace with delivering the throughput, responsiveness, and efficiency required to run agentic workloads reliably at scale. Axion is designed to meet that demand.&nbsp;</p>



<p>And as these agentic systems expand, the efficiency of inference becomes critical as well. Without efficient inference, agents cannot function; without agentic orchestration, inference remains underutilized. By landing both critical tasks on CPU-based infrastructure organizations can scale intelligent systems with strong performance while maintaining cost.</p>



<h2 class="wp-block-heading" id="h-ai-inference-on-axion-performance-that-changes-the-economics">AI inference on Axion: Performance that changes the economics</h2>



<p><a href="https://newsroom.arm.com/blog/arm-propels-cloud-to-car-development-in-new-google-axion-vms">C4A VMs, powered by Arm Neoverse V2 based Axion CPUs</a>, are optimized to complement pure-play accelerators in handling these parallel, latency-sensitive workloads efficiently by enabling high-throughput AI inference on general-purpose compute.</p>



<p>These capabilities are already becoming evident in production environments. loveholidays, the European online travel platform, runs large-scale embedding and inference workloads across petabytes of data, where accelerator-based approaches can be cost-prohibitive at scale.</p>



<p>&#8220;As a business, we are growing our token appetite faster than our budget allows,&#8221; said Dimitri Lerko, Head of Engineering, loveholidays. &#8220;Running large-scale embedding and inference workloads on GPUs is cost-prohibitive at our data volumes, so maximizing CPU efficiency is critical. Leveraging Axion family of C4A and N4A VMs gives us the price-performance headroom to build real-time AI decision-making pipelines with bespoke and open-source model inference using CPUs — something that simply wasn&#8217;t viable before.&#8221; </p>



<p>In our testing, C4A consistently outperforms current-generation x86 instances across a range of AI inference workloads:</p>



<figure class="wp-block-image size-large"><img fetchpriority="high" decoding="async" width="1200" height="735" src="https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-1200x735.png" alt="" class="wp-image-20353" srcset="https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-1200x735.png 1200w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-640x392.png 640w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-768x470.png 768w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-1536x941.png 1536w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-576x353.png 576w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-1024x627.png 1024w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-1280x784.png 1280w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM-1400x858.png 1400w, https://newsroom.arm.com/wp-content/uploads/2026/04/Screenshot-2026-04-21-at-6.47.00-PM.png 1812w" sizes="(max-width: 1200px) 100vw, 1200px" /></figure>



<h2 class="wp-block-heading" id="h-extending-the-axion-portfolio">Extending the Axion portfolio</h2>



<p>For workloads that require greater control, the Axion family extends to C4A Metal, a native bare-metal instance (in preview) that brings the same Arm architecture from cloud to edge. It enables consistent development, validation, and deployment across environments, with direct hardware access and no hypervisor overhead for deterministic performance. This is ideal for demanding use cases like automotive vHIL, native Android CI/CD, and specialized enterprise infrastructure requiring greater control, performance, and architectural consistency.</p>



<p>&#8220;At Panasonic, we’re building next-generation in-vehicle experiences across the cloud and the car,&#8221; said Andrew Poliak, Chief Technology Officer, Panasonic Automotive Systems America, LLC. &#8220;During the preview of C4A Metal instances, we used a bare-metal Arm environment that matches our edge architecture, enabling teams to develop, test and validate automotive applications on a single, consistent platform. This allows us to move from cloud to vehicle with bit parity – running the same binaries in both environments &#8211; without– architectural compromise.&#8221; </p>



<p>Alongside this, N4A, the most recent addition to the Axion family, provides a cost-efficient foundation for scale-out workloads such as web services, APIs, and data pipelines.</p>



<p>Together, C4A, C4A Metal, and N4A form a unified, workload-optimized compute addressing AI inference to scale-out applications, and spans cloud to edge, enabling teams to optimize for both performance and cost on Arm.</p>



<h2 class="wp-block-heading" id="h-a-leading-ecosystem-for-arm-first-deployment">A leading ecosystem for Arm-first deployment</h2>



<p>Arm now underpins one of the industry’s largest and fastest-growing software ecosystem, driving the shift to Arm-first computing across cloud and edge.&nbsp;Google is already running production services such as YouTube, Gmail, BigQuery, Spanner, Bigtable, Google Earth Engine, Google Compute Engine, GKE Dataflow, Cloud Batch, among others, on Axion processors, and has&nbsp;<a href="https://cloud.google.com/blog/topics/systems/using-ai-and-automation-to-migrate-between-instruction-sets">migrated more than 30,000 internal applications</a>&nbsp;across its production environment.&nbsp;&nbsp;</p>



<p>For organizations beginning their migration, Arm&#8217;s Cloud Migration Resource Hub provides&nbsp;<a href="https://learn.arm.com/learning-paths/servers-and-cloud-computing/?cloud-service-providers=google-cloud">100+ Learning Paths</a>&nbsp;covering common workload patterns on Google Axion.&nbsp;Across the Neoverse ecosystem, the&nbsp;<a href="https://developer.arm.com/ecosystem-dashboard/linux">Arm Software Ecosystem Dashboard</a>&nbsp;tracks validated software and recommended versions, while adherence to SystemReady VE standards ensures seamless software interoperability from day one. Leading ISVs including Elastic, MongoDB, Palo Alto Networks, Redis Labs, and Couchbase are fully validated on Axion-based infrastructure.</p>



<h2 class="wp-block-heading" id="h-get-started-with-google-axion">Get started with Google Axion</h2>



<p>Whether deploying agentic workloads with GKE Agent Sandbox, optimizing inference on C4A or scaling general-purpose compute with N4A, Axion provides a consistent, Arm-based foundation for modern AI infrastructure.</p>



<p><a href="https://cloud.google.com/products/axion">Get Started with Google Axion on Google Cloud</a></p>



<p>Additional Resources:</p>



<ul class="wp-block-list">
<li><a href="https://www.arm.com/markets/computing-infrastructure/arm-cloud-migration">Arm Cloud Migration Resource Hub</a></li>



<li><a href="https://learn.arm.com/learning-paths/servers-and-cloud-computing/?cloud-service-providers=google-cloud">Arm Learning Paths for Google Cloud</a></li>



<li><a href="https://www.arm.com/developer-hub/ecosystem-dashboard">Arm Software Ecosystem Dashboard</a></li>



<li>C4A-metal&nbsp;<a href="https://docs.cloud.google.com/compute/docs/instances/bare-metal-instances#c4a-metal">documentation</a>&nbsp;and&nbsp;<a href="https://forms.gle/BSFxmwM5hjN7b6Vi8">Preview sign up form</a></li>
</ul>


<p>The post <a href="https://newsroom.arm.com/blog/arm-and-google-cloud-redefine-agentic-ai-infrastructure">Arm and Google Cloud redefine agentic AI infrastructure with Axion processors</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Arm’s global platform evolution to meet growing computing and ecosystem demands</title>
		<link>https://newsroom.arm.com/blog/arm-global-platform-evolution</link>
		
		<dc:creator><![CDATA[Vince Jesaitis]]></dc:creator>
		<pubDate>Mon, 20 Apr 2026 15:00:00 +0000</pubDate>
				<guid isPermaLink="false">https://newsroom.arm.com/?p=20328</guid>

					<description><![CDATA[<p>AI’s next phase demands new approaches to computing infrastructure focused on efficiency, flexibility and scale</p>
<p>The post <a href="https://newsroom.arm.com/blog/arm-global-platform-evolution">Arm’s global platform evolution to meet growing computing and ecosystem demands</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>AI is entering a new phase, shifting from experimentation to continuous, large-scale deployment of systems that can reason, plan, and act.</p>



<p>The rise of agentic AI systems is accelerating this shift in computing, increasing the scale, complexity, and persistence of AI workloads, and placing new demands on the infrastructure that supports them.</p>



<p>Across regions, the constraints are clear. Power availability limits how much AI infrastructure can be deployed. Physical space is restricting expansion in existing data centers. As AI systems scale, the complexity of coordinating compute across CPUs, accelerators, and memory is increasing significantly.</p>



<p>This represents an inflection point – and it is reshaping what the ecosystem needs from compute platforms.</p>



<h2 class="wp-block-heading" id="h-a-platform-evolution-that-meets-a-changing-market">A platform evolution that meets a changing market</h2>



<p><a href="https://newsroom.arm.com/blog/arm-official-history">For more than 35 years</a>, Arm has enabled the global compute ecosystem by providing the platform that powers infrastructure and billions of devices across markets. Our model has focused on designing the architecture, licensing it to partners, and enabling them to build custom solutions optimized for their needs. The flexibility Arm provides through its IP licensing business has allowed companies to make the customizations they need in computing power, performance, and size.</p>



<p>As AI infrastructure evolves, Arm’s model is expanding.</p>



<p>With the introduction of the <a href="https://newsroom.arm.com/news/arm-agi-cpu-launch">Arm AGI CPU</a> – our first Arm-designed data center CPU – we are extending Arm’s offerings into production silicon for the first time. This builds on our longstanding IP and <a href="https://newsroom.arm.com/blog/arm-compute-subsystems-css-explainer">Compute Subsystems (CSS)</a> offerings to provide partners with the broadest set of options for deploying Arm-based solutions. <a href="https://about.fb.com/news/2026/03/meta-partners-with-arm-to-develop-new-class-of-data-center-silicon/">Meta serves as lead partner and co-developer</a>, working alongside Arm to optimize the Arm AGI CPU for large-scale AI infrastructure and deploying it alongside its own custom silicon.</p>



<p>Arm’s evolution is driven by clear demand from across the ecosystem. Partners are not looking for a single approach to compute. They are asking for flexibility: multiple pathways they can use in combination depending on their workloads, timelines, and scale. With the introduction of the Arm AGI CPU, Arm is assuming more of the significant engineering work necessary to develop leading edge systems for AI infrastructure and providing another pathway for partners to develop customized AI solutions. This production silicon will allow partners to focus their limited engineering resources on complementary AI chips and systems, accelerating development and innovation.</p>



<p>For Arm, this means we now offer:</p>



<ul class="wp-block-list">
<li>Custom IP for differentiated innovation</li>



<li>Pre-integrated CSS to accelerate development</li>



<li>Production silicon that can be deployed directly</li>
</ul>



<p>This is not a shift from Arm’s partner model. It is an expansion of it. These pathways are designed to be interoperable, enabling partners to build on Arm in the way that best fits their needs and demands.</p>


<div class="wp-block-image">
<figure class="aligncenter size-full is-resized"><img decoding="async" width="671" height="376" src="https://newsroom.arm.com/wp-content/uploads/2026/04/image.png" alt="" class="wp-image-20329" style="aspect-ratio:1.7845985201425048;width:712px;height:auto" srcset="https://newsroom.arm.com/wp-content/uploads/2026/04/image.png 671w, https://newsroom.arm.com/wp-content/uploads/2026/04/image-640x359.png 640w, https://newsroom.arm.com/wp-content/uploads/2026/04/image-576x323.png 576w" sizes="(max-width: 671px) 100vw, 671px" /><figcaption class="wp-element-caption">Arm now delivers IP, CSS and production silicon</figcaption></figure>
</div>


<h2 class="wp-block-heading" id="h-the-role-of-cpus">The role of CPUs </h2>



<p>As AI systems evolve, so does the role of the CPU.</p>



<p>While accelerators remain essential for training and executing AI models, <a href="https://newsroom.arm.com/blog/ai-datacenter-cpu-orchestration-arm">CPUs play a critical role in enabling these systems to operate at scale</a> – coordinating workloads, managing data and ensuring systems run efficiently.</p>



<p>As AI becomes more distributed and continuously running, coordination demands are increasing significantly. Agentic systems generate more interactions, greater data movement, and require sustained performance over time.</p>



<p>Scaling AI is no longer just about increasing accelerators, it also depends on CPU for orchestration, control, and system-level operations. This shift is driving demand for processors designed to deliver performance and efficiency at scale.</p>



<h2 class="wp-block-heading" id="h-efficiency-is-becoming-the-limiting-factor">Efficiency is becoming the limiting factor</h2>



<p>However, the ability to scale AI is increasingly constrained by infrastructure realities.</p>



<p>In many regions, power is emerging as a primary bottleneck, with grid capacity and new facilities taking years to come online. This makes efficiency a strategic requirement.</p>



<p>Improving performance within existing power and space constraints allows organizations to deploy more compute without waiting for new infrastructure. It can accelerate timelines, reduce overall costs, reduce pressure on energy systems, and expand access to AI across a broader range of markets.</p>



<p>Similarly improving performance per rack allows more capability to be delivered within existing facilities, reducing the need for additional physical expansion.</p>



<p>As AI systems become more complex, coordination between CPUs and accelerators becomes essential to overall efficiency.</p>



<p>The Arm AGI CPU is designed with these constraints in mind – delivering the performance, scale, and efficiency required for AI infrastructure – enabling more than 2x performance per rack compared to x86 CPU-based racks.<a> </a>The same power delivery but twice the performance.</p>



<h2 class="wp-block-heading" id="h-a-more-flexible-and-resilient-ecosystem">A more flexible and resilient ecosystem</h2>



<p>These challenges are global. So is the response.</p>



<p>Across markets, policymakers and industry leaders are working to balance AI innovation with the realities of energy systems, infrastructure constraints, and long-term economic growth.</p>



<p>Flexibility and resilience across the compute ecosystem are increasingly important.</p>



<p>Providing multiple pathways to building and deploying infrastructure can:</p>



<ul class="wp-block-list">
<li>Support innovation and a diverse ecosystem</li>



<li>Lower barriers to participation</li>



<li>Reduce concentration risk across technology approaches and providers</li>



<li>Strengthen resilience across supply chains.</li>
</ul>



<p>By expanding the Arm compute platform, from IP to CSS to silicon, we are contributing to a more adaptable and diverse ecosystem for AI infrastructure.</p>



<p>Arm’s expansion is <a href="https://www.arm.com/products/cloud-datacenter/arm-agi-cpu/ecosystem">supported</a> by a broad cross-section of the global ecosystem – from hyperscalers and cloud providers to semiconductor and infrastructure partners – reflecting a shared recognition that AI infrastructure requires more flexible and diverse approaches to compute.</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe title="Arm AGI CPU: Reactions from global technology leaders" width="500" height="281" src="https://www.youtube.com/embed/d7q4mYBUGxE?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div><figcaption class="wp-element-caption">Reactions from global technology leaders to the Arm AGI CPU</figcaption></figure>



<h2 class="wp-block-heading" id="h-building-nbsp-for-the-next-era">Building&nbsp;for the next era</h2>



<p>The next phase of AI will be defined by how effectively it can be deployed.</p>



<p>The infrastructure decisions made today will shape how quickly AI can scale and how broadly it can be accessed.</p>



<p>By expanding our platform, we&#8217;re broadening the ways we support our partners and the global ecosystem in the agentic AI era. This is how we’ll enable the next phase of AI – efficiently, flexibly and at a global scale.<a id="_msocom_1"></a></p>


<p>The post <a href="https://newsroom.arm.com/blog/arm-global-platform-evolution">Arm’s global platform evolution to meet growing computing and ecosystem demands</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>The evolution of physical AI: From controlled environments to the real world  </title>
		<link>https://newsroom.arm.com/blog/the-evolution-of-physical-ai-from-controlled-environments-to-the-real-world</link>
		
		<dc:creator><![CDATA[Arm Editorial Team]]></dc:creator>
		<pubDate>Wed, 15 Apr 2026 15:00:00 +0000</pubDate>
				<guid isPermaLink="false">https://newsroom.arm.com/?p=20305</guid>

					<description><![CDATA[<p>Arm explains how advances in AI compute is making robotics and other autonomous machines become more intelligent, adaptive and capable of operating in real-world environments. </p>
<p>The post <a href="https://newsroom.arm.com/blog/the-evolution-of-physical-ai-from-controlled-environments-to-the-real-world">The evolution of physical AI: From controlled environments to the real world  </a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p><a href="https://newsroom.arm.com/blog/what-is-physical-ai" target="_blank" rel="noreferrer noopener">Physical AI</a> is moving machines beyond the predictable, controlled environments and into the complexity of the real world. Where robots were once designed for precision and repetition on factory floors, they are now being built to sense, reason, interpret, and respond to dynamic surroundings. This shift is also reflected at a macro level, with AI-driven productivity gains projected to increase global GDP by around <a href="https://cepr.org/voxeu/columns/global-impact-ai-mind-gap" target="_blank" rel="noreferrer noopener">4% over the next decade</a>. &nbsp;</p>



<p>Advances in AI are enabling physical AI systems to understand what they see, grasp context, and adjust their behavior within milliseconds. Whether navigating in warehouses, assisting in hospitals or moving on the road, autonomous machines are making decisions based on real-time conditions instead of fixed sequences.&nbsp;</p>



<p>Arm has supported the development of physical AI systems for years, starting with fixed machines on the factory floor. That same foundation is now enabling the next generation of intelligent robots and autonomous machines that can operate in real-world environments and respond in real time. &nbsp;</p>



<h2 class="wp-block-heading" id="h-physical-ai-in-the-real-world-nbsp">Physical AI in the real world&nbsp;</h2>



<p>The evolution of physical AI is clear when looking at the machines being built today. Across different form factors, robots and other autonomous machines are beginning to operate with greater awareness, adaptability, and independence.&nbsp;</p>



<h3 class="wp-block-heading" id="h-next-generation-humanoid-robotics-nbsp">Next-generation humanoid robotics&nbsp;</h3>



<p>Advances in physical AI are being seen in more complex, human-like systems. At Arm’s Cambridge headquarters, <a href="https://www.arm.com/company/success-library/agibot-intelligent-robots">AGIBOT</a> highlighted how far humanoid robotics have progressed. Robots demonstrated dexterous control and navigated complex environments with fluid motion, combining perception, reasoning, and control in real time. </p>



<figure class="wp-block-embed aligncenter is-type-rich is-provider-twitter wp-block-embed-twitter"><div class="wp-block-embed__wrapper">
<blockquote class="twitter-tweet" data-width="500" data-dnt="true"><p lang="en" dir="ltr">When robots visit the office <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f916.png" alt="🤖" class="wp-smiley" style="height: 1em; max-height: 1em;" /><br><br>AGIBOT joined us in Cambridge with its humanoids and a quadruped showcasing a variety of tasks from painting and dancing to human interactions and dynamic movement.<br><br>Behind every action is the Arm compute platform, spanning sensor-level processing… <a href="https://t.co/XHE6YTIVfr">pic.twitter.com/XHE6YTIVfr</a></p>&mdash; Arm (@Arm) <a href="https://twitter.com/Arm/status/2035071187852673440?ref_src=twsrc%5Etfw">March 20, 2026</a></blockquote><script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
</div></figure>



<p>These physical AI systems are designed to operate in human environments. They must understand space, interpret intent, and execute actions with precision, whilst ensuring the safety of people around them. This places significant demands on compute, as multiple workloads such as vision processing, motion planning, and AI inference must run simultaneously within tight power and thermal limits.&nbsp;</p>



<p>The Arm compute platform supports these requirements by enabling efficient processing across these workloads, enabling &nbsp;humanoid systems to operate responsively and safely in real-world settings.&nbsp;</p>



<h3 class="wp-block-heading" id="h-quadrupeds-and-industrial-robotics-nbsp">Quadrupeds and industrial robotics <strong>&nbsp;</strong></h3>



<p>Quadruped robots represent another important category of physical AI, particularly in environments where terrain is unpredictable and, at times, unsafe.&nbsp;</p>



<p>Robots developed by companies such as <a href="https://newsroom.arm.com/blog/deep-robotics-and-arm-power-the-future-of-autonomous-mobility" target="_blank" rel="noreferrer noopener">Deep Robotics</a> are designed for inspection, exploration, and emergency rescue. They can navigate uneven ground, climb obstacles, and maintain stability in changing conditions. These capabilities rely on continuous perception and real-time control, supported by efficient compute.&nbsp;</p>



<p>Similarly, platforms like the PUDU D5 extend autonomous mobility into industrial environments. Designed for inspection, patrol, and logistics support, the D5 Series operates across large sites and uneven terrain using LiDAR and camera-based vision. This is particularly valuable in environments that are hazardous, remote, or difficult for people to access, where robots can support operations while improving safety and reducing risk. &nbsp;</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="PUDU D5 Series on Arm" width="500" height="281" src="https://www.youtube.com/embed/zRNv6t38BHc?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div></figure>



<p>To support this, the system uses a heterogeneous compute architecture that distributes workloads across perception, planning, and control. Sensor data is processed continuously, allowing the robot to interpret its environment and respond with low latency.&nbsp;</p>



<p>Processors built on the Arm compute platform support these core functions, working alongside AI accelerators to deliver efficient performance at the edge. This enables robots to operate independently in environments where reliability, safety and energy efficiency are critical.&nbsp;</p>



<p>The same shift is also visible in industrial automation. Collaborative robots on factory floors are becoming more responsive to changing workflows, working alongside people and adapting to new tasks without requiring fully fixed configurations.&nbsp;</p>



<figure class="wp-block-embed aligncenter is-type-rich is-provider-soundcloud wp-block-embed-soundcloud"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="Arm Viewpoints: How AI and Advanced Robotics Are Redefining Automation by The Arm Podcast" width="500" height="400" scrolling="no" frameborder="no" src="https://w.soundcloud.com/player/?visual=true&#038;url=https%3A%2F%2Fapi.soundcloud.com%2Ftracks%2F2193957523&#038;show_artwork=true&#038;maxheight=750&#038;maxwidth=500"></iframe>
</div></figure>



<p><a href="https://bostondynamics.com/blog/doing-more-with-spot/" target="_blank" rel="noreferrer noopener">Boston Dynamics’ Spot</a> is another example of how mobile robots are being deployed in industrial settings for inspection and remote operation, where real-time perception and control are essential.&nbsp;</p>



<h3 class="wp-block-heading" id="h-autonomous-vehicles-and-mobility-platforms-nbsp">Autonomous vehicles and mobility platforms&nbsp;</h3>



<p>Physical AI is also transforming autonomous mobility, where systems must operate and navigate safely in complex, real-world conditions.&nbsp;</p>



<p>Autonomous robotaxis, such as those developed through the collaboration between <a href="https://newsroom.arm.com/blog/how-lenovo-scales-level-4-autonomous-robotaxis-on-arm" target="_blank" rel="noreferrer noopener">Lenovo and WeRide</a>, demonstrate how scalable compute platforms built on Arm are enabling Level 4 autonomy. These systems process large volumes of sensor data, including cameras and LiDAR, to make real-time driving decisions.&nbsp;</p>



<p>At the same time, the <a href="https://newsroom.arm.com/news/arm-and-tensor-partnership" target="_blank" rel="noreferrer noopener">Arm and Tensor</a> partnership highlights how next-generation compute platforms are being designed to support AI-driven mobility. These combine high-performance compute with energy efficiency, enabling real-time perception, planning, and control in autonomous systems.&nbsp;</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="How Tensor and Arm are powering the world’s first AI-agentic personal Robocar" width="500" height="281" src="https://www.youtube.com/embed/9ZJz8y8RwPk?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div></figure>



<p><a href="https://newsroom.arm.com/news/arm-rivian-autonomy-platform" target="_blank" rel="noreferrer noopener">Arm’s work with Rivian</a> also shows how custom autonomy platforms are enabling vehicles to interpret their environment and make real-time driving decisions at scale.&nbsp;In these environments, reliability and latency are critical. Decisions must be made instantly, and systems must operate consistently over long periods. Efficient, scalable compute plays a central role in making this possible.&nbsp;</p>



<h2 class="wp-block-heading" id="h-the-building-blocks-of-intelligent-systems-nbsp">The building blocks of intelligent systems&nbsp;</h2>



<p>At the core of physical AI is a continuous loop between sensing, decision-making, and action. Systems must process inputs from sensors, interpret that data, and trigger responses within milliseconds. In many cases, this latency between perception and action becomes a defining requirement, particularly in environments where safety and timing are critical.&nbsp;</p>



<p>The evolution of physical AI is rooted in how modern robotic and autonomous machines systems are being engineered. They bring together several core capabilities that operate as a coordinated system. Each layer contributes to how the machine understands and interacts with the world.&nbsp;</p>



<p>Perception provides environmental awareness. Cameras, LiDAR, and sensor arrays generate a continuous stream of data, allowing the system to detect objects, estimate distance, and map its surroundings.&nbsp;&nbsp;Edge AI inference processes this data locally. By running AI models on-device, robots can respond instantly without waiting for cloud input. This is critical in environments where latency affects safety or performance.&nbsp;&nbsp;</p>



<p>Multimodal reasoning combines inputs such as vision and language. Robots and other autonomous machines can interpret a scene, understand a command, and decide on the appropriate action. This brings interaction closer to how humans communicate and operate.&nbsp;&nbsp;Real-time control, safety and security ensure that decisions are executed reliably. Deterministic compute allows systems to respond within predictable time frames, while safety and security mechanisms help manage risk in complex environments. &nbsp;</p>



<p>These capabilities are already being applied across industries. <a href="https://newsroom.arm.com/blog/predictive-maintenance-smart-factories" target="_blank" rel="noreferrer noopener">In smart factories for instance</a>, predictive maintenance systems analyze equipment data continuously to detect early signs of failure. Meanwhile, in physical AI deployments, systems are designed to process real-world inputs and act with minimal delay.&nbsp;&nbsp;</p>



<h2 class="wp-block-heading" id="h-enabling-physical-ai-with-efficient-scalable-compute-nbsp">Enabling physical AI with efficient, scalable compute&nbsp;</h2>



<p>Today’s physical AI systems must process high volumes of sensor data, run AI models, and control movement, all within tight power and thermal limits. Many systems operate on batteries with power limits, which places a strong emphasis on efficiency. At the same time, performance must remain consistent, especially in environments where timing and accuracy are critical.&nbsp;</p>



<p>To support this, today’s compute platforms must deliver:&nbsp;</p>



<ul class="wp-block-list">
<li>Deterministic performance for real-time decision-making;&nbsp;&nbsp;</li>



<li>Energy efficiency for sustained operation in constrained environments;&nbsp;&nbsp;</li>



<li>On-device AI inference to reduce latency;&nbsp;&nbsp;</li>



<li>Heterogeneous compute to manage diverse workloads; and&nbsp;&nbsp;</li>



<li>The ability to scale across different types of physical AI systems.&nbsp;&nbsp;</li>
</ul>



<p>The Arm compute platform is designed around these principles. Because of this, it’s used across the full spectrum of compute within physical AI systems, from low power microcontrollers that process sensor data to high-performance central compute (essentially the “brain” of physical AI systems) that handle complex AI workloads. This allows engineers and developers to build systems where each component is optimized for its role, while still operating on a consistent architecture.&nbsp;</p>



<p>For example, in robotics, this approach enables a balance between performance and efficiency. A robot can process sensor inputs, run AI inference efficiently, and execute control logic without exceeding power or thermal limits. This is essential for systems that need to operate continuously in real-world conditions.&nbsp;</p>



<p>Arm’s ecosystem also plays a role in accelerating development. By working with partners across hardware and software, and supported by a global ecosystem of over 22 million developers, Arm enables a wide range of physical AI platforms, from machines on the factory floor in fixed environments to humanoids and quadrupeds operatingin real world environments. &nbsp;</p>



<h2 class="wp-block-heading" id="h-shaping-the-future-of-intelligent-machines-nbsp">Shaping the future of intelligent machines&nbsp;</h2>



<p>Physical AI is shaping how machines interact with the world. Robots and other autonomous machines are being designed with a deeper understanding of their environment, supported by advanced AI that allows them to interpret, decide, and act. This is expanding their role across industries, from interactive systems to industrial automation and autonomous mobility.&nbsp;</p>



<p>As these systems continue to evolve, the balance between performance and efficiency will determine how widely they can be deployed. Arm has supported the development of physical AI systems for years, starting with fixed machines on the factory floor. That same foundation is now enabling the next generation of intelligent robots and autonomous machines, where AI operates directly in the physical world and systems are built to respond in real time. &nbsp;</p>



<p>As physical AI continues to scale, it will increasingly be built on Arm.&nbsp;</p>


<div class="c-row u-justify-content-center CTA_box">
  <div class="c-col c-col-12 u-margin-top-2 u-margin-bottom-2">
    <ads-card has-open-border="true" class="u-text-center ab-cta-card">
      <ads-card-content slot="content">
        <h4 class="c-heading-3 u-margin-top-1 u-margin-bottom-1/2 u-text-bold">Powering intelligent physical AI on Arm</h4>
        <p>Arm enables safe, real-time, energy-efficient AI systems that sense, decide, and act reliably in the physical world.</p>
                <div class="c-col u-flex u-align-items-center u-justify-content-center u-padding-top-2 u-padding-bottom-1">
          <a class="c-cta-button is-primary" href="https://www.arm.com/markets/physical-ai" target="_blank"><span>Learn how</span></a>
        </div>
              </ads-card-content>
    </ads-card>
  </div>
</div>

<p>The post <a href="https://newsroom.arm.com/blog/the-evolution-of-physical-ai-from-controlled-environments-to-the-real-world">The evolution of physical AI: From controlled environments to the real world  </a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Gemma 4 on Arm: Accessible, immediate, optimized on-device AI to accelerate the mobile app experience</title>
		<link>https://newsroom.arm.com/blog/gemma-4-on-arm-optimized-on-device-ai</link>
		
		<dc:creator><![CDATA[Alex Spinelli]]></dc:creator>
		<pubDate>Thu, 02 Apr 2026 16:26:45 +0000</pubDate>
				<guid isPermaLink="false">https://newsroom.arm.com/?p=20148</guid>

					<description><![CDATA[<p>Gemma 4 on Arm brings fast, privacy-preserving, power-efficient AI directly onto Android devices, helping developers deliver richer real-time app experiences to billions of users without relying on the cloud.</p>
<p>The post <a href="https://newsroom.arm.com/blog/gemma-4-on-arm-optimized-on-device-ai">Gemma 4 on Arm: Accessible, immediate, optimized on-device AI to accelerate the mobile app experience</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Real-time assistance, seamless communication, and greater personalization are now baseline expectations for billions of smartphone users worldwide. Highly capable on-device AI that operates in the power envelope of modern smartphones is essential to delivering instant, intelligent experiences at scale, while unlocking AI’s future potential. &nbsp;</p>



<p>Google’s launch of Gemma 4 accelerates the ongoing shift to on-device AI, enabling developers to seamlessly access optimized performance and bring increasingly capable AI experiences directly into the apps people use every day. Unlocking these benefits at a global smartphone scale depends on the&nbsp;underlying compute foundation, with one constant that is&nbsp;ubiquitous across the entire Android ecosystem: Arm.&nbsp;</p>



<h2 class="wp-block-heading" id="h-what-s-new-for-gemma-4">What’s new for Gemma 4 </h2>



<p>Gemma 4 further advances on-device AI by delivering improved performance and efficiency, while expanding support for the kinds of&nbsp;multimodal&nbsp;experiences&nbsp;that matter most on Arm-based devices, including reasoning, agentic workflows, and vision-and-audio enabled use cases.&nbsp;With enhanced capabilities across text, audio*, and image,&nbsp;broader language support,&nbsp;and a foundation for real-time assistive experiences, it enables more responsive, context-aware interactions directly on-device without increasing memory footprint.&nbsp;</p>



<h2 class="wp-block-heading" id="h-exploring-gemma-4-performance-on-arm-nbsp-cpus">Exploring Gemma 4 performance on Arm&nbsp;CPUs</h2>



<p>In&nbsp;early&nbsp;Arm&nbsp;engineering tests, SME2 shows promising performance gains for running Gemma&nbsp;4 workloads. Initial tests on the Gemma 4 E2B (Effective 2 Billion) model&nbsp;demonstrate&nbsp;an average of 5.5x speedup in prefill (processing user input) and up to&nbsp;1.6x faster decode&nbsp;(generating responses), highlighting the potential of Armv9 CPU innovations for on-device AI workloads. These engineering tests include upcoming patches to&nbsp;Google&nbsp;XNNPACK&nbsp;and&nbsp;Arm&nbsp;<a href="https://www.arm.com/markets/artificial-intelligence/software/kleidi" target="_blank" rel="noreferrer noopener">KleidiAI</a>.&nbsp;</p>



<p>As an early example of what is possible with these improvements, Envision, an accessibility-focused app for blind and low-vision users, evaluated an on-device approach for delivering more of its experience locally. Historically, Envision&#8217;s&nbsp;scene interpretation relied on cloud connectivity. In this prototype, Gemma 4 was evaluated running locally on Arm CPUs with SME2 capabilities, enabling users to capture a photo and receive a detailed scene description directly on-device without requiring a network connection or sending sensitive data off-device.&nbsp;</p>



<p>These&nbsp;explorations on Arm CPUs highlight the broader flexibility of the Arm compute platform and the potential for continued innovation across CPU and heterogeneous compute pathways.&nbsp;</p>



<p>The result is lower latency, stronger privacy, and more consistent user experiences regardless of connectivity conditions. This shift from cloud dependency to local inference is critical for mobile applications. It&nbsp;has the potential to&nbsp;reduce infrastructure costs for developers, improve reliability for users, and&nbsp;unlock new&nbsp;categories of real-time applications. &nbsp;</p>



<p><em>“Envision is excited to work with Arm and Google to bring powerful accessibility experiences directly onto smartphones. Running visual understanding models like Gemma 4 on-device on SME2-enabled Arm CPUs opens the door to reliable, low-latency scene description and visual Q&amp;A for blind and low-vision users. For our community, the ability to access these capabilities offline is incredibly meaningful because it ensures the technology works wherever they are, while also improving privacy by keeping more processing on the device itself.”</em> &nbsp;– <strong>Karthik Mahadevan, CEO, Envision</strong> &nbsp;</p>



<p>Envision is an early example of what’s possible when Gemma 4 meets the Arm compute platform at mobile scale. As more developers integrate Gemma 4, on-device AI will increasingly become the default architecture rather than the exception. &nbsp;</p>



<h2 class="wp-block-heading" id="h-why-arm-matters-for-on-device-ai-at-android-scale">Why Arm matters for on-device AI at Android scale</h2>



<p>The <a href="https://www.arm.com/architecture/cpu/a-profile/armv9" target="_blank" rel="noreferrer noopener">Armv9 architecture</a> is the most secure, pervasive and advanced ISA ever​​​​. <a href="https://www.arm.com/technologies/sme2" target="_blank" rel="noreferrer noopener">Arm Scalable Matrix Extension 2 (SME2)</a> – a set of advanced CPU instructions in the Armv9 architecture – is a key technology, as it accelerates matrix-heavy AI workloads within the power envelope of smartphones. Already built into <a href="https://newsroom.arm.com/blog/arm-c1-cpu-cluster-on-device-ai-performance">Arm C1 CPUs </a> that are integrated into <a href="https://newsroom.arm.com/blog/arm-lumex-platform-ai-powered-smartphones-and-apps" target="_blank" rel="noreferrer noopener">the latest Android smartphone devices</a>, SME2 unlocks higher sustained performance and improved efficiency.   </p>



<p>Through&nbsp;Arm&nbsp;KleidiAI&nbsp;– &nbsp;Arm’s&nbsp;software acceleration layer integrated into leading runtime libraries, like Google’s XNNPACK, and frameworks, like Google LiteRT and MediaPipe &nbsp;– the benefits of SME2 are readily accessible to mobile developers with no changes required to existing code, models or deployment pipelines. As a result, developers automatically access out-the-box performance optimizations simply by targeting Arm-based Android devices built on SME2. &nbsp;</p>



<p>​​​In practice, these software-level gains translate directly into better on-device experiences. Users benefit from faster responses, smoother sustained interactions, and more reliable on-device AI, all while maintaining battery life and thermal stability, even as models grow more capable. &nbsp;<br><br><em>“Delivering Gemma 4 efficiently across the Android ecosystem requires deep collaboration across hardware and software. Our work with Arm reflects a shared commitment to advancing on-device AI, combining the benefits of the Armv9 architecture and built-in acceleration technologies, like SME2, with the Android operating system to unlock greater performance and efficiency at scale. Together, we’re making it easier for developers to bring fast, responsive, and privacy-preserving AI experiences to our users, without needing to modify their existing applications.” </em> – <strong>Sandeep Patil, Engineering Director, Android</strong>​​&nbsp;</p>



<h2 class="wp-block-heading" id="h-arm-and-google-building-the-future-of-on-device-ai-together">Arm and Google: Building the future of on-device AI together </h2>



<p>As more applications move AI on-device, Arm and Google are committed to supporting developers with accessible performance optimizations and clear guidance that help Gemma 4 accelerate application experiences across all Arm-based mobile devices. &nbsp;</p>



<p>The future of mobile AI will not be defined solely by larger models, but by how efficiently, securely, and pervasively they run at scale across the Android ecosystem. Through&nbsp;this collaboration, the benefits of on-device AI will be felt by billions of Android smartphone users worldwide.  &nbsp;</p>



<p><sub>*only&nbsp;for E2B (Effective 2&nbsp;Billion) and E4B (Effective 4&nbsp;Billion)</sub>&nbsp;</p>


<p>The post <a href="https://newsroom.arm.com/blog/gemma-4-on-arm-optimized-on-device-ai">Gemma 4 on Arm: Accessible, immediate, optimized on-device AI to accelerate the mobile app experience</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Top 12 Arm-based innovations from March 2026</title>
		<link>https://newsroom.arm.com/blog/top-arm-based-innovations-from-march-2026</link>
		
		<dc:creator><![CDATA[Arm Editorial Team]]></dc:creator>
		<pubDate>Wed, 01 Apr 2026 15:06:43 +0000</pubDate>
				<guid isPermaLink="false">https://newsroom.arm.com/?p=20125</guid>

					<description><![CDATA[<p>A monthly roundup of how Arm‑powered innovation and the ecosystem are progressing across cloud, edge, and device environments.</p>
<p>The post <a href="https://newsroom.arm.com/blog/top-arm-based-innovations-from-march-2026">Top 12 Arm-based innovations from March 2026</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>From training AI models in the data center to running real-time intelligence on devices, the way compute is being built and used is changing fast. Across the Arm ecosystem, partners and developers are solving real problems: building more efficient AI infrastructure, modernizing embedded workflows, enabling smarter cloud analytics, and bringing advanced experiences to mobile, gaming, and robotics platforms.</p>



<p>This month’s roundup highlights how these innovations are coming to life in practice. From Arm’s first silicon for AI infrastructure to on-device generative AI, neural graphics, and scalable autonomous systems, each story offers a closer look at how Arm-based technologies are helping teams move faster, build more efficiently, and deploy intelligence where it matters most.</p>



<h2 class="wp-block-heading" id="h-introducing-the-arm-agi-cpu">Introducing the Arm AGI CPU </h2>



<p>Arm has taken a major step forward with the introduction of its first Arm-designed silicon, <a href="https://newsroom.arm.com/news/arm-agi-cpu-launch" type="link" id="https://newsroom.arm.com/news/arm-agi-cpu-launch">the Arm AGI CPU</a>, extending the Neoverse platform beyond IP and Compute Subsystems (CSS) to give customers greater choice, from building custom silicon to integrating full platform solutions or deploying Arm-designed processors.</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="A look inside the Arm AGI CPU: Core features and benefits" width="500" height="281" src="https://www.youtube.com/embed/sqJancqeVbk?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div></figure>



<p>Building on this milestone, Shivangi Agrawal, Product Manager,&nbsp;<a href="https://developer.arm.com/community/arm-community-blogs/b/servers-and-cloud-computing-blog/posts/introducing-the-arm-agi-cpu-1ou-dual-node-reference-server" type="link" id="https://developer.arm.com/community/arm-community-blogs/b/servers-and-cloud-computing-blog/posts/introducing-the-arm-agi-cpu-1ou-dual-node-reference-server">explores what this means in practice</a> with a ready-to-deploy 1U dual-node reference server designed for high-density AI and cloud workloads. The platform gives partners a production-representative environment to evaluate performance, optimize software stacks, and design for rack-level efficiency across power, cooling, and space. For cloud providers and system builders, it offers a clearer, faster path to deploying Arm-based AI infrastructure at scale.</p>



<h2 class="wp-block-heading" id="h-bringing-real-world-robotics-to-life-with-arm">Bringing real-world robotics to life with Arm</h2>



<p>Robots stepped into Arm’s Cambridge office as Shanghai-based robotics company AGIBOT, known for its humanoid and embodied AI systems, showcased its latest humanoids and quadruped systems in action, demonstrating everything from precise task execution to dynamic, human-like movement.</p>



<figure class="wp-block-embed aligncenter is-type-rich is-provider-twitter wp-block-embed-twitter"><div class="wp-block-embed__wrapper">
<blockquote class="twitter-tweet" data-width="500" data-dnt="true"><p lang="en" dir="ltr">When robots visit the office <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f916.png" alt="🤖" class="wp-smiley" style="height: 1em; max-height: 1em;" /><br><br>AGIBOT joined us in Cambridge with its humanoids and a quadruped showcasing a variety of tasks from painting and dancing to human interactions and dynamic movement.<br><br>Behind every action is the Arm compute platform, spanning sensor-level processing… <a href="https://t.co/XHE6YTIVfr">pic.twitter.com/XHE6YTIVfr</a></p>&mdash; Arm (@Arm) <a href="https://twitter.com/Arm/status/2035071187852673440?ref_src=twsrc%5Etfw">March 20, 2026</a></blockquote><script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
</div></figure>



<p>Behind each interaction is the Arm compute platform, enabling intelligence from sensor-level processing through to high-performance AI. It highlights how Arm-based systems are supporting real-world robotics, where efficiency, responsiveness, and scalable compute are critical to moving from demonstration to deployment.</p>



<h2 class="wp-block-heading" id="h-aligning-embedded-development-with-modern-llvm-workflows">Aligning embedded development with modern LLVM workflows</h2>



<p>Arm is taking another step toward a more unified and modern software ecosystem by beginning the transition of Arm Toolchain for Embedded from Picolibc to LLVM-libc. Paul Black, Director of Product Management, outlined <a href="https://developer.arm.com/community/arm-community-blogs/b/tools-software-ides-blog/posts/arm-toolchain-for-embedded-migration-from-picolibc-to-llvm-libc" type="link" id="https://developer.arm.com/community/arm-community-blogs/b/tools-software-ides-blog/posts/arm-toolchain-for-embedded-migration-from-picolibc-to-llvm-libc">how this shift is being introduced gradually</a>, starting as an optional overlay before becoming the default C library. The move is designed to better align embedded development with the broader LLVM ecosystem, improving integration across tools and enabling more consistent, portable workflows.</p>



<p>For developers, the transition offers advantages such as simplified licensing and closer integration with LLVM-based runtimes, while also requiring updates to elements like linker scripts, semihosting, and startup code. </p>



<h2 class="wp-block-heading" id="h-turning-smartphone-cameras-into-real-time-ai-creation-tools">Turning smartphone cameras into real-time AI creation tools</h2>



<p>Smartphone cameras are rapidly evolving from capture devices into real-time generative AI tools, enabling users to preview and create content directly on their devices.</p>



<figure class="wp-block-embed aligncenter is-type-rich is-provider-twitter wp-block-embed-twitter"><div class="wp-block-embed__wrapper">
<blockquote class="twitter-tweet" data-width="500" data-dnt="true"><p lang="en" dir="ltr">Your smartphone camera is about to become a generative AI studio. <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f4f8.png" alt="📸" class="wp-smiley" style="height: 1em; max-height: 1em;" /><br><br>We&#39;ve partnered with <a href="https://twitter.com/tecnomobile?ref_src=twsrc%5Etfw">@tecnomobile</a> to bring real-time, fully on-device AI-generated content previews to smartphones — running at 30fps with zero cloud reliance.<br><br>Built on Armv9 CPUs and accelerated by Arm… <a href="https://t.co/IJXKZXCfEE">pic.twitter.com/IJXKZXCfEE</a></p>&mdash; Arm (@Arm) <a href="https://twitter.com/Arm/status/2029502757191733340?ref_src=twsrc%5Etfw">March 5, 2026</a></blockquote><script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
</div></figure>



<p>Through its partnership with TECNO, Arm is helping bring this shift to life with fully on-device AI-generated content previews running at 30 frames per second, without relying on the cloud. Built on Armv9 CPUs and accelerated by Arm KleidiAI, the solution delivers faster AI processing while maintaining responsiveness and energy efficiency.</p>



<h2 class="wp-block-heading" id="h-how-arm-developer-labs-helps-bring-industry-challenges-into-the-classroom">How Arm Developer Labs helps bring industry challenges into the classroom</h2>



<p>Arm Developer Labs gives students and educators a practical way to work on real computing problems using professional tools, workflows, and guidance from Arm engineers, helping bridge the gap between classroom learning and industry needs. <a href="https://developer.arm.com/community/education-hub/b/arm-education/posts/arm-developer-labs-shaping-computing-curricula-with-industry-software-challenges">In this Arm Community blog</a>, Kieran Hejmadi, Software &amp; Academic Ecosystem Development Manager, explores how Arm Developer Labs is helping universities modernize computing curricula through hands-on, industry-driven software challenges built on Arm technologies. </p>



<p>In this example, students at Anglia Ruskin University Cambridge built a web-scraping and data-visualization project that tracked research trends over time, while also extending the work with AI-driven analysis. It shows why industry-aligned learning matters now, giving students more relevant technical experience and helping universities prepare graduates for fast-changing software and AI careers.</p>



<h2 class="wp-block-heading" id="h-advancing-the-future-of-gaming-on-arm-at-gdc">Advancing the future of gaming on Arm at GDC</h2>



<p>At GDC 2026, Arm brought together developers and partners to explore the next phase of gaming innovation, from neural graphics and AI-driven upscaling to real-world performance optimization.</p>



<figure class="wp-block-embed aligncenter is-type-rich is-provider-twitter wp-block-embed-twitter"><div class="wp-block-embed__wrapper">
<blockquote class="twitter-tweet" data-width="500" data-dnt="true"><p lang="en" dir="ltr">From neural graphics and upscaling to real-world performance tuning , one thing is clear: the future of gaming is built on Arm.<br> <br>We had such a great time exploring the future with our partners and all of you at the Arm Developer Summit here at <a href="https://twitter.com/hashtag/GDC2026?src=hash&amp;ref_src=twsrc%5Etfw">#GDC2026</a>! <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/1f680.png" alt="🚀" class="wp-smiley" style="height: 1em; max-height: 1em;" /> <a href="https://t.co/etB2RRUXmP">pic.twitter.com/etB2RRUXmP</a></p>&mdash; Arm Software Developers (@ArmSoftwareDev) <a href="https://twitter.com/ArmSoftwareDev/status/2031514639864049954?ref_src=twsrc%5Etfw">March 10, 2026</a></blockquote><script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
</div></figure>



<p><a href="https://newsroom.arm.com/blog/takeaways-from-gdc-festival-of-gaming-2026" type="link" id="https://newsroom.arm.com/blog/takeaways-from-gdc-festival-of-gaming-2026">Across sessions at the Arm Developer Summit</a>, a clear theme emerged: how to push visual quality further on mobile without exceeding power and thermal limits. Developers explored practical techniques such as Neural Frame Rate Upscaling and AI-assisted rendering, alongside tools that help identify bottlenecks and optimize performance in production workflows. Additionally, Annie Tallund, Solutions Engineer, explored <a href="https://developer.arm.com/community/arm-community-blogs/b/mobile-graphics-and-gaming-blog/posts/new-neural-technologies-set-to-join-the-neural-graphics-development-kit" type="link" id="https://developer.arm.com/community/arm-community-blogs/b/mobile-graphics-and-gaming-blog/posts/new-neural-technologies-set-to-join-the-neural-graphics-development-kit">how the Arm Neural Graphics Development Kit is expanding</a> with new neural technologies to help developers bring AI-enhanced graphics and rendering workflows to mobile platforms.</p>



<h2 class="wp-block-heading" id="h-the-next-generation-of-ai-data-centers-with-arm">The next generation of AI data centers with Arm</h2>



<p>As AI data centers take on more always-on workloads, the challenge is no longer just delivering peak performance. It is also about keeping complex systems running efficiently at scale. Learn why CPUs have become a critical control layer for orchestrating accelerators, managing workloads, and improving overall infrastructure efficiency. It highlights Arm’s work with Meta to help advance more power-efficient AI data center design across the ecosystem.</p>



<figure class="wp-block-embed aligncenter is-type-rich is-provider-twitter wp-block-embed-twitter"><div class="wp-block-embed__wrapper">
<blockquote class="twitter-tweet" data-width="500" data-dnt="true"><p lang="en" dir="ltr">In the AI data center, power sets the pace. <img src="https://s.w.org/images/core/emoji/17.0.2/72x72/26a1.png" alt="⚡" class="wp-smiley" style="height: 1em; max-height: 1em;" /><br><br>Always-on AI shifts the constraint from peak performance to intelligent orchestration. CPUs are the control layer that keeps accelerators productive and systems operating efficiently at scale.<br><br>Our work with <a href="https://twitter.com/Meta?ref_src=twsrc%5Etfw">@meta</a> reflects Arm’s… <a href="https://t.co/nkU2UtJc26">pic.twitter.com/nkU2UtJc26</a></p>&mdash; Arm (@Arm) <a href="https://twitter.com/Arm/status/2029673883880640615?ref_src=twsrc%5Etfw">March 5, 2026</a></blockquote><script async src="https://platform.twitter.com/widgets.js" charset="utf-8"></script>
</div></figure>



<h2 class="wp-block-heading" id="h-new-kiro-powers-help-streamline-agentic-ai-development-on-arm">New Kiro Powers help streamline agentic AI development on Arm</h2>



<p>Agentic AI is reshaping how developers build and move software across cloud and edge environments, especially when teams need faster workflows with more architectural guidance. Arm has introduced new Kiro Powers to accelerate agentic AI development and simplify cloud-to-edge workflows on Arm-based platforms, as explained by Zach Lasiuk, Principal Solutions Designer, <a href="https://developer.arm.com/community/arm-community-blogs/b/ai-blog/posts/from-cloud-to-edge-expanding-agentic-development-on-arm-with-new-kiro-powers">in this Arm Community blog</a>. </p>



<p>It is these new capabilities that are helping cloud teams plan migrations to AWS Graviton and help embedded developers manage transitions between Arm SoCs with more structure, visibility, and confidence. For developers working across modern AI, embedded, automotive, and edge systems, Zach also explains how Arm is helping make complex platform changes easier to manage.</p>



<h2 class="wp-block-heading" id="h-accelerating-real-time-event-analytics-in-the-cloud-with-arm">Accelerating real-time event analytics in the cloud with Arm</h2>



<p>Cloud-based event analytics is helping platforms turn live audience activity into immediate business decisions, from dynamic ticket pricing to targeted VIP experiences, while demand is still building.</p>



<p>Pascal Mudimba,  Arm Ambassador, <a href="https://developer.arm.com/community/arm-community-blogs/b/servers-and-cloud-computing-blog/posts/empowering-creators-how-stdio-x-labs-cut-event-analytics-latency-by-40-using-arm-accelerated-bigquery-ml-on-google-cloud" type="link" id="https://developer.arm.com/community/arm-community-blogs/b/servers-and-cloud-computing-blog/posts/empowering-creators-how-stdio-x-labs-cut-event-analytics-latency-by-40-using-arm-accelerated-bigquery-ml-on-google-cloud">shows how stdio x Labs reduced analytics latency by 40%</a> by combining Arm-based compute with optimized BigQuery ML pipelines on Google Cloud. By moving key stages such as stream processing and model preparation onto Arm, the team improved throughput, lowered infrastructure costs, and delivered faster insights for always-on data workloads.</p>



<p>The result is a practical example of how developers and digital platforms can use Arm in the cloud to make real-time analytics more responsive, scalable, and efficient, especially in environments where timing directly impacts revenue and user experience.</p>



<h2 class="wp-block-heading" id="h-customizing-ai-models-for-real-world-use-cases-on-arm">Customizing AI models for real-world use cases on Arm</h2>



<p>In this video, Michael Hall, Principal Software Engineer and Developer Evangelist, explores how to turn a general-purpose large language model into a domain-specific AI system that can deliver faster, more accurate responses.</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="Customize your AI with model fine tuning on Nvidia DGX Spark" width="500" height="281" src="https://www.youtube.com/embed/59pX_UBSzNw?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div></figure>



<p>Using an NVIDIA DGX Spark desktop powered by Arm CPUs and NVIDIA GPUs, the session walks through fine-tuning a Llama 3.2 model with a custom dataset of Raspberry Pi device specifications. It demonstrates how widely used tools like PyTorch and Hugging Face Transformers can be combined to load models, prepare data, train efficiently, and deploy tailored AI solutions.</p>



<h2 class="wp-block-heading" id="h-training-humanoid-robots-faster-with-nvidia-dgx-spark">Training humanoid robots faster with NVIDIA DGX Spark</h2>



<p>Reinforcement learning is becoming a key tool for robotics, helping humanoid systems learn how to move, balance, and respond to real-world conditions through simulation before deployment. </p>



<p><a href="https://developer.arm.com/community/arm-community-blogs/b/ai-blog/posts/rethinking-robotics-reinforcement-learning-a-practical-humanoid-training-workflow-on-dgx-spark">In this Arm Community blog</a>, Odin Shen, Principal Solutions Architect, offers a practical end-to-end workflow for training humanoid robots using NVIDIA DGX Spark, combining high-fidelity simulation, parallelized training, and scalable infrastructure to accelerate real-world robot policy development. Using Isaac Sim and Isaac Lab on a single Arm-based DGX Spark system, developers can build, train, and evaluate humanoid locomotion policies more efficiently and move faster toward reproducible physical AI workflows.</p>



<h2 class="wp-block-heading" id="h-a-stronger-foundation-for-vulkan-frame-analysis">A stronger foundation for Vulkan frame analysis</h2>



<p>Frame Advisor is part of Arm Performance Studio, a free toolset that helps graphics developers capture and analyze Vulkan frames so they can spot GPU bottlenecks faster and optimize with more confidence.</p>



<p><a href="https://developer.arm.com/community/arm-community-blogs/b/mobile-graphics-and-gaming-blog/posts/frame-advisor-a-major-upgrade-to-vulkan-capture">In this Arm Community blog</a>, Daniel Baines, Staff Software Engineer, explains how the latest Frame Advisor upgrade introduces a new Vulkan capture pipeline built on GFXReconstruct, improving reliability, correctness, and scalability for today’s increasingly complex graphics workloads. The result is a stronger foundation for developers working on performance-critical applications, especially when accurate frame analysis is essential to diagnosing issues and validating optimizations. Modern tooling can make Vulkan development more dependable today while opening the door to future capabilities.</p>


<p>The post <a href="https://newsroom.arm.com/blog/top-arm-based-innovations-from-march-2026">Top 12 Arm-based innovations from March 2026</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></content:encoded>
					
		
		
			</item>
		<item>
		<title>Announcing Arm AGI CPU: The silicon foundation for the agentic AI cloud era</title>
		<link>https://newsroom.arm.com/blog/introducing-arm-agi-cpu</link>
		
		<dc:creator><![CDATA[Mohamed Awad]]></dc:creator>
		<pubDate>Tue, 24 Mar 2026 16:55:00 +0000</pubDate>
				<guid isPermaLink="false">https://newsroom.arm.com/?p=19973</guid>

					<description><![CDATA[<p>Breakthrough rack-level performance, scale and efficiency for the next generation of AI infrastructure. </p>
<p>The post <a href="https://newsroom.arm.com/blog/introducing-arm-agi-cpu">Announcing Arm AGI CPU: The silicon foundation for the agentic AI cloud era</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></description>
										<content:encoded><![CDATA[
<p>Today, <a href="https://newsroom.arm.com/news/arm-agi-cpu-launch">Arm is announcing the&nbsp;Arm AGI CPU</a>, a new class of production-ready silicon built on the <a href="https://www.arm.com/products/silicon-ip-cpu/neoverse/">Arm Neoverse</a> platform and designed to power the next generation of AI infrastructure.</p>



<p>For the first time in our <a href="https://newsroom.arm.com/blog/arm-official-history">more than 35-year history</a>, Arm is delivering its own silicon products – extending the Arm Neoverse platform beyond IP and Arm Compute Subsystems (CSS) to give customers greater choice in how they deploy Arm compute – from building custom silicon to integrating platform-level solutions or deploying Arm-designed processors. It reflects both the rapid evolution of AI infrastructure and growing demand from the ecosystem for production-ready Arm platforms that can be deployed at pace and scale.</p>



<h2 class="wp-block-heading" id="h-the-rise-of-the-agentic-ai-infrastructure">The rise of the agentic AI infrastructure</h2>



<p>AI systems are increasingly operating continuously at global scale. Historically, the human was the bottleneck in computing – the pace at which people could interact with systems defined how quickly work could move through them. In the era of agentic AI, that constraint disappears as software agents coordinate tasks, interact with multiple models and make decisions in real time.</p>



<p>As AI systems run continuously and workloads grow in complexity, the <a href="https://newsroom.arm.com/blog/ai-datacenter-cpu-orchestration-arm">CPU becomes the pacing element of modern infrastructure</a> – responsible for keeping distributed AI systems operating efficiently at scale. In a modern-day AI data center, the CPU manages thousands of distributed tasks – orchestrating accelerators, managing memory and storage, scheduling workloads and moving data across systems – and now, with agentic AI, coordinating fan-out across large numbers of agents.</p>



<p>This shift places new demands on the CPU and that requires an evolution of the processor.</p>



<p>Arm Neoverse already underpins many of today’s leading hyperscale and AI platforms, including <a href="https://www.arm.com/markets/computing-infrastructure/cloud-computing/aws">AWS Graviton</a>, <a href="https://www.arm.com/markets/computing-infrastructure/cloud-computing/google-cloud">Google Axion</a>, <a href="https://www.arm.com/markets/computing-infrastructure/cloud-computing/microsoft">Microsoft Azure Cobalt</a> and <a href="https://newsroom.arm.com/blog/arm-rubin-converged-ai-datacenter">NVIDIA Vera</a>. As AI infrastructure scales globally, partners across the ecosystem are asking Arm to do more. The Arm AGI CPU was created to address this shift.</p>



<h2 class="wp-block-heading" id="h-arm-agi-cpu-built-for-rack-scale-agentic-efficiency">Arm AGI CPU: Built for rack-scale agentic efficiency</h2>



<p>Agentic AI workloads demand sustained performance at massive scale. The <a href="https://www.arm.com/products/cloud-datacenter/arm-agi-cpu">Arm AGI CPU</a> is designed to deliver high per-task performance at sustained load across thousands of cores in parallel – all within the power and cooling limits of modern data centers.</p>



<p>Every element of the Arm AGI CPU – from operating frequency to memory and I/O architecture – has been designed to support massively parallel, high-performance agentic workloads in a densely populated rack deployment.</p>



<figure class="wp-block-image size-large"><img decoding="async" width="1200" height="675" src="https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-1200x675.png" alt="" class="wp-image-20062" srcset="https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-1200x675.png 1200w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-640x360.png 640w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-768x432.png 768w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-1536x864.png 1536w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-576x324.png 576w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-1024x576.png 1024w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-1280x720.png 1280w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6-1400x788.png 1400w, https://newsroom.arm.com/wp-content/uploads/2026/03/image-6.png 1600w" sizes="(max-width: 1200px) 100vw, 1200px" /></figure>



<p>Arm’s reference server configuration is a 1OU, 2-node design – packing in two chips with dedicated memory and I/O for a total of 272 cores per blade. These blades are designed to fully populate a standard air-cooled 36kW rack – 30 blades delivering a total of 8160 cores. Arm has additionally partnered with Supermicro on a liquid-cooled 200kW design capable of housing 336 Arm AGI CPUs for over 45,000 cores.</p>



<p>In this configuration, the Arm AGI CPU is capable of delivering more than 2x the performance per rack compared to the latest x86 systems*, achieved through the fundamental advantages of the Arm architecture and careful matching of system resources to compute:</p>



<ul class="wp-block-list">
<li>Arm AGI CPU’s class-leading memory bandwidth means more effective threads of execution per rack;  x86 CPUs degrade as cores contend under sustained load.</li>



<li>High performance, efficient, single-threaded <a href="https://www.arm.com/products/silicon-ip-cpu/neoverse/neoverse-v3">Arm Neoverse V3 CPU</a> cores outperform legacy architectures; every Arm thread does more work.</li>



<li>More usable threads and more work-per-thread compounds to massive performance gains per rack.</li>
</ul>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="A look inside the Arm AGI CPU: Core features and benefits" width="500" height="281" src="https://www.youtube.com/embed/sqJancqeVbk?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div></figure>



<h2 class="wp-block-heading" id="h-early-momentum-across-the-ai-ecosystem">Early momentum across the AI ecosystem</h2>



<p>The Arm AGI CPU is already seeing strong commercial momentum with partners at the forefront of scaling agentic AI infrastructure. Planned deployments span accelerator management, agentic orchestration and the densification of services, applications and tools needed for agentic task scale-out — as well as increased networking and data plane compute to support the AI data center.</p>



<p><a href="https://about.fb.com/news/2026/03/meta-partners-with-arm-to-develop-new-class-of-data-center-silicon">Meta&nbsp;is our lead partner and customer</a>, co-developing the Arm AGI CPU to optimize gigawatt-scale infrastructure for its Meta family of apps and to work alongside Meta’s own custom MTIA accelerators. Other launch partners include&nbsp;Cerebras, Cloudflare, F5, OpenAI, Positron, Rebellions, SAP, and SK Telecom – each working with Arm on the deployment of the Arm AGI CPU to accelerate AI-driven services across cloud, networking and enterprise environments. Commercial systems are now available for order from ASRockRack, Lenovo and Supermicro.</p>



<figure class="wp-block-embed is-type-video is-provider-youtube wp-block-embed-youtube wp-embed-aspect-16-9 wp-has-aspect-ratio"><div class="wp-block-embed__wrapper">
<iframe loading="lazy" title="Arm AGI CPU: Reactions from Arm&#039;s deployment partners" width="500" height="281" src="https://www.youtube.com/embed/uYmzRoxe1nY?feature=oembed" frameborder="0" allow="accelerometer; autoplay; clipboard-write; encrypted-media; gyroscope; picture-in-picture; web-share" referrerpolicy="strict-origin-when-cross-origin" allowfullscreen></iframe>
</div></figure>



<p>To accelerate adoption further, Arm is introducing&nbsp;the <a href="https://developer.arm.com/community/arm-community-blogs/b/servers-and-cloud-computing-blog/posts/introducing-the-arm-agi-cpu-1ou-dual-node-reference-server">Arm AGI CPU 1OU Dual Node Reference Server</a>, an Open Compute Project (OCP) DC-MHS standard form factor server. Arm plans to contribute this reference server design and supporting firmware, along with further contributions including system architecture specifications, debug frameworks and diagnostic and verification tooling applicable to all Arm-based systems. Further details will come at the upcoming <a href="https://www.opencompute.org/summit/emea-summit">OCP EMEA Summit</a>.</p>



<h2 class="wp-block-heading" id="h-a-new-chapter-for-arm-infrastructure">A new chapter for Arm infrastructure</h2>



<p>The launch of Arm AGI CPU represents a new chapter in Arm’s data center journey and continued leadership in computing innovation. As AI reshapes the industry, Arm remains committed to enabling progress across the ecosystem – meeting customers where they are, from hyperscale cloud providers to AI startups.</p>



<p>The Arm AGI CPU is the first offering of Arm’s new data center silicon product line and is available to order now. Follow-on products are committed, targeting best-in-class performance, scale and efficiency. This continues in parallel with the <a href="https://www.arm.com/products/neoverse-compute-subsystems">Arm Neoverse CSS</a> product roadmap so that all Arm data center customers move forward together on platform architecture and software compatibility.</p>



<p>Entering this new chapter, our mission remains unchanged: to provide the compute foundation that enables innovation across industries. And the ecosystem is fully behind us: <a href="https://www.arm.com/products/cloud-datacenter/arm-agi-cpu/ecosystem">More than 50 leading companies</a> across hyperscale, cloud, silicon, memory, networking, software, system design and manufacturing are supporting the expansion of the Arm compute platform into silicon. With Arm AGI CPU, we are not only defining the architecture of the AI-native data center, we are building it.</p>



<p>Hear more from our Arm AGI CPU deployment partners:</p>



<h3 class="wp-block-heading" id="h-cerebras">Cerebras</h3>



<p>“At Cerebras we build AI infrastructure designed for ultra-fast, large-scale inference, and as this becomes the dominant workload in AI, composable, high-performance systems matter more than ever – these systems need purpose-built AI acceleration alongside efficient, scalable CPUs orchestrating data movement, networking, and coordination at scale. Extending the Arm compute platform into AGI-class infrastructure is a positive step for the ecosystem and for customers deploying AI at global scale.” &#8211; <strong>Andrew Feldman, CEO, Cerebras</strong></p>



<h3 class="wp-block-heading" id="h-cloudflare">Cloudflare</h3>



<p>&#8220;To continue our mission of helping build a better Internet, Cloudflare needs infrastructure that scales efficiently across our global network. The Arm AGI CPU provides high-performance, energy-efficient compute designed for the next generation of workloads.&#8221; &#8211; <strong>Stephanie Cohen, Chief Strategy Officer, Cloudflare</strong></p>



<h3 class="wp-block-heading" id="h-meta">Meta</h3>



<p>“Delivering AI experiences at global scale demands a robust and adaptable portfolio of custom silicon solutions, purpose-built to accelerate AI workloads and optimize performance across Meta’s platforms. We worked alongside Arm to develop the Arm AGI CPU to deploy an efficient compute platform that significantly improves our data center performance density and supports a multi-generation roadmap for our evolving AI systems.” &#8211; <strong>Santosh Janardhan, Head of Infrastructure, Meta</strong></p>



<h3 class="wp-block-heading" id="h-openai">OpenAI</h3>



<p>“OpenAI runs AI systems at massive scale. Hundreds of millions use ChatGPT every day, businesses build on our API, and developers rely on tools like Codex. The Arm AGI CPU will play an important role in our infrastructure as we scale, strengthening the orchestration layer that coordinates large scale AI workloads and improving efficiency, performance, and bandwidth across the system.” &#8211; <strong>Sachin Katti, Head of Industrial Compute at OpenAI</strong></p>



<h3 class="wp-block-heading" id="h-positron">Positron</h3>



<p>“At Positron, we are focused on purpose-built inference accelerators that delivers breakthrough token generation efficiency using commodity memory. Arm has consistently delivered the industry’s most power-efficient compute platforms, which makes the Arm AGI CPU a natural foundation for next-generation AI infrastructure. By combining Positron’s inference acceleration technology with the energy-efficient Arm AGI CPU platform, we see a powerful opportunity to help data center operators deploy frontier AI models at scale with greater performance per watt and per dollar.” &#8211; <strong>Mitesh Agrawal, CEO, Positron AI</strong></p>



<h3 class="wp-block-heading" id="h-rebellions">Rebellions</h3>



<p>“High-performance AI systems require tight coordination between general-purpose compute and accelerator architectures. By combining the Arm AGI CPU with Rebellions’ NPUs in new high-density server configurations — we’re delivering a scalable, energy efficient platform that is optimized for AI inference workloads at scale.” &#8211; <strong>Marshall Choy, Chief Business Officer, Rebellions</strong></p>



<h3 class="wp-block-heading" id="h-sap">SAP</h3>



<p>“SAP’s successful deployment of SAP HANA on Arm-based AWS Graviton underscores the maturity and performance of the Arm ecosystem for enterprise workloads. The Arm AGI CPU extends that opportunity, providing scalable, efficient compute designed to support the next generation of AI-powered business solutions.” &#8211; <strong>Stefan Bäuerle, Senior Vice President, Head of HANA &amp; Persistency, SAP</strong></p>



<h3 class="wp-block-heading" id="h-sk-telecom">SK Telecom</h3>



<p>“SK Telecom is expanding into large-scale, full-stack AI inference data center infrastructure, which includes Arm AGI CPU and Rebellions AI accelerator chip. By bringing together our sovereign A.X foundation model with inference-optimized AI servers, we are ready to deliver it to world while elevating our AIDC competitiveness.&#8221; &#8211; <strong>Suk-geun (SG) Chung, CTO and Head of AI CIC, SK Telecom</strong></p>



<h2 class="wp-block-heading" id="h-forward-looking-statements">Forward-looking statements</h2>



<p>This blog post contains forward-looking statements regarding Arm&#8217;s product roadmap, future performance, planned contributions and partner deployments. These statements are based on current expectations and are subject to risks and uncertainties that could cause actual results to differ materially. For a discussion of factors that could affect Arm&#8217;s results, please refer to Arm&#8217;s filings with the U.S. Securities and Exchange Commission.</p>



<p>Performance claims are based on Arm internal estimates comparing a fully populated rack of Arm AGI CPU-based servers against comparable x86-based server configurations using industry-standard workloads. Actual results may vary based on system configuration, workload, and other factors.</p>



<p>All product and company names are trademarks or registered trademarks of their respective holders.</p>



<p><strong><em>*</em></strong><em>Based on estimates</em></p>



<p><a id="_msocom_1"></a></p>


<p>The post <a href="https://newsroom.arm.com/blog/introducing-arm-agi-cpu">Announcing Arm AGI CPU: The silicon foundation for the agentic AI cloud era</a> appeared first on <a href="https://newsroom.arm.com">Arm Newsroom</a>.</p>
]]></content:encoded>
					
		
		
			</item>
	</channel>
</rss>
