Nvidia Unveils Vera CPU Built for Agentic AI at GTC 2026

AI News Hub Editorial
Senior AI Reporter
March 16, 2026
Nvidia Unveils Vera CPU Built for Agentic AI at GTC 2026

Nvidia unveiled its new Vera CPU during CEO Jensen Huang’s keynote at the company’s GTC event in San Jose, positioning the processor as a chip designed specifically for the growing demands of agentic AI and reinforcement learning workloads. The company said Vera has entered full production and will ship through partners in the second half of the year as part of Nvidia’s broader Vera Rubin platform.

The processor is built to support AI systems that increasingly perform complex tasks such as planning workflows, executing tools, interacting with data and validating results. Nvidia says the chip delivers twice the efficiency and runs up to 50% faster than traditional rack-scale CPUs, marking a shift toward infrastructure designed specifically for AI services.

“Vera is arriving at a turning point for AI. As intelligence becomes agentic — capable of reasoning and acting — the importance of the systems orchestrating that work is elevated,” Huang said in a statement. “The CPU is no longer simply supporting the model; it’s driving it. With breakthrough performance and energy efficiency, Vera unlocks AI systems that think faster and scale further.”

The new chip expands Nvidia’s CPU lineup following the earlier Grace processor and is intended to support what the company calls “AI factories” — large-scale computing environments built to run AI services and agents. Nvidia says Vera delivers higher throughput and responsiveness for workloads such as coding assistants, enterprise agents and consumer AI tools.

Several hyperscale cloud providers and infrastructure companies are already collaborating with Nvidia on deployments. Partners include Alibaba, CoreWeave, Meta and Oracle Cloud Infrastructure, alongside hardware manufacturers such as Dell Technologies, HPE, Lenovo and Supermicro.

The processor is part of the Vera Rubin NVL72 platform, which pairs Vera CPUs with Nvidia GPUs using the company’s NVLink-C2C interconnect. Nvidia says the connection enables 1.8 TB/s of coherent bandwidth, allowing faster data sharing between CPUs and GPUs for AI workloads.

Nvidia also introduced a rack system built around the new processor. A single rack integrates 256 liquid-cooled Vera CPUs and can support more than 22,500 concurrent CPU environments running at full performance. The design uses Nvidia’s MGX modular architecture and is supported by dozens of ecosystem partners.

The chip includes 88 custom Nvidia-designed Olympus cores built to handle AI infrastructure tasks such as orchestration, analytics pipelines and runtime systems. Each core can run two tasks simultaneously through Nvidia Spatial Multithreading to support multi-tenant environments where many AI jobs run at once.

To support demanding workloads, the processor also features a new memory system based on LPDDR5X, delivering up to 1.2 TB/s of bandwidth. Nvidia says the design provides twice the bandwidth while using half the power compared with general-purpose CPUs.

Early partners have begun testing the platform. Alex Gallego, CEO of streaming data company Redpanda, said the company saw improved performance when benchmarking workloads on Vera. “Redpanda recently tested Nvidia Vera running Apache Kafka-compatible workloads and saw dramatically better performance than other systems we’ve benchmarked, delivering up to 5.5 times lower latency,” he said in a statement.

Research institutions are also preparing to deploy the new processor. Planned users include the Leibniz Supercomputing Centre, Los Alamos National Laboratory, Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center and the Texas Advanced Computing Center.

“At TACC, we recently tested Nvidia’s Vera CPU platform as we prepare for deployment in our upcoming Horizon system — and running six of our scientific applications, we saw impressive early results,” said John Cazes, director of high-performance computing at TACC. “Vera’s per-core performance and memory bandwidth represent a giant step forward for scientific computing, and we look forward to bringing Vera-based nodes to our CPU users on Horizon later this year.”

Nvidia said Vera systems will be available from partners later this year as the company expands its hardware platform designed to support large-scale AI deployments.

This analysis is based on reporting from GamesBeat.

Image courtesy of Nvidia.

This article was generated with AI assistance and reviewed for accuracy and quality.

Last updated: March 16, 2026

About this article: This article was generated with AI assistance and reviewed by our editorial team to ensure it follows our editorial standards for accuracy and independence. We maintain strict fact-checking protocols and cite all sources.

Word count: 659Reading time: 0 minutes

AI Tools for this Article

Trending Now

📧 Stay Updated

Get the latest AI news delivered to your inbox every morning.

Browse All Articles
Share this article:
Next Article