Nvidia GTC 2026: Nvidia Launches Vera CPU, Purpose-Built for Agentic AI
Nvidia Vera CPU delivers the highest performance and energy efficiency for data processing, AI training and agentic inference at scale
This is a Press Release edited by StorageNewsletter.com on March 19, 2026 at 2:02 pmSummary:
- The Nvidia Vera CPU delivers results with twice the efficiency and 50% faster than traditional CPUs
- Customers collaborating with Nvidia to deploy Vera CPU include Alibaba Cloud, ByteDance, Meta and Oracle Cloud Infrastructure, along with CoreWeave, Lambda, Nebius and Nscale
- Manufacturing partners already adopting the Vera CPU include Dell Technologies, HPE, Lenovo and Supermicro, along with ASUS, Compal, Foxconn, GigaByte, Pegatron, Quanta Cloud Technology (QCT), Wistron and Wiwynn
Nvidia Corp. launched the Nvidia Vera CPU, the world’s first processor purpose-built for the age of agentic AI and reinforcement learning – delivering results with twice the efficiency and 50% faster than traditional rack-scale CPUs.
As reasoning and agentic AI advances, scale, performance and cost are increasingly driven by the infrastructure supporting the models that plan tasks, run tools, interact with data, run code and validate results.
The Nvidia Vera CPU builds on the success of the Nvidia Grace CPU, enabling organizations of all sizes and across industries to build AI factories that unlock agentic AI at scale. With the highest single-thread performance and bandwidth per core, Vera is a new class of CPU that delivers higher AI throughput, responsiveness and efficiency for large-scale AI services such as coding assistants, as well as consumer and enterprise agents.
Leading hyperscalers collaborating with Nvidia to deploy Vera include Alibaba Cloud, CoreWeave, Meta and Oracle Cloud Infrastructure, as well as global system makers Dell Technologies, HPE, Lenovo, Supermicro and others. This broad adoption establishes Vera as the new CPU standard for the AI workloads that matter most for developers, startups, public-private institutions and enterprises – helping democratize access to AI and accelerating innovation.
“Vera is arriving at a turning point for AI. As intelligence becomes agentic – capable of reasoning and acting – the importance of the systems orchestrating that work is elevated,” said Jensen Huang, founder and CEO, Nvidia. “The CPU is no longer simply supporting the model; it’s driving it. With breakthrough performance and energy efficiency, Vera unlocks AI systems that think faster and scale further.”
Configurable for Every Data Center
Nvidia announced a new Vera CPU rack integrating 256 liquid-cooled Vera CPUs to sustain more than 22,500 concurrent CPU environments, each running independently at full performance. AI factories can quickly deploy and scale to tens of thousands of simultaneous instances and agentic tools in a single rack.
The new Vera rack is built using the Nvidia MGX modular reference architecture, supported by 80 ecosystem partners worldwide.
As part of the Nvidia Vera Rubin NVL72 platform, Vera CPUs are paired with Nvidia GPUs through Nvidia NVLink-C2C interconnect technology, with 1.8 TB/s of coherent bandwidth – 7x the bandwidth of PCIe Gen 6 – for high-speed data sharing between CPUs and GPUs. Additionally, Nvidia introduced new reference designs that use Vera as the host CPU for Nvidia HGX Rubin NVL8 systems, coordinating data movement and system control for GPU-accelerated workloads.
Vera systems partners are providing both dual and single-socket CPU server configurations, optimal for workloads such as reinforcement learning, agentic inference, data processing, orchestration, storage management, cloud applications and high-performance computing.
Across all configurations, Vera systems integrate Nvidia ConnectX SuperNIC cards and Nvidia BlueField-4 DPUs for accelerated networking, storage and security, which are critical for agentic AI. This enables customers to optimize for their specific workloads while maintaining a single software stack across the Nvidia platform.
Designed for Agentic Scaling
By combining high-performance, energy-efficient CPU cores, a high-bandwidth memory subsystem and the second-generation Nvidia Scalable Coherency Fabric, Vera enables faster agentic responses under the extreme utilization conditions common for agentic AI and reinforcement learning.
Vera features 88 custom Nvidia-designed Olympus cores, delivering high performance for compilers, runtime engines, analytics pipelines, agentic tooling and orchestration services. Each core can run two tasks, using Nvidia Spatial Multithreading, to deliver consistent, predictable performance – ideal for multi-tenant AI factories running many jobs at once.
To further enhance energy efficiency, Vera introduces the second generation of Nvidia’s low-power memory subsystem, now built on LPDDR5X memory and delivering up to 1.2 TB/s of bandwidth – twice the bandwidth and at half the power compared with general-purpose CPUs.
Widespread Ecosystem Support
Cursor, an innovator in AI-native software development, is adopting Nvidia Vera to boost performance for its AI coding agents.
“We’re excited to use Nvidia Vera CPUs to improve overall throughput and efficiency so we can deliver faster, more responsive coding agent experiences for our customers,” said Michael Truell, cofounder and CEO, Cursor.
Redpanda, a leading streaming data and AI platform, is using Vera to dramatically boost performance.
“Redpanda recently tested Nvidia Vera running Apache Kafka-compatible workloads and saw dramatically better performance than other systems we’ve benchmarked, delivering up to 5.5x lower latency,” said Alex Gallego, founder and CEO, Redpanda. “Vera represents a new direction in CPU architecture, with more memory and less overhead per core, enabling our customers to scale real-time streaming workloads further than ever and unlock new AI and agentic applications.”
National laboratories planning to deploy Vera CPUs include Leibniz Supercomputing Centre, Los Alamos National Laboratory, Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center and the Texas Advanced Computing Center (TACC).
“At TACC, we recently tested Nvidia’s Vera CPU platform as we prepare for deployment in our upcoming Horizon system – and running six of our scientific applications, we saw impressive early results,” said John Cazes, director, high-performance computing, TACC. “Vera’s per-core performance and memory bandwidth represent a giant step forward for scientific computing, and we look forward to bringing Vera-based nodes to our CPU users on Horizon later this year.”
Leading cloud service providers planning to deploy Vera CPUs include Alibaba Cloud, ByteDance, Cloudflare, CoreWeave, Crusoe, Lambda, Nebius, Nscale, Oracle Cloud Infrastructure, Together.AI and Vultr.
Leading infrastructure providers adopting Vera CPUs include Aivres, ASRock Rack, ASUS, Compal, Cisco, Dell, Foxconn, GIGABYTE, HPE, Hyve, Inventec, Lenovo, MiTAC Computing, MSI, Pegatron, Quanta Cloud Technology (QCT), Supermicro, Wistron and Wiwynn.
Availability
Nvidia Vera is in full production and will be available from partners in the second half of this year.
Comments
Nvidia made a major impact with the announcement of its Vera CPU, a highlight strongly emphasized on stage by CEO Jensen Huang.
As a Nvidia ambassador, he claimed the new CPU delivers up to 50% better performance compared to traditional rack-scale processors - namely those from Intel and AMD.
Huang also underscored the scalability of Nvidia’s approach, both in terms of server chassis density and rack-level expansion. With the Vera architecture, deployments can scale up to 256 CPUs per rack, with an overall scalability ceiling of around 22,500 concurrent CPUs.
This development is highly strategic for Nvidia, especially considering that roughly 60% of its business comes from just five hyperscalers - a point Huang highlighted during his keynote. It also reflects the growing role of Nvidia Cloud Partners (NCPs) in expanding the ecosystem.
The Vera platform integrates closely with Nvidia’s broader portfolio, including BlueField-4, NVLink, SmartNICs, and Spectrum-X networking technologies.
Through its AI Data Platform vision, Nvidia is positioning server vendors as key enablers of adoption. Many have already embraced this shift, launching new products, publishing reference architectures, and showcasing vertical use cases.
This vertical focus - spanning industries such as robotics - was also a central theme of the keynote, pointing to what could be the next major growth driver for the Santa Clara-based company.






