The NVIDIA Vera Rubin platform is creating the next frontier of agentic AI with seven new chips: the NVIDIA Vera central processing unit (CPU), NVIDIA Rubin graphics processing unit (GPU), NVIDIA NVLink 6 switch, NVIDIA ConnectX-9 SuperNIC, NVIDIA BlueField-4 data processing unit (DPU) and NVIDIA Spectrum-6 Ethernet switch, as well as the newly-integrated NVIDIA Groq 3 language processing unit (LPU).
| Source: NVIDIA. Seven different chips are part of the NVIDIA Vera Rubin platform. |
Designed to operate together as a single AI supercomputer, the chips power every phase of AI — from massive-scale pretraining, post-training and test-time scaling to real-time agentic inference.
Through deep
codesign across compute, networking and storage, supported
by an ecosystem of more than 80 NVIDIA MGX ecosystem partners with a
global supply chain, NVIDIA Vera Rubin offers the most extensive NVIDIA
pod-scale platform — a supercomputer where multiple racks purpose-built
for AI work together as one massive, coherent system, NVIDIA said.
“Vera
Rubin is a generational leap — seven breakthrough chips, five racks,
one giant supercomputer — built to power every phase of AI,” said Jensen
Huang, founder and CEO of NVIDIA.
“The agentic AI inflection point has arrived with Vera Rubin kicking off the greatest infrastructure buildout in history.”
AI
labs and frontier model developers including Anthropic, Meta, Mistral
AI and OpenAI are looking to use the NVIDIA Vera Rubin platform to train
larger, more capable models and to serve long-context, multimodal
systems at lower latency and cost than with prior GPU generations.
“Enterprises and developers are using Claude for increasingly complex reasoning, agentic workflows and mission-critical decisions. That demands infrastructure that can keep pace,” said Dario Amodei, CEO and cofounder of Anthropic.
“NVIDIA's Vera Rubin
platform gives us the compute, networking and system design to keep
delivering while advancing the safety and reliability our customers
depend on.”
“NVIDIA infrastructure is the foundation that lets us keep pushing the frontier of AI,” said Sam Altman, CEO of OpenAI.
“With
NVIDIA Vera Rubin, we’ll run more powerful models and agents at massive
scale and deliver faster, more reliable systems to hundreds of millions
of people.”
According to NVIDIA, AI infrastructure is evolving
from discrete chips and standalone servers to fully-integrated
rack-scale systems, POD-scale deployments, AI factories and sovereign
AI. These advances are driving dramatic gains in performance, improving
cost efficiency for organisations of all sizes and across industries,
while helping democratise access to AI and improve energy efficiency to
power the world’s most demanding workloads.
New systems include the:
- NVIDIA Vera Rubin NVL72 rack
- NVIDIA Vera CPU rack
- NVIDIA Groq 3 LPX rack
- NVIDIA BlueField-4 STX storage rack
- NVIDIA Spectrum-6 SPX Ethernet rack
NVIDIA, along with over 200 data centre infrastructure partners, also
announced the NVIDIA DSX platform for Vera Rubin. DSX covers AI factory design and deployment. NVIDIA further released the Vera Rubin DSX AI Factory reference design, a blueprint for codesigned AI infrastructure that maximises tokens per
watt and overall goodput.
Details
Vera Rubin-based products will be available from partners starting 2H26. This includes cloud providers and global system manufacturers.
Hashtags: #GTC, GTC2026
*NIC stands for network interface card.
No comments:
Post a Comment