CraftRigs
Architecture Guide

Vera Rubin vs Hopper: What NVIDIA's GTC 2026 Announcement Means for Local AI Builders

By Chloe Smith 7 min read

Some links on this page may be affiliate links. We disclose it because you deserve to know, not because it changes anything. Every recommendation here comes from benchmarks, not budgets.

The number Jensen Huang dropped on Monday was "10x." Ten times lower inference cost than the previous generation. If you follow AI hardware, you already know those numbers get slippery fast — and this one is slipperier than most.

Here's what actually matters: that 10x claim is versus Blackwell, not Hopper. If you're running H100s or planning to buy used ones, the real performance gap between Vera Rubin and your hardware is closer to 30 to 50x. That's the figure nobody in the GTC 2026 press coverage is foregrounding. And it's the number that should determine what you buy this year.

Let's work through it.


What Vera Rubin Actually Is

Named after the astronomer who gave us the first strong evidence for dark matter, Vera Rubin is NVIDIA's successor to Blackwell. Jensen previewed it at GTC 2025, formally announced it at CES 2026 in January, shipped first samples to customers in late February, and put it front and center at today's keynote at SAP Center in San Jose — 30,000 attendees, 190 countries, the whole circus.

The spec sheet is genuinely absurd:

  • 288GB of HBM4 memory per GPU — versus 80GB on the H100
  • 22 TB/s memory bandwidth — versus 3.35 TB/s on the H100 (a 6.6x jump)
  • 50 petaFLOPS of NVFP4 inference per chip — 5x faster than Blackwell's B200
  • 336 billion transistors — 1.6x more than Blackwell
  • NVLink 6 interconnect for rack-scale GPU-to-GPU bandwidth
  • NVL72 rack: 72 Rubin GPUs + 36 Vera CPUs = 3.6 exaFLOPS total inference

Note

Vera Rubin at a glance: Six co-designed chips: Rubin GPU, Vera CPU, NVLink 6 Switch, ConnectX-9 SuperNIC, BlueField-4 DPU, and Spectrum-6 Ethernet. The whole system has 1.3 million components. Hyperscaler availability: H2 2026. General cloud availability for everyone else: realistically 2027.

The design philosophy NVIDIA is calling "extreme codesign" — all six chips engineered together rather than a GPU bolted onto commodity infrastructure — is actually central to the performance numbers. A lot of the gains live in bandwidth and interconnect improvements that don't show up cleanly in raw FLOPS comparisons.


The 10x Claim, Decoded

When NVIDIA says "10x lower inference cost per token versus the previous generation," they mean Blackwell. Specifically, the GB200 NVL72 rack currently shipping to hyperscalers. Not Hopper. Not H100.

The figure is also a cost number, not a raw throughput number. That distinction matters. Cost per token is a function of throughput, power efficiency, and memory efficiency simultaneously. Rubin hits 10x by combining:

  • 5x raw inference throughput per GPU over Blackwell B200
  • 10x performance per watt versus Grace Blackwell (per NVIDIA's own figures to CNBC)
  • 4x fewer GPUs needed to train MoE models, which reduces rack footprint and operational cost

So where does 10x come from if raw throughput is only 5x? Electricity. In a world where power costs dominate inference economics at hyperscaler scale, running twice the compute for the same wattage closes the other half of the gap. It's not marketing math — it's infrastructure math.

But here's what that also means: if you're benchmarking from Hopper, not Blackwell, the 10x figure is the wrong starting point entirely.


The Real Rubin vs Hopper Gap: 30–50x

Here's the arithmetic most coverage is skipping.

The H100 (Hopper architecture) delivers roughly 3.96 petaFLOPS of FP8 inference performance. Rubin delivers 50 petaFLOPS of NVFP4. Raw compute ratio: 12.6x. But that understates the real-world gap.

Blackwell's B200 delivers approximately 4x the inference throughput of H100 on actual transformer workloads — tokens per second, not FLOPS. Rubin is 5x over Blackwell. Multiply: 4x × 5x = 20x in raw throughput terms, Rubin over H100.

Caution

Why the gap is wider than the math suggests: Raw throughput (20x) undersells Rubin's real-world advantage over Hopper. The HBM4 bandwidth jump (22 TB/s vs 3.35 TB/s) is 6.6x alone — and for LLM inference, memory bandwidth is usually the binding constraint, not compute. Stack NVLink 6 interconnect gains, the disaggregated inference architecture, and six co-designed chips removing bottlenecks between compute layers, and effective inference efficiency for large-model workloads runs 30 to 50x over H100 depending on model size and batch configuration.

Think about what that means for competitive inference pricing. An H100 cluster competing with a Rubin cluster needs to charge dramatically more per token to stay solvent. That's not a sustainable position for businesses built around inference margin.

Jensen Huang said as much on stage, without quite framing it in those terms. The "10x vs Blackwell" number is technically accurate and consistently misleading for anyone still on Hopper hardware.


How Hyperscaler Upgrade Cycles Create Used Hardware Supply

There's a thing that happens every time NVIDIA ships a generation. Hyperscalers — AWS, Google, Microsoft, CoreWeave — have to choose between running current hardware longer or upgrading. When inference costs drop by an order of magnitude for competitors on newer silicon, the choice gets forced.

That creates a predictable flood of used hardware on secondary markets. It happened with Ampere when Hopper shipped. It'll happen with Hopper when Rubin ships, just with a slightly longer lag.

The timeline isn't immediate. Hyperscalers extended GPU depreciation schedules to six years after the Blackwell transition — partly accounting decisions, partly genuine inference demand for previous-gen hardware. Jensen himself joked that "when Blackwell ships, you couldn't give Hoppers away" — and turned out to be wrong. CoreWeave's H100 GPUs from 2022 contract expirations immediately rebooked at 95% of original pricing. Inference demand absorbed the supply.

Rubin changes that dynamic more aggressively. It's not a marginal improvement. The 30-50x performance gap makes running H100s for competitive inference pricing genuinely painful in a Rubin world. Secondary market supply will increase, and this time there's real question whether demand absorbs it as cleanly.


The Used H100 Market Timeline

Right now, H100s that sold new for $40,000 are trading on eBay around $6,000 — an 85% drop. Cloud rental has fallen to $2.10 to $3.50 per GPU-hour depending on provider. For continuous workloads at high utilization, purchasing hardware breaks even against cloud rental in roughly 10 to 15 months.

The secondary market cascade pattern follows a predictable arc: years 1-2, H100s handle frontier training workloads for hyperscalers; years 3-4, inference; years 5-6, batch workloads. An H100 fleet deployed by a hyperscaler in 2022 hits the 4-year mark in 2026 — exactly when Rubin ships.

Tip

The buy window to watch: Rubin hits hyperscaler production in H2 2026. The first wave of Hopper offloads will likely hit secondary markets in Q1-Q2 2027, as hyperscalers reconfigure capacity for Rubin racks. That's when used H100 prices could soften from ~$6,000 toward $3,000-$4,000. If you don't need GPU compute for the next 12 months, waiting is rational.

Will prices actually crater? Maybe not dramatically. Inference demand is still high enough that older hardware has buyers. But the rate of new supply entering secondary markets will accelerate sharply in 2027, and that's what matters for pricing — not whether there's demand for H100s in the abstract.

There's also a Blackwell wildcard here. As Rubin ships, B200 hardware starts its own secondary market journey. Used Blackwell arriving in late 2027 or 2028 will itself put pressure on H100 pricing. The cascade compresses.


What Local Builders Should Do Right Now

Vera Rubin is not a GPU you'll be running locally or renting affordably in 2026. The NVL72 rack — 72 GPUs, 36 CPUs, 1.3 million components — is a hyperscaler product. When NVIDIA says "H2 2026," they mean AWS, Google, and Microsoft get it first. Meaningful cloud availability for non-hyperscaler customers probably stretches into mid-2027. Consumer-tier access? Unknown.

So the GTC 2026 keynote matters to you not because you're buying Vera Rubin, but because it starts the used-market timer.

If you need GPU compute right now for local inference: a used H100 at $6,000 is a reasonable buy for the first time in this hardware's history. That price point was unthinkable 18 months ago. For local LLM inference running 70B parameter models, 80GB of HBM3 handles the job without quantization compromises. The break-even math against cloud rental works.

If you can wait 9 to 12 months: do it. The Rubin announcement signals hyperscaler upgrade cycles are now in motion. The secondary market in late 2026 and early 2027 will have substantially more H100 supply than today. Prices will follow. This is the most favorable used-H100 buying environment in the hardware's lifetime, and it's about to get better.

If you're already running H100s: you're not in crisis. Inference demand is high enough that the H100 still pencils out for production workloads. You're not competing with Rubin-equipped hyperscalers on price, and your customers probably aren't either. Run the hardware, depreciate it honestly, and plan your next upgrade around 2027 when used Blackwell hardware enters secondary markets at scale.

The headline from GTC 2026 is Vera Rubin. The actual story for local AI builders is what happens to the H100 market in the six months after it ships.


See Also

vera-rubin h100 hopper gtc-2026 nvidia datacenter used-gpu-market local-ai

Technical Intelligence, Weekly.

Access our longitudinal study of hardware performance and architectural optimization benchmarks.