CraftRigs
articles

NVIDIA's Inference Moat: Why 18 Years of CUDA Still Beats the Competition

By Charlotte Stewart 10 min read

Some links on this page may be affiliate links. We disclose it because you deserve to know, not because it changes anything. Every recommendation here comes from benchmarks, not budgets.


**NVIDIA's inference advantage isn't about the GPU. It's about 18 years of library optimization that AMD hasn't had time to replicate. For power users running serious inference workloads, that moat is real and measurable — but it's not infinite, and it doesn't matter equally for every build.**

## What NVIDIA's Moat Actually Is (And What It Isn't)

Here's the thing people get wrong: NVIDIA's dominance in local inference isn't because their GPUs are better. AMD's RDNA3 and CDNA architectures have comparable theoretical compute. The MI300X has 1.5x the theoretical FLOPs of an H100.

The moat is the software. Specifically: 18 years of hand-tuned kernels in [cuBLAS](/glossary/cublas), [cuDNN](/glossary/cudnn), NCCL, and [TensorRT](/glossary/tensorrt)-LLM. These aren't just libraries — they're thousands of micro-optimized operations for specific GPU architectures, accumulated over nearly two decades of iteration.

[CUDA](/glossary/cuda) launched in 2007. ROCm launched in 2016. That's a 9-year head start, and the gap doesn't disappear when you release a competitive chip. You have to write all the kernels.

### The Maturity Trap

AMD's libraries exist. rocBLAS works. MIOpen handles deep learning ops. But the same workload that runs on a battle-hardened cuDNN kernel — one that's been tuned across four GPU generations with feedback from thousands of production deployments — is running on an AMD equivalent that's had maybe three years of optimization attention.

Torch and Hugging Face pipelines are written CUDA-first. Every time a developer hits a weird edge case, they fix it on CUDA. ROCm gets the fix weeks or months later, if it gets it at all. This is the chicken-and-egg problem that's very hard to break out of: fewer users means fewer bug reports means slower optimization means fewer users.

> [!NOTE]
> This isn't a knock on AMD's engineering. It's a description of how ecosystems compound. ROCm is a genuinely good platform in 2026 — it just started 9 years later.

---

## CUDA's Inference Advantages in Real Numbers

The abstract ecosystem argument is one thing. Here's what it looks like in actual tokens per second.

As of March 2026, [CUDA](/glossary/cuda) typically outperforms ROCm by 10–30% on compute-intensive LLM inference workloads. That gap widens at scale. At 16 concurrent users, NVIDIA's throughput advantage grows meaningfully. At 128 concurrent users, the CUDA gap scores rise further — the batching and out-of-order execution efficiency that NVIDIA has tuned for years shows up most clearly when you're slamming a model with parallel requests.

[TensorRT-LLM](/glossary/tensorrt) is the software that does the heavy lifting here. Against vLLM running on the same NVIDIA hardware, TensorRT-LLM delivers 15–30% higher throughput on H100s, and 20–100% faster in quantized workloads where FP8 optimization applies. That's not a small gap.

For a production deployment where you're burning server time 24/7, a 20% throughput improvement means 20% fewer GPUs to serve the same load. At datacenter pricing, that's real money.

> [!TIP]
> If you're evaluating inference engines for a production deployment, benchmark TensorRT-LLM vs vLLM on your specific model and quantization level. The gap varies significantly — FP8 workloads see the biggest difference, Q4 quantized consumer workloads see the smallest.

### Concrete Benchmark: What Runs What

Before we get into comparisons, a critical hardware reality that most articles get wrong.

The RTX 5090 has 32 GB VRAM. A Q4-quantized Llama 3.1 70B requires approximately 35–40 GB VRAM minimum. A **single** RTX 5090 cannot cleanly run 70B models — you'll get 1–2 tokens/second as the model thrashes into system RAM. The RTX 5080, at 16 GB, doesn't even get that far.

Here's what the hardware actually runs (as of March 2026):

Tokens/sec


~213 tok/s


~61 tok/s


~27 tok/s


~35–40 tok/s (batch)
*Sources: DatabaseMart dual-5090 benchmark, RunPod RTX 5090 benchmarks. As of March 2026. Software: Ollama / llama.cpp. Single-stream inference.*

For the professional tier — H100 vs MI300X — AMD's datacenter card has 192 GB of HBM3 and enormous theoretical throughput. But per SemiAnalysis benchmarking, the MI300X achieves only 37–66% of H100/H200 **realized** inference performance for Llama 3.1 70B workloads. That gap between theoretical and realized is entirely software. See our [GPU benchmarks guide](/guides/gpu-benchmarks-2026-local-llm/) for extended comparisons across model sizes.

---

## Why the Moat Matters for Professionals (And When It Doesn't)

The right question isn't "is NVIDIA better?" It's "how much better, for what workload?"

For a single-user local inference rig, the difference between CUDA and ROCm is roughly 10–30% on tokens per second. If you're running Llama 3.1 8B for coding assistance and it runs at 60 tok/s instead of 80 tok/s, you won't notice. Human reading speed is the bottleneck, not the hardware.

For a business serving 50+ concurrent inference requests, it's a different calculation. NVIDIA's batching efficiency compounds. At 128 concurrent requests, the platform gap grows — meaning you need significantly more AMD iron to serve the same throughput. At scale, that premium on NVIDIA hardware starts paying for itself.

The moat is largest for **batch inference** — multiple requests processed simultaneously. It's smallest for single-request, low-concurrent latency scenarios.

### When AMD Pulls Ahead

AMD wins on **price-per-GB-VRAM** for the datacenter tier, and by a lot. The MI300X has 192 GB HBM3 at approximately $10,000–$18,000 (enterprise pricing, March 2026) — no consumer MSRP exists. Compare that to an NVIDIA H100 at $30,000+. If your workload fits in AMD's ecosystem and you can absorb the ROCm overhead, the cost-per-token math can favor AMD.

AMD also wins on **specific long-context reasoning workloads**. For 1k-input / 4k-output reasoning tasks on Llama 3.1 70B, the MI325X actually surpasses the H100 in peak throughput at longer latency budgets — AMD's larger HBM capacity shines when you're generating long outputs and memory bandwidth is the bottleneck.

### When NVIDIA's Moat Doesn't Exist

Running a single 8B or 14B model for one user? The moat is invisible. The 10–30% software gap represents maybe 15 tokens per second on a model that's already running comfortably fast for a single person.

Running quantized models at Q4 or Q5? The moat shrinks further. Quantized inference is largely memory-bandwidth-bound, not compute-bound — so the compute-level kernel optimizations matter less than raw memory speed. AMD's HBM bandwidth advantage can partially compensate.

> [!WARNING]
> Be careful with benchmark comparisons that show NVIDIA and AMD at near-parity for inference. Most of those benchmarks are single-user, short-context, non-batched workloads — exactly the scenario where the moat is smallest. Scale the workload and the gap reappears.

---

## The Ecosystem Lock-In (Intentional and Real)

[CUDA](/glossary/cuda) isn't one library. It's Torch + ONNX Runtime + vLLM + TensorRT + cuBLAS + cuDNN orchestrated together, and every layer assumes NVIDIA hardware downstream. Switching to AMD means either rewriting inference infrastructure OR accepting the performance hit from using ROCm compatibility layers.

The lock-in runs three layers deep:

**Layer 1 — Model weights.** Many quantized models are produced with CUDA-specific tooling. GPTQ and AWQ quantization libraries were CUDA-first; ROCm support arrived later and is still patchier for edge cases.

**Layer 2 — Inference frameworks.** vLLM, Text Generation WebUI, and LM Studio all prioritize CUDA. ROCm compatibility exists, but you'll hit friction on newer features faster. When a framework ships a new batch inference optimization, NVIDIA gets it first.

**Layer 3 — Production deployment tooling.** [TensorRT-LLM](/glossary/tensorrt), Triton Inference Server, and DeepSpeed are NVIDIA-native. These are the tools that run production AI services. There's no AMD equivalent for TensorRT-LLM — you use vLLM on ROCm, which is good, but it's not the same.

Switch one layer to AMD: maybe 5% performance loss. Switch all three: you're looking at 25–40% total, per most real-world migration experiences. That's what makes NVIDIA's moat structural rather than just technical. See our [NVIDIA RTX 5080 vs MI300X comparison](/comparisons/nvidia-rtx-5080-vs-mi300x/) for more on the framework compatibility differences.

---

## AMD and Intel's Counter-Moves (And Why They Haven't Won Yet)

AMD is not sitting still. ROCm 7 delivered 3.5× better inference performance versus ROCm 6. The MI350 (targeting H1 2026) approaches FP8 parity with comparable NVIDIA parts in inference workloads. AMD's roadmap projections suggest 80–90% CUDA parity by the ROCm 7.x cycle.

That's real progress. NVIDIA held 90% of data center GPU revenue in 2024. In March 2026, that number is down to 86% — AMD is gaining.

But "80–90% parity" still means a 10–20% gap, and in production inference, 10% is meaningful. The MI355X shows faster inference than NVIDIA's B200 on some Llama 3.1 405B workloads — which tells you that AMD can win on specific benchmarks. The challenge is winning across the full software stack consistently.

Intel's Xe is not a factor in professional inference. The Arc A770 with 16 GB VRAM is a hobbyist card. Intel's datacenter Gaudi product exists but doesn't have the ecosystem depth to compete here.

### Will the Moat Persist?

Hardware will converge. The MI325X and MI350 are real competitors to NVIDIA's current offerings on raw silicon. The moat that remains is software, and software gaps close faster than hardware gaps.

A reasonable estimate: by late 2026 or 2027, AMD will be a viable choice for 80% of inference workloads. NVIDIA will retain the advantage for the 20% that involves extreme throughput requirements, strict latency SLAs, or production deployments where TensorRT-LLM's optimization depth is the deciding factor. The moat won't disappear — it'll shrink to a defensible niche.

---

## Should You Care? The Decision Tree

Here's how to think about it based on what you're actually building:

| Your Situation | Recommendation
|---|---|
| Hobbyist, 1–2 models, single user | AMD RX 7900 XTX or NVIDIA RTX 5080 ($999) — performance delta invisible at this scale |
| Power user, 34B+ models, occasional batching | NVIDIA RTX 5090 ($1,999) — TensorRT-LLM optimization and 32 GB VRAM justify the cost |
| Power user, 70B models required | Dual RTX 5090 setup — only realistic consumer path to 70B at acceptable speed |
| Professional, inference API, 10+ req/sec | NVIDIA RTX 6000 Ada or H100 — TensorRT-LLM is table stakes here |
| Enterprise, cost-per-token priority, long context | AMD MI300X — HBM3 VRAM advantage and pricing are compelling if ROCm overhead is acceptable |

*Prices as of March 2026 MSRP.*

For the AMD setup path, check our [AMD local LLM setup guide for 2026](/guides/amd-local-llm-setup-2026/) — ROCm on consumer GPUs is actually usable now in ways it wasn't 18 months ago.

---

## The Honest Take: Moat, Not Monopoly

NVIDIA's inference advantage is built on legitimate engineering, not lock-in tricks. They invested in CUDA for 18 years when nobody was certain AI inference would become a trillion-dollar market. That investment compounded. The performance edge is real and measurable.

But it's shrinking. AMD's ROCm 7 improvements, the MI350 hardware launch, and the broader open-source inference ecosystem are all chipping away at ground that seemed immovable two years ago. By 2027, the decision tree above will look different — some of those NVIDIA checkboxes will have AMD alternatives next to them.

Choosing AMD today is a calculated bet that software catches up faster than you need it to. That bet is more defensible in March 2026 than it was in 2024. Choosing NVIDIA is acknowledging that proven, mature software is worth a premium when your workload demands it. Both choices are rational. The wrong move is pretending the trade-off doesn't exist.

---

## FAQ

**What exactly is NVIDIA's CUDA moat?**
CUDA is a parallel computing platform and programming model that NVIDIA launched in 2007 — nearly two decades before AMD's ROCm reached production quality. The moat isn't the platform itself; it's the libraries built on top of it: cuBLAS (linear algebra), cuDNN (neural network ops), TensorRT-LLM (inference optimization). These have been tuned across multiple GPU generations with production feedback from thousands of deployments. Replicating that depth of optimization requires the same investment and time, which AMD is now doing — but the head start compounds.

**Is TensorRT-LLM really that much faster than vLLM?**
For batch inference on NVIDIA hardware, yes — 15–30% higher throughput on H100s, and up to 2× in FP8 quantized workloads. For single-user streaming inference, the gap is smaller (10–15%) and often imperceptible. The key advantage of TensorRT-LLM is that it doesn't exist for AMD — vLLM on ROCm is your best option on AMD hardware, and it's genuinely good, but you're leaving TensorRT's optimizations on the table.

**Can an RTX 5090 run 70B models?**
A single RTX 5090 (32 GB VRAM) cannot run Llama 3.1 70B cleanly — the Q4 quantized version needs approximately 35–40 GB VRAM, so the model spills into system RAM and runs at 1–2 tok/s. A dual RTX 5090 setup solves this, delivering around 27 tok/s on 70B models — comparable to H100 speeds in single-stream inference. If 70B is your target, budget for the dual-GPU path.

**Is AMD's MI300X a realistic NVIDIA alternative?**
For enterprise deployments where cost-per-token matters and you have the engineering resources to work with ROCm, yes. The MI300X's 192 GB of HBM3 is a genuine advantage for long-context workloads and large batch sizes. For individual power users or small teams, no — MI300X starts at $10,000–$18,000 and has no consumer equivalent. The comparison isn't RTX 5090 vs MI300X; it's H100 vs MI300X.

**When does ROCm actually catch up to CUDA?**
ROCm 7 delivered 3.5× better inference performance over ROCm 6, which shows the trajectory. AMD is targeting near-CUDA parity by late 2026. A realistic read: AMD will be competitive for most inference workloads by end of 2026 or mid-2027. NVIDIA will retain a meaningful edge for extreme-throughput, strict-latency production deployments for at least 2–3 more years after that. The moat will shrink to a smaller but still real premium.
cuda inference nvidia amd-rocm local-llm

Technical Intelligence, Weekly.

Access our longitudinal study of hardware performance and architectural optimization benchmarks.