Nobody told the used GPU market that the RTX 3090 is five years old.
A card that launched in September 2020 at $1,499 — back when that felt absurd — is still selling on eBay for $800 to $900 in early 2026. Not because gamers want it. Because an entirely different crowd showed up with money and a very specific problem: they need 24GB of VRAM and they don't have $3,500 for an RTX 5090.
Meanwhile, AMD quietly released the RX 9070 XT in March 2025 and priced it at $599 MSRP. That price aged about as well as most MSRPs do right now — street pricing has climbed to $729 for the entry variants — but it's new hardware, it's in stock, and it ships with a three-year warranty. Against a used 3090 at $800-900, that's actually a real conversation.
So let's have it.
Why the RTX 3090 Still Costs This Much
People expected the RTX 3090 to crater in price once the 40-series and 50-series landed. Didn't happen. The card peaked around $800-1,000 on the used market and has stayed there.
The reason is AI hobbyists. Specifically, the growing population of people running local LLMs who discovered that VRAM is the bottleneck for everything — not GPU speed, not CUDA cores, not bandwidth. Just raw VRAM capacity.
The 3090's 24GB GDDR6X is the critical number. At Q4 quantization, a 13B parameter model needs roughly 8-10GB. A 30B model needs 18-20GB. The 3090 handles 30B models comfortably. At Q8 — significantly better quality — it handles 13B models without sweating.
No consumer GPU in 2026 touches this at this price. The RTX 5090 has 32GB but costs around $3,500 minimum. The RTX 4090 has 24GB and runs $1,600-2,000 used. Two RTX 3090s using NVLink pool their VRAM into 48GB for roughly $1,700 total. That's enough for 70B models at Q4.
The AI community did the math and kept buying. That's why your $400 expectation on a 2020 GPU was wrong.
Note
RTX 3090 quick specs (2026): 24GB GDDR6X | 384-bit bus | 936.2 GB/s bandwidth | 10,496 CUDA cores | 350W TDP | Ampere architecture | Used price: $800-900
What the RX 9070 XT Actually Is
The 9070 XT is AMD's RDNA 4 mid-range flagship, built on TSMC's 4nm N4P node. It has 64 Compute Units, a 2970 MHz boost clock, 16GB GDDR6 on a 256-bit bus, and a 304W TDP. FP32 performance sits at 48.66 GFLOPS versus the 3090's 35.58 GFLOPS.
The MSRP was $599 at launch. That never materialized in practice — the cheapest current variants are $729 on Amazon (the ASRock Challenger), with premium models like the Sapphire Nitro+ hitting $799.
AMD launched this as an intentional value play. They explicitly abandoned chasing Nvidia at the high end and put everything into RDNA 4's price-to-performance ratio. For gaming, it worked.
Note
RX 9070 XT quick specs (2026): 16GB GDDR6 | 256-bit bus | 644.6 GB/s bandwidth | 4,096 shading units | 304W TDP | RDNA 4 architecture | New retail: $729+
Gaming: The 9070 XT Wins, And It's Not That Close
In synthetic benchmarks, the 9070 XT scores around 29,992 in 3DMark versus the 3090's 19,901. That's a 50.7% higher score on a card that costs less at retail right now.
Real-world gaming is less dramatic but still decisive. In Cyberpunk 2077 at 1440p, the 9070 XT averages around 83 FPS against the 3090's 66 FPS — about 26% faster. Across most modern titles, the 9070 XT leads by somewhere between 15% and 30% depending on the game. The aggregate effective 3D speed is roughly 11% faster when you blend across many titles.
The 3090 was designed for a different era of games. It's 8nm Ampere. The 9070 XT is running 4nm RDNA 4 with twice the pixel fill rate (380.2 GPixel/s vs 189.8 GPixel/s). The architectural gap shows up in rendering-heavy modern titles.
There's also the efficiency angle. The 3090 pulls 350W under load. The 9070 XT's official TDP is 304W, though real-world consumption can hit 351W in extreme scenarios — basically a wash, except the 9070 XT is faster while doing it.
For a pure gamer, the comparison doesn't really need to go any further. A new card with better performance, a warranty, and $100-170 less in cost is the obvious choice.
Tip
If you're gaming at 1440p high refresh and don't care about AI at all, the RX 9070 XT at $729 is the better purchase. The 3090's extra 8GB won't help you in games — 16GB is fine for everything currently shipping.
Local LLM: The 3090 Wins Decisively
This is where the 3090's age stops mattering.
A 13B model at Q8 quantization needs roughly 14-16GB of VRAM. The 9070 XT's 16GB will fit that, barely, but you're running right at the limit — no headroom for the KV cache, which grows with context length. Longer conversations eat into available memory fast.
A 20B or 30B model at Q4 needs 18-20GB. The 9070 XT can't hold that at all. Falls to CPU offloading, which tanks inference speed by an order of magnitude.
The 3090 handles 30B models at Q4 comfortably and runs 13B at Q8 with room to spare. Inference speeds on the 3090 run around 40-60 tokens per second on Llama 3.1 8B and 15-25 tokens per second on a 13B model. That's responsive — fast enough for real interactive use.
But there's another problem with the 9070 XT for AI: the software stack. CUDA, Nvidia's compute platform, outperforms AMD's ROCm by 10-30% on equivalent hardware. More importantly, ROCm's compatibility with AI tools is still catching up. PyTorch officially supports ROCm now, and tools like llama.cpp work with AMD GPUs — but when something breaks, CUDA has a decade of documentation and a much larger community forum to debug against. AMD has been closing the gap, but as of March 2026, the CUDA ecosystem for local AI still has meaningfully better tooling coverage.
Caution
ROCm support has improved but is still narrower than CUDA. If you're planning to run anything beyond mainstream llama.cpp/Ollama setups — fine-tuning, custom quantization pipelines, ComfyUI with specific nodes — expect more friction on AMD hardware.
For a direct comparison that includes the RTX 5090 in the same conversation, see the RTX 5090 vs RX 9070 XT breakdown.
The Price-Performance Breakdown
Here's the honest math, split by use case:
For gaming:
- RX 9070 XT at $729: ~26-30% faster in current titles, new with warranty
- RTX 3090 at $850 average used: slower, no warranty, unknown hardware history
The 9070 XT wins for gaming. You're paying less for a better gaming experience, and you're not gambling on a used card that may or may not have spent two years in a crypto mining farm.
For local AI:
- RX 9070 XT at $729: 16GB caps your models at 13B range, ROCm friction
- RTX 3090 at $850 used: 24GB opens the 20-30B model range, mature CUDA support
The 3090 wins for AI. That extra 8GB is not a minor spec difference — it's the difference between a whole tier of model quality being available or not.
The overlap case: If you want to do both gaming and AI, the 3090 is the worse gaming card but the much better AI card. Whether that tradeoff is worth it depends entirely on how seriously you're running LLMs.
Who Should Actually Buy Each One
Buy the RX 9070 XT at $729 if:
- Gaming is your primary use case
- You want something under warranty that you can return
- You run smaller models (7B, 13B at Q4) if you do any AI at all
- You don't want to think about used GPU inspection checklists
Buy the used RTX 3090 at $800-900 if:
- You're specifically running local LLMs and want models in the 20-30B range
- You want the option to NVLink two cards together later for 48GB pooled VRAM
- CUDA compatibility with your existing AI tooling matters
- You understand how to inspect a used GPU (check temps under load, run VRAM diagnostics before the return window closes)
Don't buy the 3090 if:
- You're mainly a gamer who's "curious about AI." You'll spend more money for a worse gaming card and probably never push past 13B models anyway.
- You were expecting sub-$600 prices. Those aren't coming. The AI hobbyist floor has been established at $800+.
The Verdict
The RTX 3090 is not a bad GPU. It's a genuinely weird GPU — a five-year-old card that got a second life from an entirely different audience and now costs more than most people expected.
For gamers, the RX 9070 XT is the rational purchase. Faster, cheaper, new, under warranty. The 3090's extra VRAM does nothing for gaming in 2026.
For local AI hobbyists who are serious about running larger models, the 3090 remains the best price-per-gigabyte of VRAM on the consumer market. The math doesn't lie: no other card gives you 24GB at that price. The CUDA ecosystem advantage is real, and 24GB genuinely unlocks a tier of model performance that 16GB can't.
They're just not competing for the same buyer anymore. The 3090 belongs to the AI crowd. The 9070 XT belongs to gamers. The only mistake is buying the wrong one for your actual use case.
See Also
- RTX 5090 vs RX 9070 XT for Local LLM: The Real Numbers — when you need 32GB vs 16GB
- Gemma 4 Is Coming: The GPU Sweet Spot for Google's Next Open Model — why 24GB still wins for the next Google open model
- GTC 2026 for Home Lab Builders: What Jensen's Announcements Mean for Your GPU Budget — the post-GTC 3090 buying window