The RX 9060 XT 16GB launched at $349. As of March 2026, you'll pay $439 for the cheapest model on Amazon — and the gap keeps widening. That's not a temporary shortage or a scalper tax. It's AMD making a deliberate strategic bet, and if you're building a local LLM rig, you need to understand what's happening before you make a purchase you'll regret.
Here's the short version: memory costs are blowing up, AMD is pivoting to 8GB configurations to protect margins, and the 16GB cards that local inference actually needs are going to get harder and more expensive to find. The window to buy at anything resembling reasonable prices is probably already closing.
Why DRAM Is the Whole Problem
GDDR6 pricing has been on a tear. TrendForce tracked DRAM prices rising 172% year-over-year by end of Q3 2025, and they've accelerated going into 2026. The culprit isn't mysterious: hyperscale AI infrastructure is consuming HBM at unprecedented rates, pulling semiconductor fab capacity away from conventional GDDR6. Less production capacity means higher prices, and GPU add-in board (AIB) partners are eating the difference.
Both AMD and NVIDIA quietly raised their GPU-and-memory bundle costs to partners in January 2026 — an increase of roughly 5 to 10%. Partners absorbed the first hit, then passed it to retail. That's why you saw the RX 9060 XT 16GB drift from $349 to $389 on Newegg and Best Buy within days of launch, then keep climbing.
[!INFO] The 16GB problem in hardware terms: A 16GB GDDR6 card requires eight memory chips. An 8GB card uses four. At current GDDR6 spot prices, that memory delta is no longer trivial — it's enough to wreck margin math on a mid-range GPU that needs to sell at $349.
TrendForce's February 2026 forecast made the situation more explicit: PC DRAM prices are expected to roughly double as HBM4 production ramps and absorbs remaining fab capacity. If that holds, every GPU with 16GB of GDDR6 becomes structurally expensive to make. Not briefly. Not seasonally. Structurally.
AMD's Response: Go 8GB First
AMD's AIB partners weren't subtle about this. Posts from the Board Channels forum in early February — picked up by VideoCardz and confirmed by Club386, KitGuru, and ThinkComputers — laid it out plainly: AMD is shifting Radeon production to prioritize 8GB models. The 16GB variants aren't being killed off, but they'll be scarce, and the ones that do get made will carry higher prices.
This mirrors what NVIDIA already did with RTX 50-series — the 5060 Ti launched in an 8GB configuration at MSRP, with the 16GB version commanding a significant premium and being far harder to find.
The RX 9070 situation tells the same story from a different angle. AMD is reportedly pulling production away from the standard RX 9070 toward the RX 9070 XT, not because the XT is more popular, but because both cards use the same Navi 48 die and the same 16GB of GDDR6. They cost almost the same to make. But the XT carries a $599 price tag versus $549 for the standard — so AMD can sell the same expensive-to-make product at a better margin by simply making more XTs.
Warning
Don't expect the RX 9070 XT to stay at MSRP. It briefly touched $599 in early March 2026 in the UK market. By the time most buyers see "MSRP deal" headlines and act, the window is usually closed. The XDA report from March 2025 documented exactly this pattern at the original RX 9070 XT launch — cards sold out in minutes, restocks came back at higher prices, and AMD's promised rebate system caused more confusion than relief.
What This Means for Local LLM Work
If you're running Ollama, LM Studio, or llama.cpp on an AMD GPU, VRAM capacity is the only number that matters for inference. Everything else — clock speed, shader count, ROCm version — is secondary. The model either fits in VRAM or it doesn't.
Here's the practical breakdown as of March 2026:
8GB VRAM gives you:
- Llama 3.1 8B at Q4_K_M quantization (~4.7GB) — fast, fits cleanly
- Qwen2.5 7B at Q5_K_M — borderline, works
- Anything 13B or above — no, not without heavy layer offloading to RAM, which tanks your tokens-per-second to CPU-level speeds
16GB VRAM gives you:
- Llama 3.1 13B at Q8_0 — fits and runs well
- Qwen2.5 14B, Mistral 7B at high quality — comfortable headroom
- 32B+ models at aggressive quantization — possible, not fast
- Context window flexibility — 16GB lets you push context length without falling off the VRAM cliff mid-conversation
The difference between 8GB and 16GB isn't just a quantitative step. It's the difference between running the models that have become genuinely useful for coding work and research versus running glorified toys. A 7B model at Q4 is fine for quick lookups. For anything that requires multi-step reasoning, document analysis, or extended context — 13B minimum, 14B preferred.
Tip
ROCm is actually usable now. AMD's open-source GPU compute stack has historically been the reason people defaulted to CUDA and paid the Nvidia premium. That calculus has shifted in 2026. Ollama, LM Studio, and llama.cpp all have solid ROCm support. If you haven't tried AMD for local inference in the past year, your old experience is out of date.
The Actual Buying Guide
Given all this, here's what the market looks like and what makes sense.
Option 1: RX 9060 XT 8GB ($299 MSRP)
Available. Stable price. And basically useless for serious local LLM work. 8GB hard caps you at 7B-8B models. If you're buying a GPU specifically for local AI inference, this is the wrong card. If gaming is the primary use case and LLM is secondary, fine — but don't mistake "available at MSRP" for "good value for this workload."
Option 2: RX 9060 XT 16GB ($439–$529, March 2026)
The card people actually want. It launched at $349 and now sits $90–$180 above that. Given the direction DRAM costs are heading, it's probably not going back to $349. Whether $439 is worth it depends on your alternative. For a first 16GB AMD card with solid RDNA 4 performance and ROCm support, it's defensible — barely.
Option 3: RX 9070 XT ($599 MSRP, often $650–$720)
The better card for local LLM given its higher memory bandwidth and extra compute, and it does occasionally hit near MSRP. But "occasionally" is the operative word. If you see it at $610 or under, that's probably worth grabbing. At $720, you're paying an effective tax for AMD's production prioritization decisions and you should look elsewhere.
Option 4: Used RTX 3090 (~$600–$750)
24GB of VRAM, a mature CUDA ecosystem, and prices that have stabilized on the secondary market. This is the honest answer for serious local inference work on a budget. Yes, it's old silicon. It runs Llama 3.1 70B at Q4 quantization. For an LLM rig, "old but 24GB" beats "new but 8GB" every single time.
The main friction is ROCm vs CUDA — llama.cpp on CUDA is more widely tested, slightly more stable, and doesn't require the occasional ROCm setup headache. For AMD loyalists or people building primarily for gaming with inference as a secondary use, AMD makes sense. For someone whose primary goal is running large models locally, the 3090's ecosystem is hard to beat.
Option 5: RX 7900 GRE (16GB, ~$400–$500 used)
Often overlooked. RDNA 3, 16GB GDDR6, and it actually runs local LLMs well via ROCm. No RDNA 4 features, but inference doesn't care about ray tracing or upscaling. If you find a clean one under $450, that's a sensible buy.
The Bigger Picture
AMD isn't abandoning 16GB GPUs. But when a company is structurally incentivized to make 8GB cards — because they're cheaper to produce, still carry healthy margins, and sell at close to the same price as last year — that's where production capacity goes. The 16GB variants become increasingly the "if you can find one" SKU.
This is awkward timing. Local LLM adoption has been growing fast, and the GPU market has responded by making the cards that actually serve that use case harder to buy at fair prices. The irony is pretty thick: the memory shortage driving 8GB prioritization is itself partly caused by AI infrastructure demand — which is the same force pushing people toward local inference in the first place.
What actually makes sense: if you're building a dedicated inference rig right now, lean toward the used RTX 3090 market unless you find an RX 9070 XT within $50 of MSRP. If you're upgrading an existing AMD system and 16GB is the priority, the RX 9060 XT 16GB at $439 is overpriced but not absurd given where things are heading. And if 8GB is all you can swing — use it for gaming, run a smaller model for quick tasks, and plan to upgrade the moment 16GB card prices normalize.
They might not normalize for a while.