CraftRigs
articles

How Chinese Open-Source Models Changed What You Need in a Local AI Rig

By Charlotte Stewart 11 min read

Some links on this page may be affiliate links. We disclose it because you deserve to know, not because it changes anything. Every recommendation here comes from benchmarks, not budgets.


What took 40GB of VRAM and a $1,600 RTX 4090 running with heavy CPU offloading in 2024 — producing Llama 2 70B output at 15–18 tokens per second — now runs faster and with better quality on a 24GB RTX 4090 you can buy used for around $1,400, loading a Qwen 2.5 32B model that fits entirely on-GPU with no offloading at all. Chinese open-source labs didn't add incremental improvements. They changed what "good enough hardware" actually means. **If you're still planning a local AI rig around 2024 assumptions, you're likely overspending by $400–$800 — or capping yourself at models that are now one generation behind.**

## The 2024-to-2026 Hardware Equation Flipped

In early 2024, the consensus in local AI communities was clear: if you wanted quality output, you needed a 70B model. Llama 2 70B was the benchmark everyone referenced. The problem is that at [Q4](/glossary/quantization) quantization, Llama 2 70B weighs approximately 35–40GB — too large for any single consumer GPU at the time. Running it on a single RTX 4090 (24GB VRAM) required significant CPU offloading, which tanked [tok/s](/glossary/tok-s) to the 15–18 tok/s range on llama.cpp (as tested and documented in the llama.cpp benchmarking community).

The hardware tax was steep. A single RTX 4090 launched at $1,599. That was the floor if you wanted a usable 70B setup. Two RTX 4090s for proper multi-GPU inference meant $3,200+ in GPU spend alone — before touching RAM, CPU, storage, or case.

2026 Setup


RTX 4090 (~$1,400–1,600 used)


24GB (fully on-GPU)


Qwen 2.5 32B (native, ~18–20GB)


~18–22 tok/s, no offloading


~87%+ (Qwen 2.5 32B)


Single RTX 4090
That's the same dollar spend — possibly less — producing a model that fits entirely on the GPU, runs faster, and substantially outperforms the 2024 gold standard.

### Why Most Builders Haven't Caught Up

Most local AI communities are still discussing 70B as the baseline. Reddit's r/LocalLLaMA, YouTube creator benchmarks, and most buyer's guides reference 2024-era model performance. The content cycle for hardware guides runs 12–18 months behind model releases. DeepSeek-V2 dropped in May 2024. Qwen 2.5 launched in September 2024. DeepSeek-V3 in December 2024. But the "you need 70B" advice is still being handed out to beginners in early 2026 as if none of that happened.

---

## What Makes Chinese Models Different (It's Not Just Parameters)

There are two separate improvements happening in Chinese open-source models — and conflating them causes confusion.

The first is architectural: [Mixture of Experts](/explainers/moe-vs-dense/). The second is data quality and training methodology. Both matter, and they solve different problems.

### MoE vs. Dense Models Explained

A dense model like Llama 2 70B activates all 70 billion parameters on every single inference step. Every token you generate runs through the full network. That's why it needs 35–40GB of VRAM at Q4 — those parameters all have to be accessible.

A Mixture of Experts model works differently. DeepSeek-V3, for example, has 671B total parameters — but a learned router selects only 37 billion of them to activate per token. The rest sit idle. The total model file still requires multi-GPU or enterprise hardware to load in full. But this architecture — sparse activation of parameters — is what enabled DeepSeek to train a model that punches at 671B capacity while using only 37B compute per step. Smaller distilled variants of DeepSeek (the R1-Distill 7B and 14B series) strip this down to something a single consumer GPU can handle.

The practical implication: the architectural research from DeepSeek's MoE work fed directly into the training philosophy behind Qwen 2.5 — a dense model family that achieves far higher accuracy per parameter than 2024's generation.

### Why Quality Improved Without More Parameters

Qwen 2.5 32B is a dense model — not MoE. Every parameter is active. But it was trained on higher-quality instruction data, uses better alignment techniques, and benefits from 18 months of post-GPT4 fine-tuning research that simply wasn't available when Llama 2 was built. 

The result is measurable. Llama 2 70B scored approximately 56.8% on GSM8K (multi-step math problems). Qwen 2.5 32B is in the 87%+ range on the same benchmark. DeepSeek-V3 scores 89.3% on GSM8K per its official technical report — from a model that technically has 70B worth of active compute per token, not 671B.

More parameters weren't what 2024 models needed. Better training data was.

---

## Hardware Implications by Tier (What You Actually Need Now)

> [!NOTE]
> The VRAM numbers below assume Q4_K_M quantization via llama.cpp or Ollama. Switching to higher quants (Q6, Q8) improves output quality slightly but requires more VRAM.

### Budget Builder Tier (12–16GB VRAM, ~$700–$1,100 build)

**2024 reality:** A 12GB RTX 4070 Ti got you Llama 2 13B at modest quality. The 70B model was completely out of reach — not even worth attempting without a second GPU.

**2026 reality:** That same 12GB runs Qwen 2.5 14B at Q4, which fits in ~9GB and leaves headroom for context. Qwen 2.5 14B outperforms Llama 2 70B on reasoning and math. Same price range, dramatically better output.

2026 Setup


RTX 4070 Ti (12GB, ~$590 used, ~$849 new as of March 2026)


Qwen 2.5 14B at Q4 (~9GB)


~30–40 tok/s


Competitive reasoning, strong math
The 14B tier changed most dramatically for budget builders. You didn't get a faster GPU. You got a better model that fits the same hardware.

> [!TIP]
> If you're buying new, the RTX 5070 ($549 MSRP, but check current retail — prices vary due to supply constraints) offers the same 12GB VRAM with ~10–15% better power efficiency than the RTX 4070 Ti. For a budget build, used RTX 4070 Ti at ~$590 is the better value right now.

### Mid-Range Tier (24GB VRAM, ~$1,400–$1,800 build)

This is where the 2026 shift is most dramatic. In 2024, the 24GB RTX 4090 was the starting point for 70B access — and even then, only with CPU offloading. The Llama 2 70B model file at ~35–40GB exceeded the GPU's 24GB by a significant margin.

Now, Qwen 2.5 32B at Q4_K_M fits in approximately 18–20GB — entirely within the RTX 4090's 24GB. No offloading. No performance cliff. Just native inference at a model that scores 30+ percentage points higher than Llama 2 70B on benchmark tasks.

2026 Setup


RTX 4090 (24GB, ~$1,400–$1,600 used today)


Qwen 2.5 32B (on-GPU, ~18–22 tok/s)


~4–6GB VRAM free for context/KV cache


Substantially outperforms 2024 gold standard
The 24GB GPU didn't change. The model did.

For buyers in this tier, the used RTX 4090 market is increasingly compelling. RTX 50-series supply shortages have pushed used 40-series prices down. A used RTX 4090 at ~$1,400–$1,600 paired with a Qwen 2.5 32B setup is a better local AI rig than a new RTX 5090 at MSRP — and easier to actually buy.

### Power User Tier ($2,000–$3,500)

Power users want 70B+ model access at usable speeds, fine-tuning capability, or multi-user serving. This tier changed less than the others in absolute terms, but the options improved significantly.

**Option A — RTX 5090 ($1,999 MSRP, currently $2,500–$5,000+ retail due to supply shortages):** 32GB GDDR7 VRAM. Runs Qwen 2.5 72B at Q4 (~38–40GB — partial offloading still needed) and handles all 32B models natively with substantial headroom. Best for: single-user professional deployments.

**Option B — Dual GPU setup (RTX 4090 + RTX 4090, ~$2,800–$3,200 used):** Combined 48GB enables full 70B model access with no offloading, distributed inference via llama.cpp tensor splitting. Access to 70B dense models and DeepSeek-R1 70B distills at ~30–35 tok/s combined.

> [!WARNING]
> The RTX 5090 MSRP of $1,999 exists on paper only as of March 2026. Actual retail is running $2,500 to over $5,000. Don't build a budget around MSRP — check live prices on [BestValueGPU](https://bestvaluegpu.com) before committing.

---

## Model Comparison: 2024 Baseline vs. 2026 Alternatives

All benchmarks below from official technical reports and leaderboard evaluations. Llama 2 70B benchmarks from Meta's published technical report. Qwen 2.5 scores from the Qwen2.5 technical report (arXiv:2412.15115). DeepSeek-V3 scores from the DeepSeek-V3 technical report (arXiv:2412.19437).

### Performance Benchmarks (Verified, March 2026)

**Math Reasoning — GSM8K (8-shot):**

Fits on Single Consumer GPU?


No (24GB max)


Yes (any 12GB card)


Yes (24GB card)


No
The Qwen 2.5 14B outperforms Llama 2 70B on math by more than 20 percentage points — from a model that runs on hardware that costs $590 used.

**Coding — HumanEval:**

The most meaningful number here isn't Llama 2 70B vs Qwen 2.5 base. It's the specialized coding variant: **Qwen2.5-Coder-32B-Instruct scores 92.7% on HumanEval** (per NVIDIA NIM model card). Llama 2 70B scored approximately 29.9% on the same benchmark. If coding is your use case, the gap is not incremental — it's generational.

> [!NOTE]
> HumanEval tests isolated single-function problems, which overestimates real-world coding performance for all models. A 92.7% HumanEval score does not mean Qwen2.5-Coder solves production coding tasks at that rate. Use it as a relative comparison, not an absolute claim.

### VRAM Requirements at Q4 Quantization

Minimum Card


Dual RTX 4090 or RTX A6000


RTX 4070 Ti (12GB)


RTX 4090 (24GB)


Multi-GPU / cloud only


RTX 4070 Ti (12GB)
The per-parameter efficiency of Chinese models didn't eliminate the need for VRAM — it reset which quality tier each VRAM bucket can reach. 12GB now accesses quality that used to require 40GB.

---

## Why 2024 Recommendations Don't Work for 2026 Builders

The parameter count became less reliable as a quality signal starting in mid-2024. DeepSeek released MoE models in January 2024. DeepSeek-V2 arrived in May 2024. Qwen 2.5 launched in September 2024. By December 2024, when DeepSeek-V3 dropped, the research community had established that instruction-tuned 14B–32B models were competitive with 2024-era 70B baselines on a wide range of tasks.

The hardware guides haven't caught up. Most still recommend the RTX 4090 as "the safe 70B bet" — correct framing for 2024, but now obsolete advice. You don't need 70B if you don't want to.

See the [GPU comparison guide](/guides/gpu-comparison-2026/) for an updated side-by-side of all current options if you want to cross-reference the numbers here.

### The Overspend Trap

Budget builders hear "you need a 70B model" and start calculating how to afford $1,600+ in GPU spend. Some of them buy used RTX 4090s thinking they've solved the problem — only to find that the model barely fits, runs with CPU offloading, and still underperforms what a properly trained 32B model would deliver on the same hardware.

Mid-range builders skip the Qwen 2.5 32B entirely, assuming a 32B model is a downgrade from 70B. They're operating on 2024 intuition about parameter scaling that no longer holds.

For a deeper comparison of Qwen vs Llama families to see how this plays out across tasks, see [Qwen vs Llama: Which Model Family Should You Run?](/comparisons/qwen-vs-llama/)

### What Actually Changed in the Open-Source Ecosystem

The timeline matters for understanding why the advice gap exists:

- **January 2024:** DeepSeek MoE initial models — not yet competitive for general use
- **May 2024:** DeepSeek-V2 — first real signal that sparse MoE could challenge dense 70B
- **September 2024:** Qwen 2.5 family — dense models at 7B/14B/32B/72B with dramatically improved training data
- **December 2024:** DeepSeek-V3 — 671B MoE that achieves near-frontier performance, validates the architecture
- **Early 2026:** Community consensus lagging, but hardware requirements have already shifted

The models that changed local AI hardware requirements emerged in a 7-month window (May–December 2024). Most existing guides were written before that window opened.

---

## What to Actually Buy Now (Tier-by-Tier)

### Budget Builder: The ~$700–$900 Build

**GPU:** Used RTX 4070 Ti (~$590, 12GB VRAM) or new RTX 5070 (~$549 MSRP, check current retail)
**Model:** Qwen 2.5 14B or DeepSeek-R1-Distill-14B at Q4_K_M (~9GB VRAM)
**Performance:** ~30–40 tok/s, outperforms 2024's Llama 2 70B on reasoning/math

The case for used RTX 4070 Ti is strong right now. RTX 50-series supply pressure has pushed used 40-series prices down 25%+ (per GPU market reports, March 2026). The 12GB VRAM ceiling is real — you won't run 32B models — but Qwen 2.5 14B is not a consolation prize. It's a generationally different model than what 12GB could access in 2024.

### Mid-Range: The ~$1,400–$1,800 Build

**GPU:** Used RTX 4090 (~$1,400–$1,600, 24GB VRAM)
**Model:** Qwen 2.5 32B or Qwen2.5-Coder-32B-Instruct at Q4_K_M (~18–20GB VRAM)
**Performance:** ~18–22 tok/s fully on-GPU, with ~4–6GB VRAM free for context

This is the tier where the 2026 shift is most tangible. A mid-range RTX 4090 — same GPU that cost $1,599 new in late 2023 — now natively loads a model that would have required $3,200+ in GPUs to run properly in 2024. If you're doing serious coding, analysis, or writing workflows, the 32B tier is where the quality becomes professional-grade.

### Power User: $2,000+

If you need 70B dense model access, fine-tuning, or multi-user serving, the calculus shifts:

**Single-GPU option:** RTX 5090 (MSRP $1,999, retail reality $2,500–$5,000+ as of March 2026). 32GB VRAM. Handles 32B natively with headroom and gets close on 70B at aggressive quantization.

**Dual-GPU option:** Two used RTX 4090s (~$2,800–$3,200 total). 48GB combined. Handles 70B dense models via tensor split, plus Qwen 2.5 72B at Q4 (~38–40GB). Best option if you need true 70B+ access without the RTX 5090 supply lottery.

For more on multi-GPU inference setups, the [dual GPU local LLM stack guide](/articles/102-dual-gpu-local-llm-stack/) covers tensor splitting, NVLink, and performance expectations in detail.

### Should You Wait for the Next GPU Generation?

RTX 60-series is on the horizon — likely late 2026 or early 2027. If you can defer 6+ months, waiting has merit. But if you have workloads today, the used RTX 4090 market is arguably the best it's ever been for local AI. You're not waiting for better models — the models are already here. You're just waiting for the hardware market to catch up to what the models already need.

---

## FAQ

**Isn't 32B too small? Don't I need 70B?**

In 2024, yes. A well-tuned 2024 Llama 2 70B outperformed most 32B models on reasoning tasks. But that was a function of training data quality, not an immutable law of parameter scaling. Qwen 2.5 32B scores 87%+ on GSM8K math; Llama 2 70B scored 56.8%. The 32B model isn't a downgrade — it's running on better training and a different VRAM budget. Try Qwen 2.5 14B (12GB) or 32B (24GB) for a week before concluding you need to chase 70B.

**Will my 2024 RTX 4090 build be obsolete?**

No. Your RTX 4090 now runs better models than it did in 2024 — Qwen 2.5 32B fits natively in 24GB with no offloading. You could previously only access 70B models via CPU offloading and 15–18 tok/s. Now you can run a model that outperforms that 70B, faster, with VRAM to spare for longer context windows. The hardware didn't become obsolete. Its capabilities unlocked.

**If I downsize now, won't models get bigger and require upgrades?**

The research direction is toward efficiency, not parameter inflation. MoE scaling — more expert modules, sparser activation — uses memory more efficiently than stacking dense parameters. If anything, the trend from 2024 to 2026 suggests the 2026–2027 "state-of-art" small models will fit in even less VRAM than today's equivalents. Betting on 2026 hardware being future-proof is a reasonable wager.

**Should I buy used or wait for new?**

Used RTX 4090 under $1,500: strong buy if you want 32B model access right now. Used RTX 4070 Ti under $600: reasonable for 14B workflows. New RTX 5070 at MSRP ($549): good if you're building from scratch and don't need 32B models. Avoid buying new RTX 5090 or 5070 Ti at current retail markups — you're paying 30–70% above MSRP for supply-constrained inventory. Check [BestValueGPU's price history](https://bestvaluegpu.com) before pulling the trigger on any 50-series card. For a full updated ranking of GPU options by local AI performance, see the [hardware upgrade ladder](/articles/100-local-llm-hardware-upgrade-ladder/).
local-llm gpu-buying-guide deepseek qwen hardware-requirements

Technical Intelligence, Weekly.

Access our longitudinal study of hardware performance and architectural optimization benchmarks.