Skip to content
llm-speed

Best rig for Qwen3-Coder-Next

Qwen3-Coder-Next is an 80B-parameter MoE with ~3B active. The activation pattern means it punches above its weight on Apple Silicon and sane consumer GPUs. Here's the data we have.

Verdict

No Qwen3-Coder-Next submissions yet. The 80B-A3B MoE shape means even an M-series Ultra or a single MI300X should hold the 4-bit quant in unified / VRAM with room for a useful context window — be the first to submit and this verdict will rewrite around your numbers.

No data submitted for this task yet.

Run the suite to be the first benchmark for this guide:

$ pipx install llm-speed && llm-speed bench

Qwen3-Coder-Next's 80B-A3B mixture-of-experts means raw VRAM matters less than for a dense model of comparable footprint, but you still need ~40-50 GB of memory for a 4-bit quant. That puts it in reach of an M-series Ultra, a pair of 24 GB consumer GPUs, or a single MI300X / H100. Below is every config we have a real submitted run for.

Side-by-side comparisons

See also: All hardware · All models · Methodology