Qwen Models Guide: The AI Family You're Missing
๐ More on this topic: Llama 3 Guide ยท DeepSeek Models Guide ยท Best Models for Coding ยท VRAM Requirements
While everyone was talking about Llama and DeepSeek, Alibaba quietly built the best open-source model family in the world.
Qwen 3 beats Llama 3 at every comparable size. Qwen 2.5 Coder 32B matches GPT-4o on coding benchmarks. Qwen-VL handles vision tasks that most local users assumed needed cloud APIs. And unlike DeepSeek’s 671B behemoths that need a server rack, Qwen ships practical sizes that run on the GPU you already own.
The problem is the lineup is massive. Qwen 3, Qwen 2.5, Coder, VL, MoE variants, thinking mode, non-thinking mode โ it’s a lot. This guide maps out every model that matters for local use, what hardware it needs, and when to pick Qwen over the alternatives.
The Qwen Family at a Glance
Qwen isn’t one model โ it’s a sprawling ecosystem. Here’s what exists and what actually matters for running locally:
| Family | What It Is | Sizes | Why It Matters |
|---|---|---|---|
| Qwen 3 | Latest generation, general purpose | 0.6B, 1.7B, 4B, 8B, 14B, 32B, 30B-A3B (MoE), 235B-A22B (MoE) | Best open-source at every size tier |
| Qwen 2.5 Coder | Code-specialized | 0.5B, 1.5B, 3B, 7B, 14B, 32B | 32B matches GPT-4o on code |
| Qwen3-VL | Vision-language | 2B, 4B, 8B, 32B | Best local vision model |
| Qwen 2.5 | Previous gen general | 0.5Bโ72B | Superseded by Qwen 3 for most uses |
Everything is Apache 2.0 licensed โ fully open for personal and commercial use. Alibaba trained Qwen 3 on 36 trillion tokens across 119 languages. For context, Llama 3.1 was trained on 15 trillion.
Why Qwen Leads Right Now
Three things set Qwen 3 apart from every other open model family in early 2026:
1. Benchmark dominance at every size
Each Qwen 3 model performs like the previous generation one tier up:
| Qwen 3 Model | Performs Like | Implication |
|---|---|---|
| 8B | Qwen 2.5 14B | 14B-class performance in 5 GB VRAM |
| 14B | Qwen 2.5 32B | 32B-class performance in 10-12 GB VRAM |
| 32B | Qwen 2.5 72B | 72B-class performance in 20 GB VRAM |
This isn’t marketing โ it holds up across MMLU, HumanEval, MATH, and GPQA benchmarks. The Qwen 3 32B at Q4 quantization fits on an RTX 3090 and delivers what previously required a dual-GPU setup.
2. Hybrid thinking mode
Qwen 3 can switch between two modes in the same conversation:
- Thinking mode (
/think): The model reasons step-by-step before answering. Uses more tokens, slower, but dramatically better on math, logic, and complex analysis. - Non-thinking mode (
/no_think): Fast, direct responses for simple questions and chat. Same speed as a standard model.
No other model family offers this toggle. DeepSeek R1 always thinks (and dumps reasoning tokens you don’t need for “what’s the weather?”). Llama 3 never thinks. Qwen 3 lets you choose per-message.
3. 119 languages
Most open models handle English well, maybe a few European languages passably. Qwen 3 was explicitly trained on 119 languages with strong multilingual benchmarks. If you work in Chinese, Japanese, Korean, Arabic, or basically anything besides English, Qwen is the clear choice.
Qwen 3: Every Size Ranked
0.6B and 1.7B โ Edge Only
These exist for phones, IoT devices, and embedded systems. They’re not useful for general conversation or any task requiring real understanding. If you’re building something for a Raspberry Pi, they’re there. Otherwise, skip them.
ollama run qwen3:0.6b
ollama run qwen3:1.7b
4B โ Minimum Viable Model
The 4B is where Qwen 3 starts being genuinely useful. It handles simple Q&A, text formatting, basic summarization, and straightforward coding tasks. Fits on 4 GB VRAM at Q4 quantization.
It’s not going to write complex code or handle nuanced reasoning, but for a model that fits on a potato, the quality is surprising.
ollama run qwen3:4b
VRAM: ~2.5 GB at Q4_K_M
8B โ The New Default Small Model
This is the one most people should start with. Qwen 3 8B outperforms Llama 3.1 8B on nearly every benchmark โ MMLU, HumanEval, MATH, GSM8K. It matches what Qwen 2.5 14B could do, but in an 8 GB VRAM envelope.
| Benchmark | Qwen 3 8B | Llama 3.1 8B | Winner |
|---|---|---|---|
| MMLU | 73.8 | 69.4 | Qwen |
| HumanEval | 72.0 | 62.2 | Qwen |
| MATH | 62.8 | 47.2 | Qwen |
| GSM8K | 84.2 | 79.6 | Qwen |
With thinking mode enabled, it gets even better on math and reasoning โ at the cost of 2-3x the tokens per response.
ollama run qwen3:8b
VRAM: ~5 GB at Q4_K_M. Runs comfortably on an RTX 3060 12GB with room for context.
14B โ The Sweet Spot for 16 GB GPUs
If you have 16 GB VRAM (RTX 4060 Ti 16GB, RTX 5060 Ti 16GB, or an older card), the 14B is your model. It performs at the level of Qwen 2.5’s 32B โ strong coding, solid reasoning, good creative writing.
This is where thinking mode really starts to shine. Toggle /think for a coding problem, get step-by-step reasoning. Toggle /no_think for casual chat, get instant responses.
ollama run qwen3:14b
VRAM: ~10-12 GB at Q4_K_M. Fits on 16 GB cards with reasonable context windows.
32B โ Best Model on 24 GB
The Qwen 3 32B is arguably the single best model you can run on a 24 GB GPU. It matches Qwen 2.5 72B performance โ a model that previously needed 48+ GB โ and fits on a single RTX 3090 or 4090 at Q4 quantization.
For coding, reasoning, analysis, and creative work, this is the local model that makes cloud APIs optional for most tasks.
ollama run qwen3:32b
VRAM: ~20 GB at Q4_K_M. Tight fit on 24 GB but works with moderate context lengths (~4K-8K tokens). At Q3_K_M (~17 GB), you get more context headroom.
30B-A3B MoE โ Promising but Problematic
On paper, the 30B-A3B is exciting: it’s a Mixture of Experts model that only activates 3 billion parameters per token, so it should run fast while having 30B total knowledge. In practice, there’s a known Ollama issue (GitHub #10458) where GPU utilization drops significantly compared to dense models.
The full model needs ~19-21 GB at Q4_K_M, which defeats the speed advantage since it’s nearly as large as the dense 32B. Until the GPU utilization issue is fixed, stick with the dense models.
# You can try it, but expect suboptimal GPU utilization
ollama run qwen3:30b-a3b
VRAM: ~19-21 GB at Q4_K_M. Not recommended over the dense 32B right now.
235B-A22B MoE โ Competition for DeepSeek R1
The flagship. A 235B parameter MoE that activates 22B per token. Competitive with DeepSeek R1 and GPT-4o on benchmarks, including a CodeForces ELO of 2056. But at ~140 GB for Q4 quantization, this is multi-GPU territory or heavy CPU offloading.
If you’re running a home server with multiple GPUs or substantial RAM for CPU inference, it’s the most capable open model available. For everyone else, the 32B dense gets you 80% of the way there on realistic hardware.
ollama run qwen3:235b-a22b
VRAM: ~140 GB at Q4_K_M. Requires multi-GPU or CPU offloading.
Thinking Mode: Qwen 3’s Killer Feature
Hybrid thinking is what separates Qwen 3 from everything else. Here’s how to actually use it.
How it works
When you type /think before a prompt (or in the system prompt), the model generates internal reasoning tokens before its answer. These reasoning tokens are visible โ you’ll see the model’s chain of thought. When you type /no_think, it responds directly like a standard model.
When to use thinking mode
| Use Case | Mode | Why |
|---|---|---|
| Math problems | /think | Step-by-step reasoning catches errors |
| Code debugging | /think | Model traces through logic systematically |
| Complex analysis | /think | Breaks down multi-part questions |
| Quick chat | /no_think | No overhead, instant responses |
| Text formatting | /no_think | Simple task, thinking wastes tokens |
| Translation | /no_think | Direct task, no reasoning needed |
Controlling thinking in Ollama
Per-message toggle:
>>> /think What is the integral of x^2 * e^x?
>>> /no_think Translate "hello" to French
Interactive session toggle:
>>> /set think # Enable thinking for all messages
>>> /set nothink # Disable thinking for all messages
CLI flag:
ollama run qwen3:8b --think=false # Start with thinking off
The token cost
Thinking mode generates significantly more tokens โ typically 2-5x more than non-thinking mode for the same question. This means slower responses and more memory usage for the KV cache. On hard math problems, the tradeoff is worth it. For chat, it’s not.
Permanent thinking control via Modelfile
If you want a version of Qwen 3 that never thinks (or always thinks), create a custom Modelfile:
FROM qwen3:8b
# Never think โ fast chat mode
PARAMETER think false
# Or: always think โ reasoning mode
# PARAMETER think true
ollama create qwen3-chat -f Modelfile
ollama run qwen3-chat
Qwen 2.5 Coder: Still the Best for Code
Qwen 3 is the better general-purpose model, but for pure coding tasks, Qwen 2.5 Coder 32B remains dominant. It was trained specifically on code and scores 88.4% on HumanEval โ matching GPT-4o.
| Model | HumanEval | MBPP | Best For |
|---|---|---|---|
| Qwen 2.5 Coder 32B | 88.4% | 82.5% | Code generation, completion |
| DeepSeek Coder V2 16B | 81.1% | 77.8% | Lighter code tasks |
| Llama 3.1 8B | 62.2% | 68.4% | General + some code |
| Qwen 3 32B (thinking) | ~85% | ~80% | Code with reasoning |
The 32B Coder fits on 24 GB at Q4 quantization. If coding is your primary use case โ pair programming, code completion, debugging โ this is still the model to run.
Smaller sizes exist (7B, 14B) for tighter VRAM budgets. The 7B Coder is a solid option for 8 GB cards when you need code-specific performance over general chat.
ollama run qwen2.5-coder:32b
ollama run qwen2.5-coder:14b
ollama run qwen2.5-coder:7b
For day-to-day coding with occasional non-code tasks, Qwen 3 32B with thinking mode is catching up and may be the better all-rounder. For dedicated code generation pipelines, the specialized Coder model still wins.
Qwen-VL: Vision That Actually Works Locally
Qwen3-VL is the most capable vision-language model you can run locally. It handles image description, document reading, chart analysis, GUI interaction, and even video understanding.
| Size | VRAM (Q4) | Context | Best For |
|---|---|---|---|
| 2B | ~2 GB | 256K | Phone/edge, basic OCR |
| 4B | ~3 GB | 256K | Lightweight image tasks |
| 8B | ~6 GB | 256K | General vision, documents |
| 32B | ~20 GB | 256K | Complex analysis, GUI agent |
The 8B is the practical choice for most users. It handles document OCR, image Q&A, and chart interpretation well on a 12-16 GB GPU. The 32B is genuinely impressive โ it can act as a GUI agent, understanding screenshots and suggesting interactions โ but needs 24 GB.
All sizes support 256K native context (expandable to 1M with the Qwen3-2507 update), which is relevant for processing long documents with embedded images.
ollama run qwen3-vl:8b
Using vision in Ollama:
# Describe an image
ollama run qwen3-vl:8b "What's in this image?" ./photo.jpg
# Or from the interactive prompt
>>> What does this diagram show? [image: /path/to/diagram.png]
Compared to Llama 3.2 Vision 11B, Qwen3-VL 8B is competitive while using less VRAM. The 32B significantly outperforms Llama’s vision models on complex tasks.
VRAM Requirements
Here’s what you actually need for each Qwen model at practical quantization levels:
| Model | Q4_K_M | Q5_K_M | Q8_0 | FP16 |
|---|---|---|---|---|
| Qwen 3 0.6B | ~0.5 GB | ~0.6 GB | ~0.8 GB | ~1.2 GB |
| Qwen 3 1.7B | ~1.2 GB | ~1.4 GB | ~2 GB | ~3.4 GB |
| Qwen 3 4B | ~2.5 GB | ~3 GB | ~4.5 GB | ~8 GB |
| Qwen 3 8B | ~5 GB | ~6 GB | ~9 GB | ~16 GB |
| Qwen 3 14B | ~10 GB | ~12 GB | ~16 GB | ~28 GB |
| Qwen 3 32B | ~20 GB | ~24 GB | ~36 GB | ~64 GB |
| Qwen 3 30B-A3B | ~19 GB | ~22 GB | ~34 GB | ~60 GB |
| Qwen 2.5 Coder 32B | ~20 GB | ~24 GB | ~36 GB | ~64 GB |
| Qwen3-VL 8B | ~6 GB | ~7 GB | ~10 GB | ~18 GB |
Which quantization? Q4_K_M is the default recommendation โ best balance of quality and size. Move to Q5_K_M if you have the VRAM to spare. Drop to Q3_K_M only if you’re truly VRAM-constrained. More on quantization.
What to run on your GPU
| GPU | VRAM | Best Qwen Model | Ollama Command |
|---|---|---|---|
| RTX 3060 / 4060 | 8 GB | Qwen 3 8B Q4 | ollama run qwen3:8b |
| RTX 3060 12GB | 12 GB | Qwen 3 14B Q4 | ollama run qwen3:14b |
| RTX 4060 Ti 16GB | 16 GB | Qwen 3 14B Q5 | ollama run qwen3:14b |
| RTX 3090 / 4090 | 24 GB | Qwen 3 32B Q4 | ollama run qwen3:32b |
| 2x RTX 3090 | 48 GB | Qwen 3 32B FP16 | ollama run qwen3:32b-fp16 |
| Mac M1 Pro 16GB | 16 GB | Qwen 3 14B Q4 | ollama run qwen3:14b |
| Mac M2/M3 Ultra | 64-192 GB | Qwen 3 32B Q8 / 235B-A22B | ollama run qwen3:32b-q8_0 |
โ Use our Planning Tool to check exact VRAM for your setup.
Qwen vs the Competition
Qwen 3 8B vs Llama 3.1 8B
Qwen 3 wins across the board. The gap is largest on math (MATH: 62.8 vs 47.2) and coding (HumanEval: 72.0 vs 62.2). Llama 3.1 8B’s advantage is the massive fine-tune ecosystem โ if you need Dolphin, Hermes, or other community tunes, they’re all built on Llama. For the base model, Qwen 3 is the better choice.
Qwen 3 32B vs Llama 3.3 70B
This is the interesting comparison. Llama 3.3 70B is technically stronger on some benchmarks, but it needs ~43 GB VRAM versus Qwen 3 32B’s ~20 GB. Per-VRAM-gigabyte, Qwen 3 32B offers dramatically better value. If you have 24 GB, Qwen 3 32B is your model. If you have 48+ GB, Llama 3.3 70B is worth considering.
Qwen 3 vs DeepSeek R1 Distills
DeepSeek R1 distills still lead on pure math and formal reasoning (R1-Distill-Qwen-14B scores 69.7% on AIME 2024). But they always output reasoning tokens, making them slower for casual use. Qwen 3’s thinking toggle means you get the reasoning when you need it and fast responses when you don’t. For a single all-purpose model, Qwen 3 wins. For dedicated math/logic pipelines, the R1 distills are still competitive.
Qwen 2.5 Coder vs DeepSeek Coder V2
Qwen 2.5 Coder 32B (88.4% HumanEval) beats DeepSeek Coder V2 16B (81.1%) convincingly, but it’s also 2x the size. At the same model size, they’re closer. The Coder 32B needs 24 GB; if you only have 16 GB, DeepSeek Coder V2 16B is the pragmatic choice.
Setup Guide
Basic Ollama setup
Install Ollama if you haven’t, then pull the model for your GPU:
# 8 GB VRAM
ollama run qwen3:8b
# 12-16 GB VRAM
ollama run qwen3:14b
# 24 GB VRAM
ollama run qwen3:32b
# Coding-focused
ollama run qwen2.5-coder:32b
# Vision
ollama run qwen3-vl:8b
Custom Modelfile for optimized settings
Create a file called Modelfile:
FROM qwen3:8b
# Set context length (adjust based on VRAM)
PARAMETER num_ctx 8192
# Temperature for creative tasks
PARAMETER temperature 0.7
# System prompt
SYSTEM """You are a helpful coding assistant. Be concise and provide working code examples."""
ollama create my-qwen -f Modelfile
ollama run my-qwen
Choosing the right quantization
# Default (Q4_K_M) โ recommended
ollama run qwen3:8b
# Higher quality if you have VRAM
ollama pull qwen3:8b-q5_K_M
# Maximum quality
ollama pull qwen3:8b-q8_0
# Lower quality for tight VRAM
ollama pull qwen3:8b-q3_K_M
Using with Open WebUI
If you’re running Open WebUI, Qwen models work out of the box. Pull the model in Ollama, and it appears in the model dropdown automatically. Thinking mode tokens are displayed in a collapsible section in most UIs.
Common Problems
Thinking tokens are verbose
Qwen 3’s thinking mode can generate hundreds of tokens of reasoning before answering a simple question. Use /no_think for straightforward queries, or create a Modelfile with PARAMETER think false for a permanently non-thinking version.
Which Qwen should I pick?
If you’re overwhelmed by the lineup:
- Just chatting? Qwen 3 at whatever size fits your GPU
- Coding? Qwen 2.5 Coder 32B (24 GB) or 14B (16 GB)
- Images/documents? Qwen3-VL 8B
- Math/reasoning? Qwen 3 with
/thinkenabled - Non-English? Qwen 3 at any size โ best multilingual support
MoE models running slow
The 30B-A3B MoE has a known GPU utilization issue in Ollama (tracked in GitHub issue #10458). Until this is resolved, dense models (8B, 14B, 32B) will give you better real-world performance despite theoretically lower parameter counts.
Context length limits
Qwen 3 supports long contexts (32K+ natively, 1M with Qwen3-2507 update), but your actual usable context depends on VRAM. Each token in the KV cache costs memory. At Q4_K_M:
- 8B with 8K context: ~6 GB total
- 14B with 8K context: ~12 GB total
- 32B with 4K context: ~21 GB total
If you’re hitting out-of-memory errors, reduce num_ctx in your Modelfile before dropping quantization quality.
Qwen 2.5 vs Qwen 3 โ when to keep the older model
Qwen 3 supersedes Qwen 2.5 for general use. The exceptions:
- Qwen 2.5 Coder: Still the best for pure code tasks
- Qwen 2.5 72B: If you have the VRAM for it, the 72B hasn’t been fully replaced until you step up to the 235B MoE
- Existing fine-tunes: Community fine-tunes built on Qwen 2.5 haven’t all been ported to Qwen 3 yet
For everything else, Qwen 3 is the upgrade.
Bottom Line
Qwen 3 is the model family to beat in 2026. At every size that matters for consumer hardware โ 8B, 14B, and 32B โ it outperforms the competition while fitting the same VRAM budget. The hybrid thinking mode is genuinely useful, not a gimmick. The coding and vision variants fill the gaps.
If you’re still running Llama 3.1 8B as your default small model, try Qwen 3 8B. The difference is immediately obvious. If you have 24 GB and you’re not running Qwen 3 32B, you’re leaving performance on the table.
The Qwen ecosystem is catching up on fine-tunes and community tooling โ Llama still has the edge there. But for base model quality out of the box, Qwen is where the benchmarks point, and where the experience confirms it.
# Just try it
ollama run qwen3:8b