DeepSeek R1 Distill: The Six Reasoning Models You Can Run Locally
DeepSeek R1 Distill brings R1's reasoning to 1.5B–70B local models. Compare benchmarks, licences and hardware needs — pick the right size today.
Comprehensive guides to every DeepSeek AI model including V3, V3.2, R1, Coder, Math, VL, and more. Compare capabilities, architecture, benchmarks, and real-world performance in one place.
DeepSeek R1 Distill brings R1's reasoning to 1.5B–70B local models. Compare benchmarks, licences and hardware needs — pick the right size today.
DeepSeek Coder V2 hit 90.2% on HumanEval as an open MoE coder. See specs, benchmarks, pricing and how to access it today.
DeepSeek Math 7B hit 51.7% on MATH and pioneered GRPO. See benchmarks, access options and how it compares to V4 — read the full breakdown.
DeepSeek VL is the original 1.3B/7B open vision-language model. See architecture, benchmarks, licensing and how to run it. Read the practitioner guide.
DeepSeek VL2 is an open-weight MoE vision-language model with strong OCR and grounding. Compare variants, benchmarks and access — read the practitioner review.
DeepSeek LLM was the first DeepSeek release: 7B and 67B open-weight models trained on 2T tokens. Read the architecture, benchmarks and verdict.

DeepSeek MoE explained: fine-grained experts, shared experts, and how V4-Pro and V4-Flash use it. Compare specs and pricing — read the full breakdown.
DeepSeek Prover hits 88.9% on MiniF2F for Lean 4 theorem proving. See architecture, benchmarks, access, and limits — read the full breakdown.
DeepSeek V2 introduced MLA and DeepSeekMoE in 2024. See specs, benchmarks, pricing legacy and how it compares to V4 — read the full breakdown.
DeepSeek V2.5 merged Chat and Coder V2 into one model. See specs, benchmarks, pricing context and migration paths — read the full breakdown.