Skip to content

DeepSeek R1 Distill Qwen 32B

DeepSeek
Code Multilingual Thinking Tool Calls

DeepSeek R1 Distill Qwen 32B is a 32.76-billion-parameter dense transformer from DeepSeek, distilled from the larger R1 reasoning model into a Qwen-based architecture. It excels at chain-of-thought reasoning, code generation, and multilingual tasks with built-in thinking capabilities. Compared to standard 30B-class instruct models, it provides stronger logical and mathematical reasoning. The model supports nine languages and a 128K context window, making it suitable for developers and researchers who need reasoning-focused inference on mid-range GPU setups.

Hardware Configuration

Optional — for precise deployment recommendations
Quantization Quality Size Fit
FP16 Full precision 61.03 GB
Q8_0 High 32.43 GB
Q6_K High 25.04 GB
Q5_K_M Medium 21.66 GB
Q4_K_M Medium 18.49 GB
Q3_K_M Low 14.84 GB
Q2_K Low 11.47 GB
Q2_K_L Low 11.64 GB
Last updated: March 5, 2026