DeepSeek R1 Distill Qwen 32B
DeepSeek
Code Multilingual Thinking Tool Calls
DeepSeek R1 Distill Qwen 32B is a 32.76-billion-parameter dense transformer from DeepSeek, distilled from the larger R1 reasoning model into a Qwen-based architecture. It excels at chain-of-thought reasoning, code generation, and multilingual tasks with built-in thinking capabilities. Compared to standard 30B-class instruct models, it provides stronger logical and mathematical reasoning. The model supports nine languages and a 128K context window, making it suitable for developers and researchers who need reasoning-focused inference on mid-range GPU setups.
Hardware Configuration
Optional — for precise deployment recommendations
| Quantization | Quality | Size | Fit |
|---|---|---|---|
| FP16 | Full precision | 61.03 GB | — |
| Q8_0 | High | 32.43 GB | — |
| Q6_K | High | 25.04 GB | — |
| Q5_K_M | Medium | 21.66 GB | — |
| Q4_K_M | Medium | 18.49 GB | — |
| Q3_K_M | Low | 14.84 GB | — |
| Q2_K | Low | 11.47 GB | — |
| Q2_K_L | Low | 11.64 GB | — |
Last updated: March 5, 2026