Overview
Deepseek
Models
22
Cheapest Input
$0.01
Max Context
164K
Top Intelligence
32.1
Microsoft
Models
7
Cheapest Input
$0.05
Max Context
131K
Top Intelligence
33.1
Flagship Model Comparison
DeepSeek V3.2vsWizardLM-2 8x22B
| Metric | DeepSeek V3.2 | WizardLM-2 8x22B |
|---|---|---|
| Input $/1M tokens | $0.26 | $0.62 |
| Output $/1M tokens | $0.38 | $0.62 |
| Context Window | 164K | 66K |
| Intelligence | 32.1 | 33.1 |
| Coding | 34.6 | N/A |
| Math | 59.0 | N/A |
All Models Pricing
Deepseek Models
| Model | Input/1M | Output/1M | Context |
|---|---|---|---|
| DeepSeek V3.2 | $0.26 | $0.38 | 164K |
| DeepSeek V3.2 Speciale | $0.40 | $1.20 | 164K |
| DeepSeek V3.1 Terminus | $0.21 | $0.79 | 164K |
| DeepSeek V3.2 Exp | $0.27 | $0.41 | 164K |
| DeepSeek V3.1 | $0.15 | $0.75 | 33K |
| R1 0528 | $0.45 | $2.15 | 164K |
| R1 | $0.55 | $2.19 | 64K |
| R1 Distill Qwen 32B | $0.29 | $0.29 | 33K |
| DeepSeek V3 0324 | $0.20 | $0.77 | 164K |
| DeepSeek R1 0528 Qwen3 8B | $0.20 | $0.20 | 128K |
| R1 Distill Llama 70B | $0.70 | $0.80 | 131K |
| R1 Distill Qwen 14B | $0.15 | $0.15 | 33K |
| R1 Distill Llama 8B | $0.04 | $0.04 | 33K |
| DeepSeek V3 | $0.01 | $0.03 | 164K |
| DeepSeek Coder 1.3B Base | $0.10 | $0.10 | 16K |
| DeepSeek Coder 7B Base | $0.20 | $0.20 | 4K |
| R1 Distill Qwen 7B | $0.20 | $0.20 | 33K |
| DeepSeek Coder 7B Base v1.5 | $0.20 | $0.20 | 4K |
| DeepSeek Coder 7B Instruct v1.5 | $0.20 | $0.20 | 4K |
| R1 Distill Qwen 1.5B | $0.20 | $0.20 | 131K |
| DeepSeek Prover V2 | $0.50 | $2.18 | 164K |
| DeepSeek Coder 33B Instruct | $0.80 | $0.80 | 16K |
Microsoft Models
| Model | Input/1M | Output/1M | Context |
|---|---|---|---|
| WizardLM-2 8x22B | $0.62 | $0.62 | 66K |
| Phi-3 Medium 128K Instruct | $1.00 | $1.00 | 128K |
| Phi-3 Mini 128K Instruct | $0.10 | $0.10 | 128K |
| Phi 4 | $0.06 | $0.14 | 16K |
| Phi 4 Multimodal Instruct | $0.05 | $0.10 | 131K |
| Phi 4 Reasoning Plus | $0.07 | $0.35 | 32K |
| Phi-3.5 Mini 128K Instruct | $0.10 | $0.10 | 128K |
Benchmark Comparison
Best Scores by Provider
| Benchmark | Deepseek | Microsoft |
|---|---|---|
| Intelligence | 32.1 DeepSeek V3.2 | 33.1 WizardLM-2 8x22B |
| Coding | 37.9 DeepSeek V3.2 Speciale | 11.2 Phi 4 |
| Math | 96.7 DeepSeek V3.2 Speciale | 18.0 Phi 4 |
| MMLU Pro | 86.3 DeepSeek V3.2 Speciale | 71.4 Phi 4 |
| GPQA | 87.1 DeepSeek V3.2 Speciale | 57.5 Phi 4 |
| LiveCodeBench | 89.6 DeepSeek V3.2 Speciale | 23.1 Phi 4 |
| Aider | 74.2 DeepSeek V3.2 Exp | 44.4 WizardLM-2 8x22B |
| AIME | 89.3 R1 0528 | 14.3 Phi 4 |
| BBH | N/A | 48.6 WizardLM-2 8x22B |
Capabilities
| Capability | Deepseek | Microsoft |
|---|---|---|
| Vision | — | ✓ |
| Tool Calls | ✓ (16 models) | ✓ (7 models) |
| Reasoning | ✓ (15 models) | ✓ (2 models) |
| Audio Input | — | — |
| Audio Output | — | — |
| PDF Input | — | — |
| Web Search | — | — |
| Prompt Caching | ✓ (4 models) | — |
| Open Source Models | ✓ (16 models) | ✓ (7 models) |
Model-Level Comparisons
Compare specific models head-to-head:
DeepSeek V3.2vsWizardLM-2 8x22B
DeepSeek V3.2vsPhi-3 Medium 128K Instruct
DeepSeek V3.2vsPhi-3 Mini 128K Instruct
DeepSeek V3.2 SpecialevsWizardLM-2 8x22B
DeepSeek V3.2 SpecialevsPhi-3 Medium 128K Instruct
DeepSeek V3.2 SpecialevsPhi-3 Mini 128K Instruct
DeepSeek V3.1 TerminusvsWizardLM-2 8x22B
DeepSeek V3.1 TerminusvsPhi-3 Medium 128K Instruct
DeepSeek V3.1 TerminusvsPhi-3 Mini 128K Instruct