DeepSeek V3.2

DeepSeek-V3.2 is a large language model designed to harmonize high computational efficiency with strong reasoning and agentic tool-use performance. It introduces DeepSeek Sparse Attention (DSA), a fine-grained sparse attention mechanism that reduces training and inference cost while preserving quality in long-context scenarios. A scalable reinforcement learning post-training framework further improves reasoning, with reported performance in the GPT-5 class, and the model has demonstrated gold-medal results on the 2025 IMO and IOI. V3.2 also uses a large-scale agentic task synthesis pipeline to better integrate reasoning into tool-use settings, boosting compliance and generalization in interactive environments.

Users can control the reasoning behaviour with the reasoning enabled boolean. Learn more in our docs

Share

Model details

Context window163,840 tokens
Max completion size51 tokens
Prompt cost / 1K tokens$0.00000028
Completion cost / 1K tokens$0.0000004
Accepts
Produces

Benchmark performance

Overall

60
score
25th
placement

Cost

98
score
3rd
placement

Logic

80
score
7th
placement

Speed

58
score
30th
placement

Scoring

20
score
20th
placement

Tool Use

19
score
9th
placement

Hallucination

97
score
2nd
placement

Classification

29
score
3rd
placement

Structured Output

33
score
8th
placement

Pricing

Usage pricing
Prompt
$0.00000028
Completion
$0.0000004
Request
FREE
Image
FREE
Web Search
FREE
Internal Reasoning
FREE
Input Cache Read
FREE
Input Cache Write
FREE

Best Overall scoring LLMs

xAI

Grok 4 Fast

88
score
1st
placement
Qwen

Qwen3 VL 235B A22B Instruct

86
score
2nd
placement
xAI

Grok 4.1 Fast

84
score
3rd
placement
OpenAI

GPT-5.1 Chat

82
score
4th
placement
OpenAI

GPT-5.1-Codex

82
score
4th
placement
Anthropic

Claude Haiku 4.5

80
score
5th
placement