DeepSeek-V3.2 Series Open Source, Performance Rivals Gemini-3.0-Pro
š Surprise Release on ChatGPTās Third Anniversary
On the third anniversary of ChatGPT, DeepSeek suddenly unveiled two new AI models:
- DeepSeekāV3.2 ā designed for balanced practicality, ideal for daily Q&A, general Agent tasks, and realāworld tool invocation.
- š§ Reasoning ability: on par with GPTā5, slightly below Geminiā3.0āPro.
- DeepSeekāV3.2āSpeciale ā built for extreme reasoning, matching Geminiā3.0āPro in benchmarks, excelling in advanced mathematical and logical tasks.
- š Won gold medals in IMO 2025, CMO 2025, ICPC World Finals 2025, and IOI 2025, with ICPC second among humans and IOI tenth among humans.

---
DeepSeekāV3.2 ā Balanced, HighāPerformance Agent Model
Key Characteristics
- GPTā5ālevel reasoning
- Shorter outputs than KimiāK2āThinking ā faster response times
- Unified āthinking into tool invocationā ā works in thinking and nonāthinking tool modes
- Trained with largeāscale Agent datasets: 1,800+ environments, 85,000+ complex instructions
Benchmark Highlights
- Outperforms existing openāsource models in Agent evaluations.
- Wins toolāinvocation tests without specific training for those tools.

---
DeepSeekāV3.2āSpeciale ā Extreme Reasoning Edition
- Integrates DeepSeekāMathāV2 theoremāproving
- Best for mathematical proofs, programming competitions, academic research
- Not optimized for casual conversation or writing
- Researchāonly, no tool invocation
- Outperforms V3.2 in complex tasks ā much higher token usage & cost

---
Availability
- App & Web use DeepSeekāV3.2 by default
- Speciale available via temporary API
- Technical report, paper, and benchmarks released publicly
---
š Technical Innovations in V3.2
DeepSeek Sparse Attention (DSA)
- New attention mechanism
- Traditional: O(L²) complexity ā slow for long contexts
- DSA: O(LĀ·k), with k āŖ L
- Faster reasoning for long sequences with no performance drop
- FP8 precision, works with MLA (MultiāQuery Attention)
- Two key components:
- Lightning Indexer ā quickly ranks token relevance
- Fineāgrained token selection

---
Training Process ā V3.1 Terminus to V3.2
Phase 1: Dense Warmāup
- Dense attention maintained
- Train lightning indexer to align with main attention
- 1,000 steps, 2.1āÆB tokens
Phase 2: Introduce Sparse Mechanism
- Each query token selects 2,048 keyāvalue pairs
- 15,000 steps, 943.7āÆB tokens
---
š Performance Gains
On 128āÆk sequences:
| Stage | V3.1āTerminus Cost | V3.2 Cost |
|-------------|--------------------|-----------|
| Prefill | $0.70 / 1M tokens | ~$0.20 |
| Decode | $2.40 / 1M tokens | $0.80 |

---
š” Heavy Investment in RL
- RL compute budget >10% of preātraining ā rare in openāsource LLMs
- Developed stable, scalable RL protocol
- Targeted postātraining improvements for complex tasks

---
RL Scalability Enhancements
- Unbiased KL Estimation ā removed systemic bias, stabilized gradients
- Offline Sequence Masking ā prevents offāpolicy degradation
- Keep Routing for MoE ā keeps training/inference expert paths consistent
---
Expert Distillation Approach
Task Domains:
- Mathematics
- Programming
- Logical reasoning
- Agent general tasks
- Agent programming
- Agent search
Each domain: supports āthinkingā & ānonāthinkingā modes
ā Expert models feed highāquality domain data into final training

---
š§ Breakthrough in Agent Capabilities
- Combines reasoning + tool usage
- Optimized thinking context management:
- Discards reasoning traces only on new user messages, not tool actions
- Toolācall history & results preserved
---
Prompt Engineering for Cold Starts
- Special prompts encourage tool calls during reasoning
- For competitions: clearly separates āthinkingā from final answer
---
RealāWorld Applicability
Platforms like AiToEarnå®ē½ enhance model utility by:
- Multiāplatform publishing (Douyin, Bilibili, LinkedIn, X, etc.)
- AI content creation, analytics, and monetization
- Public AI model rankings (AI樔åęå)
---
ā Automated Environment Synthesis
- 1,827 environments, 85,000 complex prompts
- Example: travel plan under constraints (budget balance, no repeat cities)
- āHard to solve, easy to verifyā ā perfect for RL


---
Code & Search Agent Data
- Code Agent: millions of GitHub issueāPR pairs ā executable environments in Python, Java, JavaScript
- Search Agent: multiāagent pipeline, longātail entity sampling, verified answers
Results:
- SWEāVerified: 73.1%
- Terminal Bench 2.0: 46.4%
- Tool benchmarks: near closedāsource model performance

---
š Limitations
- Lower training FLOPs ā narrower world knowledge than top closedāsource models
- Less tokenāefficient ā longer trajectories needed for equal quality
- Improvement targets set for future releases
---
DeepSeek, when will R2 arrive?! šÆ
---
For researchers, developers, and AI creators, DeepSeekāV3.2 is a blueprint for integrating reasoning, tool usage, and efficiency.
Platforms like AiToEarnå®ē½ let builders:
- Connect AI generation with publishing
- Analyze performance across platforms
- Monetize creativity at scale
Openāsource + multiāplatform reach = realāworld AI impact.