π DeepSeek-V4 Preview is officially live & open-sourced! Welcome to the era of cost-effective 1M context length.
πΉ DeepSeek-V4-Pro: 1.6T total / 49B active params. Performance rivaling the world's top closed-source models.
πΉ DeepSeek-V4-Flash: 284B total / 13B active params. Your fast, efficient, and economical choice.
Try it now at chat.deepseek.com via Expert Mode / Instant Mode. API is updated & available today!
πΉ Enhanced Agentic Capabilities: Open-source SOTA in Agentic Coding benchmarks.
πΉ Rich World Knowledge: Leads all current open models, trailing only Gemini-3.1-Pro.
πΉ World-Class Reasoning: Beats all current open models in Math/STEM/Coding, rivaling top closed-source models.
2/n
DeepSeek-V4-Flash
πΉ Reasoning capabilities closely approach V4-Pro.
πΉ Performs on par with V4-Pro on simple Agent tasks.
πΉ Smaller parameter size, faster response times, and highly cost-effective API pricing.
πΉ Novel Attention: Token-wise compression + DSA (DeepSeek Sparse Attention).
πΉ Peak Efficiency: World-leading long context with drastically reduced compute & memory costs.
πΉ 1M Standard: 1M context is now the default across all official DeepSeek services.
4/n
Dedicated Optimizations for Agent Capabilities
πΉ DeepSeek-V4 is seamlessly integrated with leading AI agents like Claude Code, OpenClaw & OpenCode.
πΉ Already driving our in-house agentic coding at DeepSeek.
The figure below showcases a sample PDF generated by DeepSeek-V4-Pro.
5/n
API is Available Today!
πΉ Keep base_url, just update model to deepseek-v4-pro or deepseek-v4-flash.
πΉ Supports OpenAI ChatCompletions & Anthropic APIs.
πΉ Both models support 1M context & dual modes (Thinking / Non-Thinking): api-docs.deepseek.com/guides/thinkinβ¦
β οΈ Note: deepseek-chat & deepseek-reasoner will be fully retired and inaccessible after Jul 24th, 2026, 15:59 (UTC Time). (Currently routing to deepseek-v4-flash non-thinking/thinking).
6/n
πΉ Amid recent attention, a quick reminder: please rely only on our official accounts for DeepSeek news. Statements from other channels do not reflect our views.
πΉ Thank you for your continued trust. We remain committed to longtermism, advancing steadily toward our ultimate goal of AGI.
7/n
β’ β’ β’
Missing some Tweet in this thread? You can try to
force a refresh
π Launching DeepSeek-V3.2 & DeepSeek-V3.2-Speciale β Reasoning-first models built for agents!
πΉ DeepSeek-V3.2: Official successor to V3.2-Exp. Now live on App, Web & API.
πΉ DeepSeek-V3.2-Speciale: Pushing the boundaries of reasoning capabilities. API-only for now.
πΉ V3.2: Balanced inference vs. length. Your daily driver at GPT-5 level performance.
πΉ V3.2-Speciale: Maxed-out reasoning capabilities. Rivals Gemini-3.0-Pro.
π₯ Gold-Medal Performance: V3.2-Speciale attains gold-level results in IMO, CMO, ICPC World Finals & IOI 2025.
π Note: V3.2-Speciale dominates complex tasks but requires higher token usage. Currently API-only (no tool-use) to support community evaluation & research.
2/n
π€ Thinking in Tool-Use
πΉ Introduces a new massive agent training data synthesis method covering 1,800+ environments & 85k+ complex instructions.
πΉ DeepSeek-V3.2 is our first model to integrate thinking directly into tool-use, and also supports tool-use in both thinking and non-thinking modes.
β¨ Built on V3.1-Terminus, it debuts DeepSeek Sparse Attention(DSA) for faster, more efficient training & inference on long context.
π Now live on App, Web, and API.
π° API prices cut by 50%+!
1/n
β‘οΈ Efficiency Gains
π€ DSA achieves fine-grained sparse attention with minimal impact on output quality β boosting long-context performance & reducing compute cost.
π Benchmarks show V3.2-Exp performs on par with V3.1-Terminus.
2/n
π» API Update
π Lower costs, same access!
π° DeepSeek API prices drop 50%+, effective immediately.
Introducing DeepSeek-V3.1: our first step toward the agent era! π
π§ Hybrid inference: Think & Non-Think β one model, two modes
β‘οΈ Faster thinking: DeepSeek-V3.1-Think reaches answers in less time vs. DeepSeek-R1-0528
π οΈ Stronger agent skills: Post-training boosts tool use and multi-step agent tasks
Try it now β toggle Think/Non-Think via the "DeepThink" button: chat.deepseek.com
1/5
API Update βοΈ
πΉ deepseek-chat β non-thinking mode
πΉ deepseek-reasoner β thinking mode
π§΅ 128K context for both
π Anthropic API format supported: api-docs.deepseek.com/guides/anthropβ¦
β Strict Function Calling supported in Beta API: api-docs.deepseek.com/guides/functioβ¦
π More API resources, smoother API experience
2/5
Tools & Agents Upgrades π§°
π Better results on SWE / Terminal-Bench
π Stronger multi-step reasoning for complex search tasks
β‘οΈ Big gains in thinking efficiency
π¬ Distilled from DeepSeek-R1, 6 small models fully open-sourced
π 32B & 70B models on par with OpenAI-o1-mini
π€ Empowering the open-source community
π Pushing the boundaries of **open AI**!
π 2/n
π License Update!
π DeepSeek-R1 is now MIT licensed for clear open access
π Open for the community to leverage model weights & outputs
π οΈ API outputs can now be used for fine-tuning & distillation
π Exciting news! Weβve officially launched DeepSeek-V2.5 β a powerful combination of DeepSeek-V2-0628 and DeepSeek-Coder-V2-0724! Now, with enhanced writing, instruction-following, and human preference alignment, itβs available on Web and API. Enjoy seamless Function Calling, FIM, and Json Output all-in-one!
Note: Due to significant updates in this version, if performance drops in certain cases, we recommend adjusting the system prompt and temperature settings for the best results!
DeepSeek-V2.5 outperforms both DeepSeek-V2-0628 and DeepSeek-Coder-V2-0724 on most benchmarks.
In our internal Chinese evaluations, DeepSeek-V2.5 shows a significant improvement in win rates against GPT-4o mini and ChatGPT-4o-latest (judged by GPT-4o) compared to DeepSeek-V2-0628.