DeepSeek Profile picture
Unravel the mystery of AGI with curiosity. Answer the essential question with long-termism.
7 subscribers
Jan 20 β€’ 5 tweets β€’ 3 min read
πŸš€ DeepSeek-R1 is here!

⚑ Performance on par with OpenAI-o1
πŸ“– Fully open-source model & technical report
πŸ† MIT licensed: Distill & commercialize freely!

🌐 Website & API are live now! Try DeepThink at today!

πŸ‹ 1/n chat.deepseek.comImage πŸ”₯ Bonus: Open-Source Distilled Models!

πŸ”¬ Distilled from DeepSeek-R1, 6 small models fully open-sourced
πŸ“ 32B & 70B models on par with OpenAI-o1-mini
🀝 Empowering the open-source community

🌍 Pushing the boundaries of **open AI**!

πŸ‹ 2/n Image
Dec 26, 2024 β€’ 4 tweets β€’ 2 min read
πŸš€ Introducing DeepSeek-V3!

Biggest leap forward yet:
⚑ 60 tokens/second (3x faster than V2!)
πŸ’ͺ Enhanced capabilities
πŸ›  API compatibility intact
🌍 Fully open-source models & papers

πŸ‹ 1/n
Image
πŸŽ‰ What’s new in V3?

🧠 671B MoE parameters
πŸš€ 37B activated parameters
πŸ“š Trained on 14.8T high-quality tokens

πŸ”— Dive deeper here:
Model πŸ‘‰ github.com/deepseek-ai/De…
Paper πŸ‘‰ github.com/deepseek-ai/De…

πŸ‹ 2/n
Sep 6, 2024 β€’ 4 tweets β€’ 2 min read
πŸš€ Exciting news! We’ve officially launched DeepSeek-V2.5 – a powerful combination of DeepSeek-V2-0628 and DeepSeek-Coder-V2-0724! Now, with enhanced writing, instruction-following, and human preference alignment, it’s available on Web and API. Enjoy seamless Function Calling, FIM, and Json Output all-in-one!

Note: Due to significant updates in this version, if performance drops in certain cases, we recommend adjusting the system prompt and temperature settings for the best results!Image DeepSeek-V2.5 outperforms both DeepSeek-V2-0628 and DeepSeek-Coder-V2-0724 on most benchmarks.
Image
Image