DeepSeek has officially released the first version of its new series model, DeepSeek-V3, and made it open source. This marks a significant milestone in the ongoing evolution of AI technology, offering developers and researchers access to a powerful new tool.
The DeepSeek-V3 model represents a significant leap forward in AI capabilities. You can interact with the latest V3 model by logging into the official website chat.deepseek.com. The API service has been updated simultaneously, with no configuration changes required for the interface. Note that the current version of DeepSeek-V3 does not support multimodal input and output.
One of the most remarkable improvements in DeepSeek-V3 is its generation speed. Through algorithmic and engineering innovations, it has increased from 20 TPS (tokens per second) to 60 TPS, a threefold increase compared to the V2.5 model. This enhancement provides users with a faster and smoother experience.
With the launch of the more powerful and faster DeepSeek-V3, the pricing for the model API service has been updated to:
DeepSeek-V3 is trained using FP8 and open-sources the native FP8 weights. Supported by the open-source community, SGLang and LMDeploy have immediately supported the native FP8 inference of the V3 model. TensorRT-LLM and MindIE have implemented BF16 inference. Additionally, a conversion script from FP8 to BF16 is provided to facilitate community adaptation and expansion of application scenarios. You can download the model weights and find more local deployment information at Hugging Face.
DeepSeek's unwavering commitment to open source and long-termism aims to democratize AGI. They are excited to share their progress in model pre-training with the community, observing the narrowing gap between open-source and closed-source models. DeepSeek plans to continue building richer functionalities such as deep thinking and multimodality on the DeepSeek-V3 base model, and to continuously share its latest explorations with the community.
For further information and engagement, you can join the DeepSeek community through:
Read more about other DeepSeek releases: