DeepSeek-R1 represents a significant advancement in language model capabilities, leveraging reinforcement learning to achieve impressive reasoning performance with limited labeled data. This article guides you through the process of deploying and utilizing DeepSeek-R1 on Tencent Cloud's High-Performance Application Service (HAI), enabling you to quickly test and integrate this powerful model into your applications.
DeepSeek-R1 stands out due to its enhanced reasoning capabilities achieved through extensive reinforcement learning during its post-training phase. This approach allows the model to excel in tasks such as:
Its performance in these areas rivals that of OpenAI's o1
model and other leading models.
Tencent Cloud's HAI provides a pre-configured environment for DeepSeek-R1, making it easy to get started without the complexities of setting up your own infrastructure.
For optimal performance, consider the following compute plan recommendations based on the DeepSeek-R1 model size:
Refer to the Compute Package Types documentation for detailed compute plan specifications.
Once the instance creation is complete, you’ll receive a login password via internal messaging. You now have multiple options for interacting with the DeepSeek-R1 model:
OpenWebUI (Recommended):
ChatbotUI:
Terminal Connection (SSH):
ollama run deepseek-r1
JupyterLab:
ollama run deepseek-r1
Internal Links: Explore other HAI guides such as Managing Python Virtual Environments or Building a Stable Diffusion API Service to further enhance you development.
If the default model doesn't meet your requirements, use the following commands to customize the model parameter size:
ollama run deepseek-r1:1.5b
ollama run deepseek-r1:7b
ollama run deepseek-r1:8b
ollama run deepseek-r1:14b
ollama run deepseek-r1:32b
The environment comes pre-installed and active with Ollama serve, which can be used with with REST APIs. Refer to the Ollama API Documentation for specific instructions.
To build a personal knowledge base follow the steps below.
deepseek-r1:7b
or deepseek-r1:1.5b
model ID.ollama pull bge-m3
.bge-m3:latest
model ID.What model parameter sizes are supported?
HAI currently supports 1.5B, 7B, 8B, 14B, and 32B versions of DeepSeek-R1. 70B and 671B versions are coming soon.
What are the port numbers for Ollama/API?
The API port for calling Ollama in HAI is 6399. OpenWebUI uses port 6699, and ChatbotUI uses 6889. For details, please see common ports.
How do I use the model through the API?
Ollama serve is pre-installed and started in the instance environment. This service supports calls through the REST API. Please refer to the Ollama API documentation for specific call methods.
What if the Ollama download speed is slow in mainland China?
Resources in Beijing, Shanghai, and Guangzhou can be accelerated by clicking on the "Acceleration Settings" in the HAI console and enabling academic acceleration. For relevant capabilities please see enable academic acceleration.
What should I do if I get a resource shortage message?
Due to the popularity of DeepSeek, some regions may be out of stock, and instances cannot be created successfully. Payments will be refunded. Try changing regions or check back later.
For any usage issues, join the Tencent Cloud DeepSeek Deployment Exchange Group. Your suggestions and feedback are welcome!