Model Gallery is a Model as a Service (MaaS) component of PAI that provides state-of-the-art models across large language model (LLM), AI-generated content (AIGC), computer vision (CV), and natural language processing (NLP). This topic describes how to deploy, fine-tune, distill, and build applications with DeepSeek models using Model Gallery.
Available DeepSeek models
Choose the model that best fits your use case and resource availability.
| Model | Description | Resource requirements |
|---|---|---|
| DeepSeek-V3 | Full-size DeepSeek V3 model | Higher compute resources required |
| DeepSeek-R1 | Full-size DeepSeek R1 reasoning model | Higher compute resources required |
| DeepSeek-R1-Distill-Qwen-7B | Smaller distilled model based on Qwen2.5-7B. Recommended for quick practice | Low. Can be deployed with free trial resources |
Before you start
Activate PAI and create a workspace
PAI workspaces centralize management of computing resources, permissions, and AI assets. Activating PAI creates a default workspace. Object Storage Service (OSS) is activated by default for storing code, models, and datasets.
Region and resource specifications
Most Alibaba Cloud services, including PAI workspaces and OSS buckets, are region-specific. Some regions are not interconnected, so select your region carefully.
Key considerations:
-
Resource availability varies by region. If resources are unavailable in one region, check other regions.

-
Billing: PAI offers pay-as-you-go and subscription billing. Pay-as-you-go resources are shared across users, so resource shortages may occur.
-
Restricted specifications: Some resource specifications are restricted to whitelisted users. Consult your sales manager for recommendations.
-
Lingjun resources: PAI also supports Lingjun AI Computing Service resources with high-speed networks for distributed training or deployment. Lingjun resources are restricted to whitelisted users. Contact your sales manager if needed.

(Optional) Create a virtual private cloud (VPC) for distributed training or deployment
Deploy a DeepSeek model
One-click deployment is available for DeepSeek-V3 and DeepSeek-R1 models. For detailed instructions, see One-click deployment of DeepSeek-V3 and DeepSeek-R1 models.
To get started quickly, try DeepSeek-R1-Distill-Qwen-7B. This smaller distilled model has low resource requirements and can be deployed with free trial resources.
Fine-tune and distill a DeepSeek model
Fine-tuning trains the model with your data to improve accuracy for your specific use case.
Distillation transfers knowledge from a larger teacher model to a smaller student model, retaining accuracy while reducing compute and storage costs.
Fine-tuning success depends on dataset quality, hyperparameters, and experimentation. For many use cases, retrieval-augmented generation (RAG) may be simpler and sufficient.
For detailed instructions, see One-click fine-tuning of DeepSeek-R1 distill models.
Build AI applications with LangStudio
Develop applications with LangStudio
PAI LangStudio simplifies enterprise LLM application development with built-in templates for RAG, web search, and other AI application types.
The following tutorials walk through common application patterns using DeepSeek models in LangStudio:
| Application pattern | Tutorial |
|---|---|
| DeepSeek + Knowledge base | Use LangStudio to create a DeepSeek- and RAG-based Q&A application flow for finance and healthcare |
| DeepSeek + Web Search | Use LangStudio and Alibaba Cloud Information Query Service to build a DeepSeek web search application flow |
| DeepSeek + Knowledge base + Web Search | Use LangStudio and DeepSeek to deploy a RAG- and web search-based chatbot |



