By Wang Chen and Liu Jun
During 2025 Spring Festival, new versions of top-tier open-source Chinese AI models were launched. They showcase performance on par with OpenAI's proprietary models.
• On January 20, DeepSeek-R1 was launched, demonstrating performance comparable to OpenAI o1 in maths, coding, and reasoning tasks [1].
• On January 27, Qwen2.5-1M was introduced, supporting a 1 million token context window. Qwen2.5-1M 14B achieves performance similar to GPT-4o-mini in short-context tasks while offering a context window eight times larger. In long-context tasks, Qwen2.5-1M 14B consistently outperforms GPT-4o mini across multiple benchmark datasets [2].
• On January 27, DeepSeek unveiled Janus-Pro, a unified, multimodal understanding and generation model. In benchmark evaluations, Janus-Pro-7B surpassed both OpenAI's DALL-E 3 and Stability AI's Stable Diffusion in GenEval and DPG-Bench [3].
• On January 28, Qwen2.5-VL was released. This flagship vision-language model surpasses GPT-4o in document understanding, visual question answering, video understanding, and visual agent [4].
• On January 29, Qwen2.5-Max was launched, outperforming DeepSeek-V3 and GPT-4o in benchmark assessments like Arena-Hard, LiveBench, LiveCodeBench, and GPQA-Diamond [5].
There's a growing consensus in this industry that open-source models have evolved from simply following their closed-source counterparts to taking the lead in AI development. DeepSeek and Qwen are the frontrunners of current open-source projects. This article will walk you through the process of deploying an AI model, building a test application, and leverage the model's capabilities to perform tasks.
• Free-of-charge computing: The computing work is done on your device.
• Free-of-charge API calls: API requests are made in your local network.
• Securing sensitive data: Sensitive data is stored locally, which makes local deployment an ideal option for personal developers.
• Device limitations: A local device only supports a quantized or distilled version of DeepSeek-R1, not the full version with 671 billion parameters. To deploy DeepSeek-R1 FP4, a device requires at least 350 GB of memory.
• Accessibility: Released under the MIT license, DeepSeek-R1 supports free distillation. Additionally, DeepSeek-R1 distilled Qwen models are available for download and use.
This section will guide you in deploying a DeepSeek model on your local device and creating an application in Spring AI Alibaba to utilize DeepSeek:
The page shows that Ollama supports deploying DeepSeek-R1 locally:
You can click DeepSeek-R1 to see the detailed introduction:
2. Click Download and install Ollama. After installation, press Command + Space to open your Terminal, and run the following command:
Terminal window
# Install DeepSeek-R1-Distill-Qwen-1.5B.
ollama run deepseek-r1:1.5b
Terminal window
# Install Ollma
curl -fsSL https://ollama.com/install.sh | sh
# Install DeepSeek-R1-Distill-Qwen-1.5B.
ollama run deepseek-r1:1.5b
3. Choose a model size appropriate for your device.
DeepSeek-R1 comes in these sizes: 1.5B, 7B, 8B, 14B, 32B, 70B, and 671B. In this guide, the 1.5B model is chosen for demo purposes. Generally speaking, deploying the 8B model requires 8 GB of memory, and the 32B model requires 24 GB of memory.
Create an application with Spring AI Alibaba, and call the local model
For the full code, see https://github.com/springaialibaba/spring-ai-alibaba-examples/tree/main/spring-ai-alibaba-chat-example/ollama-deepseek-chat
Download the sample code:
Terminal window
git clone https://github.com/springaialibaba/spring-ai-alibaba-examples.git
cd spring-ai-alibaba-examples/spring-ai-alibaba-chat-example/ollama-deepseek-chat/ollama-deepseek-chat-client
Terminal window
./mvnw compile exec:java -Dexec.mainClass="com.alibaba.cloud.ai.example.chat.deepseek.OllamaChatClientApplication"
Enter http://localhost:10006/ollama/chat-client/simple/chat
in your browser to use DeepSeek locally.
When using Spring AI Alibaba to build an application, you need to add the spring-ai-alibaba-starter dependency and inject the ChatClient bean, which are not required when you use Spring Boot.
Add the spring-ai-alibaba-starter dependency. You also need to add the spring-ai-ollama-spring-boot-starter dependency to run the DeepSeek model with Ollama:
<dependency>
<groupId>org.springframework.ai</groupId>
<artifactId>spring-ai-ollama-spring-boot-starter</artifactId>
<version>1.0.0-M5</version>
</dependency>
Configure an address for the model. Specify the base URL and name of the model in application.properties:
spring.ai.ollama.base-url=http://localhost:11434
spring.ai.ollama.chat.model=deepseek-r1
Inject ChatClient:
@RestController
public class ChatController {
private final ChatClient chatClient;
public ChatController(ChatClient.Builder builder) {
this.chatClient = builder.build();
}
@GetMapping("/chat")
public String chat(String input) {
return this.chatClient.prompt()
.user(input)
.call()
.content();
}
}
Spring AI Alibaba DingTalk Group ID: 105120009405
[1] https://github.com/deepseek-ai/DeepSeek-R1
[2] https://qwenlm.github.io/blog/qwen2.5-1m
[3] https://github.com/deepseek-ai/Janus?tab=readme-ov-file
[4] https://qwenlm.github.io/blog/qwen2.5-vl/
[5] https://qwenlm.github.io/zh/blog/qwen2.5-max/
546 posts | 52 followers
FollowAlibaba Cloud Native Community - March 10, 2025
Alibaba Cloud Native Community - March 6, 2025
Alibaba Cloud Native Community - February 20, 2025
Alibaba Cloud Native Community - February 26, 2025
Alibaba Cloud Native Community - March 20, 2025
Alibaba Cloud Native Community - April 15, 2025
546 posts | 52 followers
FollowAccelerate AI-driven business and AI model training and inference with Alibaba Cloud GPU technology
Learn MoreTop-performance foundation models from Alibaba Cloud
Learn MoreOffline SDKs for visual production, such as image segmentation, video segmentation, and character recognition, based on deep learning technologies developed by Alibaba Cloud.
Learn MoreSelf-service network O&M service that features network status visualization and intelligent diagnostics capabilities
Learn MoreMore Posts by Alibaba Cloud Native Community