Some Coding Plan models, such as qwen3.5-plus and kimi-k2.5, support image understanding natively. But for text-only models like glm-5 and MiniMax-M2.5, add a local skill for visual capabilities.
Image understanding skills consume Coding Plan quota. No additional charges apply.
Prerequisites
You have subscribed to Coding Plan. See Getting started.
You have set up Coding Plan and can use it normally. See Set up AI tools.
Visual support status
Model | Visual support | Description |
| Yes | No configuration needed. Pass images directly. |
| No | Requires a skill or agent for visual capabilities. |
Method 1: Use a visual model directly (recommended)
qwen3.5-plus and kimi-k2.5 support image understanding natively. If you frequently work with images, switch to these models.
Tool | How to switch |
Claude Code |
|
OpenCode |
|
Qwen Code |
|
For other tools, see Set up AI tools. After switching, reference image paths directly or drag and drop images in conversations.
Method 2: Add visual capabilities using a skill or agent
For models without visual capabilities (glm-5, MiniMax-M2.5), configure a skill or agent.
Claude Code
Add a skill
Create a
skills/image-analyzerfolder in the.claudedirectory:mkdir -p .claude/skills/image-analyzerCreate a
SKILL.mdfile with the following content:--- name: image-analyzer description: Helps models without visual capabilities understand images. Use this skill when you need to analyze image content, extract information, text, or UI elements from an image, or understand any visual content such as screenshots, charts, or architecture diagrams. Pass the image path to get a description. model: qwen3.5-plus --- qwen3.5-plus has visual understanding capabilities. Use the qwen3.5-plus model directly for image understanding.Folder structure:
.claude/ └── skills/ └── image-analyzer/ └── SKILL.mdGet started
Start Claude Code in your project directory, then run
claude. Switch toglm-5with/model glm-5.Download and alibabacloud.png to your project directory. Then ask:
Load the image-analyzer skill and describe the information displayed in the alibabacloud.png banner.Response:
OpenCode
Add an agent
Create an
agentsfolder in the.opencodedirectory:mkdir -p .opencode/agentsCreate an
image-analyzer.mdfile with the following content:NoteThe model field must use the provider and model name from the OpenCode configuration. For example, based on the OpenCode setup, use
bailian-coding-plan/qwen3.5-plus.--- description: Analyzes images using a vision-capable model. Use this agent when the user needs to understand image content, extract information from screenshots, diagrams, UI mockups, or any visual content. Invoke with @image-analyzer followed by the image path and your question. mode: subagent model: bailian-coding-plan/qwen3.5-plus tools: write: false edit: false --- You have vision capabilities. Analyze the provided image and return a clear, structured description focused on what the user is asking about.Folder structure:
.opencode/ └── agents/ └── image-analyzer.mdGet Started
Start OpenCode in your project directory, then switch to
glm-5.Download alibabacloud.png to the project folder. Use
@to invokeimage-analyzer, then ask:@image-analyzer describe the information displayed in the alibabacloud.png banner.