Vision for coding models
Models like qwen3.5-plus and kimi-k2.5 support image understanding natively. Text-only models (glm-5, MiniMax-M2.5) need a local skill for visual capabilities.
Native support: qwen3.5-plus, kimi-k2.5 — pass images directly, no configuration needed.
Via skill/agent: qwen3-max-2026-01-23, qwen3-coder-next, qwen3-coder-plus, glm-5, glm-4.7, MiniMax-M2.5 — add a skill or agent for visual capabilities.
Switch to these models if you frequently work with images.
Reference image paths directly or drag images into conversations.
For text-only models (glm-5, MiniMax-M2.5), configure a skill or agent.
Image understanding skills consume Coding Plan quota. No additional charges apply.
Prerequisites
- Subscribe to Coding Plan. See Getting started.
- Set up Coding Plan. See your tool's setup guide.
Visual support status
Native support: qwen3.5-plus, kimi-k2.5 — pass images directly, no configuration needed.
Via skill/agent: qwen3-max-2026-01-23, qwen3-coder-next, qwen3-coder-plus, glm-5, glm-4.7, MiniMax-M2.5 — add a skill or agent for visual capabilities.
Method 1: Use a visual model directly (recommended)
Switch to these models if you frequently work with images.
| Tool | How to switch |
|---|---|
| Claude Code | /model qwen3.5-plus or /model kimi-k2.5 |
| OpenCode | /models then search for and select qwen3.5-plus or kimi-k2.5 |
| Qwen Code | /model then select qwen3.5-plus or kimi-k2.5 |
Qwen Code using OpenAI-compatible API doesn't support image input. For image understanding tasks, use Claude Code or OpenCode instead.
Method 2: Add visual capabilities using a skill or agent
For text-only models (glm-5, MiniMax-M2.5), configure a skill or agent.
- Claude Code
- OpenCode
Add a skill
- Create a
skills/image-analyzerfolder in the.claudedirectory:
- Create
SKILL.md:
- Folder structure:
Getting started
- Start Claude Code in your project directory. Switch to
glm-5with/model glm-5. - Place an image in your project directory, then ask:
Load the image-analyzer skill and describe the information in <your-image>.
FAQ
Why doesn't OpenCode + qwen3.5-plus understand images?
Why doesn't OpenCode + qwen3.5-plus understand images?
Cause: OpenCode doesn't enable visual capabilities by default — declare
modalities in the configuration.Solution: In the OpenCode configuration, add modalities and set input to ["text", "image"]:Replace sk-sp-xxx with your API key.
Why doesn't OpenClaw + qwen3.5-plus understand images?
Why doesn't OpenClaw + qwen3.5-plus understand images?
Cause: OpenClaw determines visual support based on the input field in the configuration.Solution:
- In
~/.openclaw/openclaw.json, ensure the model definition includes"input": ["text", "image"].
- Clear the model cache and restart OpenClaw: