Zero-setup AI chat client
Chatbox is a cross-platform desktop AI chat application that supports text, images, and documents. Connect it directly to Qwen Cloud's pay-as-you-go API for cost-effective access to all Qwen models through the OpenAI-compatible endpoint.
Get running in a few minutes:
You should see: qwen3.5-plus provides a detailed explanation of machine learning
Configure Chatbox to use Qwen Cloud:
When adding models, select appropriate capabilities:
"Failed to connect to Custom Provider"
Chatbox conversations can consume many tokens due to:
Quick start
Get running in a few minutes:
Configuration
Basic setup
Configure Chatbox to use Qwen Cloud:
- API endpoint:
https://dashscope-intl.aliyuncs.com/compatible-mode/v1 - Authentication: API key required
- Model selection: Add any Qwen model by entering its ID
Step-by-step configuration
1
Add custom provider
Settings → Model Provider → Add
- Name:
QwenCloud - API Mode:
OpenAI API Compatible
2
Configure API
- API Key: Your API key
- API Host:
https://dashscope-intl.aliyuncs.com/compatible-mode/v1 - Leave API Path empty
3
Add models
In Model section, click New:
- Model ID: Enter model name (e.g.,
qwen3.5-plus) - Select capabilities: Reasoning, Tool use, Vision (as applicable)
4
Optimize chat settings
Chat Settings:
- Max Message Count in Context: 5-10 (for casual chat)
- Temperature: 0.1-0.9 (lower = more focused)
- Top P: ≤1.0
Model capabilities
When adding models, select appropriate capabilities:
| Capability | Use for | Example models |
|---|---|---|
| Reasoning | Thinking mode | qwen3-max, qwq-32b-preview |
| Tool use | Function calling | Most Qwen models |
| Vision | Image understanding | qwen3-vl-plus, qvq-72b-preview |
Limitations
- Does not support: Audio/video files in chat
- Document parsing: Cannot extract images from documents
- Context limits: Multi-turn conversations accumulate tokens quickly
Examples
- Text conversation
- Image understanding
- Document analysis
Troubleshooting
"Failed to connect to Custom Provider"
Solution: Verify API key and endpoint are correct. Check quota in Qwen Cloud."Range of input length should be [1, xxx]"
Solution:"'temperature' must be Float"
- First turn: Input too long, use model with larger context
- Multi-turn: Reduce "Max Message Count in Context" or start new chat
- Switch to 1M context model for long conversations
Solution: Set Temperature to less than 2.0 in Chat Settings"xx is greater than the maximum of 1 - 'top_p'"
Solution: Set Top P to 1.0 or less in Chat SettingsHigh token consumption
Solution:
- Lower "Max Message Count in Context" to 5-10
- Start new chats for unrelated topics
- Use
qwen3.5-flashwith context cache for document Q&A
Cost optimization
Chatbox conversations can consume many tokens due to:
- Multi-turn context accumulation
- Document parsing
- Image processing
- Context management: Set Max Message Count to 5-10
- Model selection: Use
qwen3.5-flashfor routine tasks - Context cache: Models like
qwen-flashsupport caching for repeated content - New chats: Start fresh for unrelated topics
Related resources
- Models: Available models and pricing →
- Vision models: Image understanding guide →
- API docs: OpenAI-compatible reference →
- Context cache: Reduce costs for repeated content →
