Alibaba's Qwen team has released Qwen3.5, a 397-billion-parameter mixture-of-experts model that activates only 17 billion parameters per inference pass. The model represents a major step toward what the team calls "native multimodal agents" — AI systems that process text, images, code, and tool calls within a single unified architecture.
Architecture
Qwen3.5 uses a sparse mixture-of-experts (MoE) design where specialized sub-networks activate depending on the input type and task. This approach delivers the reasoning capacity of a much larger dense model while keeping inference costs manageable.
Key specifications:
- 397B total parameters, 17B activated per pass
- 201 language support — the widest language coverage of any open-weight model
- 128K context window with efficient attention scaling
- Native tool use — function calling, code execution, and API interaction built into the architecture
- Multimodal input — text, images, and structured data processed natively
Agent-First Design
What distinguishes Qwen3.5 from previous releases is its explicit focus on agentic workflows. The model includes built-in capabilities for:
- Planning — Breaking complex goals into executable sub-tasks
- Tool selection — Choosing appropriate APIs or code execution paths
- Memory management — Maintaining context across multi-turn agent interactions
- Self-correction — Detecting and recovering from errors during task execution
In Alibaba's internal benchmarks, Qwen3.5 outperforms previous Qwen models and several competing open-weight models on agentic task completion, coding, and multilingual reasoning.
Open-Weight Release
Qwen3.5 is released as open-weight under Alibaba's standard license, which permits commercial use with some restrictions for very large-scale deployments. Smaller distilled variants are expected in the coming weeks to support deployment on consumer hardware.
The 201-language coverage is particularly notable. While most frontier models focus on English and a handful of other high-resource languages — Cohere's TinyAya being a notable exception with its 67-language support, Qwen3.5 aims to serve the global developer community — a strategic advantage for Alibaba's cloud business in markets across Asia, Africa, and Latin America.
Competitive Context
The release intensifies competition in the open-weight model space. Meta's Llama, Mistral's models, and Zhipu AI's recently released GLM-5 are all vying for developer adoption. Qwen3.5's combination of scale, efficiency, multilingual breadth, and agent-native design gives it a distinctive position in this increasingly crowded field.
Weights and documentation are available on Hugging Face and ModelScope.


