Back to stories
Models

Alibaba's Qwen3.5 Arrives: 397B Parameters Built for Native Multimodal Agents

Michael Ouroumis2 min read
Alibaba's Qwen3.5 Arrives: 397B Parameters Built for Native Multimodal Agents

Alibaba's Qwen team has released Qwen3.5, a 397-billion-parameter mixture-of-experts model that activates only 17 billion parameters per inference pass. The model represents a major step toward what the team calls "native multimodal agents" — AI systems that process text, images, code, and tool calls within a single unified architecture.

Architecture

Qwen3.5 uses a sparse mixture-of-experts (MoE) design where specialized sub-networks activate depending on the input type and task. This approach delivers the reasoning capacity of a much larger dense model while keeping inference costs manageable.

Key specifications:

Agent-First Design

What distinguishes Qwen3.5 from previous releases is its explicit focus on agentic workflows. The model includes built-in capabilities for:

In Alibaba's internal benchmarks, Qwen3.5 outperforms previous Qwen models and several competing open-weight models on agentic task completion, coding, and multilingual reasoning.

Open-Weight Release

Qwen3.5 is released as open-weight under Alibaba's standard license, which permits commercial use with some restrictions for very large-scale deployments. Smaller distilled variants are expected in the coming weeks to support deployment on consumer hardware.

The 201-language coverage is particularly notable. While most frontier models focus on English and a handful of other high-resource languages — Cohere's TinyAya being a notable exception with its 67-language support, Qwen3.5 aims to serve the global developer community — a strategic advantage for Alibaba's cloud business in markets across Asia, Africa, and Latin America.

Competitive Context

The release intensifies competition in the open-weight model space. Meta's Llama, Mistral's models, and Zhipu AI's recently released GLM-5 are all vying for developer adoption. Qwen3.5's combination of scale, efficiency, multilingual breadth, and agent-native design gives it a distinctive position in this increasingly crowded field.

Weights and documentation are available on Hugging Face and ModelScope.

More in Models

Microsoft Releases Phi-4-Reasoning-Vision-15B: A Small Model That Knows When to Think
Models

Microsoft Releases Phi-4-Reasoning-Vision-15B: A Small Model That Knows When to Think

Microsoft open-sources Phi-4-reasoning-vision-15B, a compact 15B-parameter multimodal model that selectively activates chain-of-thought reasoning and rivals models many times its size.

8 hours ago2 min read
Anthropic Releases Claude Opus 4.6 — Its Most Capable Agentic Coding Model
Models

Anthropic Releases Claude Opus 4.6 — Its Most Capable Agentic Coding Model

Anthropic launches Claude Opus 4.6, a frontier model purpose-built for autonomous coding agents that can plan, execute, and debug multi-file projects with minimal human oversight.

1 day ago2 min read
Meta Releases Llama 4 Maverick With 400B Parameters Under Open Weights
Models

Meta Releases Llama 4 Maverick With 400B Parameters Under Open Weights

Meta releases Llama 4 Maverick, a 400-billion parameter mixture-of-experts model under its open weights license, matching GPT-5 on key benchmarks and reigniting the open-source AI debate.

1 day ago2 min read