Back to stories
Models

Alibaba's Qwen3.5 Arrives: 397B Parameters Built for Native Multimodal Agents

Michael Ouroumis2 min read
Alibaba's Qwen3.5 Arrives: 397B Parameters Built for Native Multimodal Agents

Alibaba's Qwen team has released Qwen3.5, a 397-billion-parameter mixture-of-experts model that activates only 17 billion parameters per inference pass. The model represents a major step toward what the team calls "native multimodal agents" — AI systems that process text, images, code, and tool calls within a single unified architecture.

Architecture

Qwen3.5 uses a sparse mixture-of-experts (MoE) design where specialized sub-networks activate depending on the input type and task. This approach delivers the reasoning capacity of a much larger dense model while keeping inference costs manageable.

Key specifications:

Agent-First Design

What distinguishes Qwen3.5 from previous releases is its explicit focus on agentic workflows. The model includes built-in capabilities for:

In Alibaba's internal benchmarks, Qwen3.5 outperforms previous Qwen models and several competing open-weight models on agentic task completion, coding, and multilingual reasoning.

Open-Weight Release

Qwen3.5 is released as open-weight under Alibaba's standard license, which permits commercial use with some restrictions for very large-scale deployments. Smaller distilled variants are expected in the coming weeks to support deployment on consumer hardware.

The 201-language coverage is particularly notable. While most frontier models focus on English and a handful of other high-resource languages — Cohere's TinyAya being a notable exception with its 67-language support, Qwen3.5 aims to serve the global developer community — a strategic advantage for Alibaba's cloud business in markets across Asia, Africa, and Latin America.

Competitive Context

The release intensifies competition in the open-weight model space. Meta's Llama, Mistral's models, and Zhipu AI's recently released GLM-5 are all vying for developer adoption. Qwen3.5's combination of scale, efficiency, multilingual breadth, and agent-native design gives it a distinctive position in this increasingly crowded field.

Weights and documentation are available on Hugging Face and ModelScope.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

Moonshot Kimi K2.6 lands open-source, scales to 300 sub-agents and 4,000 coordinated steps
Models

Moonshot Kimi K2.6 lands open-source, scales to 300 sub-agents and 4,000 coordinated steps

Moonshot AI shipped Kimi K2.6 as a generally available open-source model on April 20, posting 58.6 on SWE-Bench Pro — ahead of GPT-5.4 and Claude Opus 4.6 — while scaling agent swarms to 300 sub-agents and 4,000 coordinated steps.

9 hours ago3 min read
OpenAI's 'Spud' Caught Live in API Testing, Polymarket Jumps to 81% for April 23 Launch
Models

OpenAI's 'Spud' Caught Live in API Testing, Polymarket Jumps to 81% for April 23 Launch

API monitors detected OpenAI's next frontier model — codenamed Spud — running in production-scale testing on April 19, sending Polymarket traders to an 81% implied probability of a public launch on April 23.

1 day ago2 min read
OpenAI Launches GPT-Rosalind, Its First Domain-Specific Model Built for Life Sciences
Models

OpenAI Launches GPT-Rosalind, Its First Domain-Specific Model Built for Life Sciences

OpenAI debuts GPT-Rosalind, a specialized AI model for biology, drug discovery, and genomics, with launch partners including Amgen, Moderna, and Los Alamos National Laboratory.

4 days ago2 min read