Back to stories
Models

Alibaba's Qwen3.5 Arrives: 397B Parameters Built for Native Multimodal Agents

Michael Ouroumis2 min read
Alibaba's Qwen3.5 Arrives: 397B Parameters Built for Native Multimodal Agents

Alibaba's Qwen team has released Qwen3.5, a 397-billion-parameter mixture-of-experts model that activates only 17 billion parameters per inference pass. The model represents a major step toward what the team calls "native multimodal agents" — AI systems that process text, images, code, and tool calls within a single unified architecture.

Architecture

Qwen3.5 uses a sparse mixture-of-experts (MoE) design where specialized sub-networks activate depending on the input type and task. This approach delivers the reasoning capacity of a much larger dense model while keeping inference costs manageable.

Key specifications:

Agent-First Design

What distinguishes Qwen3.5 from previous releases is its explicit focus on agentic workflows. The model includes built-in capabilities for:

In Alibaba's internal benchmarks, Qwen3.5 outperforms previous Qwen models and several competing open-weight models on agentic task completion, coding, and multilingual reasoning.

Open-Weight Release

Qwen3.5 is released as open-weight under Alibaba's standard license, which permits commercial use with some restrictions for very large-scale deployments. Smaller distilled variants are expected in the coming weeks to support deployment on consumer hardware.

The 201-language coverage is particularly notable. While most frontier models focus on English and a handful of other high-resource languages — Cohere's TinyAya being a notable exception with its 67-language support, Qwen3.5 aims to serve the global developer community — a strategic advantage for Alibaba's cloud business in markets across Asia, Africa, and Latin America.

Competitive Context

The release intensifies competition in the open-weight model space. Meta's Llama, Mistral's models, and Zhipu AI's recently released GLM-5 are all vying for developer adoption. Qwen3.5's combination of scale, efficiency, multilingual breadth, and agent-native design gives it a distinctive position in this increasingly crowded field.

Weights and documentation are available on Hugging Face and ModelScope.

Learn AI for Free — FreeAcademy.ai

Take "AI Essentials: Understanding AI in 2026" — a free course with certificate to master the skills behind this story.

More in Models

xAI Launches Grok Voice Think Fast 1.0, Tops τ-Voice Bench and Powers Starlink Support
Models

xAI Launches Grok Voice Think Fast 1.0, Tops τ-Voice Bench and Powers Starlink Support

xAI's new voice model scored 67.3% on the τ-voice Bench — well ahead of Gemini 3.1 Flash Live and GPT Realtime — and is now powering Starlink's phone sales and support with a 70% autonomous resolution rate.

2 days ago2 min read
Tencent Drops Hy3 Preview: 295B Open-Source MoE Model Kicks DeepSeek Out of Yuanbao
Models

Tencent Drops Hy3 Preview: 295B Open-Source MoE Model Kicks DeepSeek Out of Yuanbao

Tencent has open-sourced Hy3 Preview, a 295B/21B-activated mixture-of-experts model built in under three months. The Yuanbao chatbot is switching its primary engine from DeepSeek to the new in-house model.

4 days ago2 min read
DeepSeek V4 Preview Lands: 1.6T-Parameter Open Model With 1M Context, Flash Pricing at $0.14/M
Models

DeepSeek V4 Preview Lands: 1.6T-Parameter Open Model With 1M Context, Flash Pricing at $0.14/M

DeepSeek on April 24 released preview versions of V4-Pro and V4-Flash, an open-weight MoE family with a 1M-token context window and pricing that undercuts Western frontier labs.

4 days ago2 min read