Alibaba Drops 397B Open-Weight AI Model That's Actually Tiny

February 17, 2026

Alibaba Drops 397B Open-Weight AI Model That's Actually Tiny

Published: February 17, 2026 at 12:25 AM

Updated: February 17, 2026 at 12:25 AM

100-word summary

Alibaba just released Qwen3.5-397B-A17B, the first open-weight flagship in its Qwen3.5 family. Despite boasting 397 billion parameters, only 17 billion activate per task—making it wildly efficient. The model uses hybrid sparse Mixture-of-Experts with linear attention for blazing inference speeds and natively handles text, image, and video inputs. It supports 201 languages, offers a million-token context window via Qwen3.5-Plus, and runs 60% cheaper with 8x efficiency versus predecessors. Weights are live on HuggingFace and ModelScope. This could democratize advanced multimodal AI for developers, putting enterprise-grade agentic capabilities into open-source hands at fraction of the cost.

What happened

Alibaba just released Qwen3.5-397B-A17B, the first open-weight flagship in its Qwen3.5 family. Despite boasting 397 billion parameters, only 17 billion activate per task—making it wildly efficient. The model uses hybrid sparse Mixture-of-Experts with linear attention for blazing inference speeds and natively handles text, image, and video inputs. It supports 201 languages, offers a million-token context window via Qwen3.5-Plus, and runs 60% cheaper with 8x efficiency versus predecessors. Weights are live on HuggingFace and ModelScope.

Why it matters

This could democratize advanced multimodal AI for developers, putting enterprise-grade agentic capabilities into open-source hands at fraction of the cost.

Sources