Qwen Releases Sparse Coding MoE

Qwen Releases Sparse Coding MoE

16 April 2026

What happened

Qwen open-sourced Qwen3.6-35B-A3B, a sparse Mixture-of-Experts (MoE) model with 35 billion total parameters and 3 billion active during inference. This model delivers agentic coding performance, surpassing its predecessor Qwen3.5-35B-A3B and rivaling larger dense models like Qwen3.5-27B and Gemma4-31B on benchmarks such as Terminal-Bench 2.0. Qwen3.6-35B-A3B also demonstrates strong multimodal perception, matching or exceeding Claude Sonnet 4.5 on vision-language tasks, including a 92.0 score on RefCOCO. It is available as open weights, via Qwen Studio, and an upcoming Alibaba Cloud API.

Why it matters

Access to advanced agentic coding and multimodal capabilities now requires significantly less compute. Qwen3.6-35B-A3B's sparse MoE architecture, with only 3 billion active parameters, enables deployment on single GPUs while maintaining performance competitive with much larger models. For platform engineers and founders, this reduces infrastructure costs and broadens accessibility for self-hosted agentic workflows. Procurement teams can evaluate this model for cost-efficient integration into existing systems, particularly for coding assistants like OpenClaw.

Source:qwen.ai

AI generated content may differ from the original.

Published on 16 April 2026

Subscribe for Weekly Updates

Stay ahead with our weekly AI and tech briefings, delivered every Tuesday.

Qwen Releases Sparse Coding MoE