Alibaba's Qwen3.5 is a ~400B parameter open-source vision-language model (VLM) built for native multimodal agents, using a hybrid MoE and Gated Delta Networks architecture with 17B active parameters and a 256K context window extensible to 1M tokens. It supports UI navigation, visual reasoning, coding, and complex search across
•3m read time• From developer.nvidia.com
Sort: