Alibaba's Qwen3.5 is a ~400B parameter open-source vision-language model (VLM) built for native multimodal agents, using a hybrid MoE and Gated Delta Networks architecture with 17B active parameters and a 256K context window extensible to 1M tokens. It supports UI navigation, visual reasoning, coding, and complex search across

3m read time From developer.nvidia.com
Post cover image
Table of contents
Build with NVIDIA endpointsCustomize with NVIDIA NeMoGet started with Qwen3.5

Sort: