Junyang Lin
3.1K posts






🚀 Introducing the Qwen 3.5 Small Model Series Qwen3.5-0.8B · Qwen3.5-2B · Qwen3.5-4B · Qwen3.5-9B ✨ More intelligence, less compute. These small models are built on the same Qwen3.5 foundation — native multimodal, improved architecture, scaled RL: • 0.8B / 2B → tiny, fast, great for edge device • 4B → a surprisingly strong multimodal base for lightweight agents • 9B → compact, but already closing the gap with much larger models And yes — we’re also releasing the Base models as well. We hope this better supports research, experimentation, and real-world industrial innovation. Hugging Face: huggingface.co/collections/Qw… ModelScope: modelscope.cn/collections/Qw…







Qwen3.5 is now updated with improved tool-calling & coding performance! Run Qwen3.5-35B-A3B on 22GB RAM. See improvements via Claude Code, Codex. We also benchmarked GGUFs & removed MXFP4 layers from 3 quants. GGUFs: huggingface.co/unsloth/Qwen3.… Analysis: unsloth.ai/docs/models/qw…

✨ Qwen3.5 — new from @Alibaba_Qwen — introduces a frontier‑class VLM built for native multimodal agents. With a ~400B‑parameter architecture combining MoE and Gated Delta Networks, Qwen3.5 can reason across text, code, and vision — and even understand and navigate user interfaces. Learn how to: ✅ Run Qwen3.5 on free NVIDIA GPU endpoints ✅ Deploy with NIM ✅ Fine‑tune using NVIDIA NeMo See the details in our technical blog ➡️ developer.nvidia.com/blog/develop-n…





