An exploration of the new Qwen3.6-27B open weight model, which claims flagship-level agentic coding performance that surpasses previous larger MoE models while being significantly smaller in size. The author tests a quantized version using llama-server and demonstrates its impressive ability to generate complex SVG graphics locally.
Key points:
- Qwen3.6-27B outperforms the older Qwen3.5-397B-A17B on coding benchmarks.
- Dramatic reduction in model size from 807GB to approximately 55.6GB for the base version.
- Successful local execution using a 16.8GB quantized GGUF version via llama.cpp.
- High-quality SVG generation capabilities for complex prompts like a pelican riding a bicycle.