Qwen3.6-27B Outperforms 397B Model at a Fraction of the Size, Runs Locally at 16.8GB
Summary
Qwen's new 27B model stuns the AI world by outperforming a massive 397B model across all major coding benchmarks while running locally on consumer hardware at just 16.8GB, generating complex code at 25 tokens per second.
Key Points
- Qwen releases Qwen3.6-27B, a dense 27B model claiming flagship-level agentic coding performance that surpasses the much larger Qwen3.5-397B-A17B across all major coding benchmarks.
- The new model weighs in at just 55.6GB compared to the previous flagship's 807GB, and a quantized Q4_K_M version from Unsloth runs locally at only 16.8GB using llama.cpp.
- Testing on a local machine shows impressive results, with the quantized model generating complex SVG images at roughly 25 tokens per second, demonstrating strong creative and coding capabilities for its size.