Qwen3 is the kind of release that makes the open-model shelf look less like leftovers and more like a parts catalog. Qwen open-weighted two MoE models, Qwen3-235B-A22B and Qwen3-30B-A3B, plus six dense models from 0.6B through 32B under Apache 2.0.

That spread matters. A single flagship is nice for screenshots. A family of models is useful for routing, local testing, edge deployment, fine-tuning, and all the other unglamorous work that actually ships.

Source credit: Qwen's original source material.

Thinking mode is not always the answer

The most interesting design choice is Qwen3's hybrid thinking mode. The models can run in a deeper step-by-step mode for harder tasks or a faster non-thinking mode for simpler prompts. In less ceremonial language: reasoning becomes a budget you can control.

That is a big deal for anyone building agents. You do not want every calendar lookup treated like a PhD qualifying exam. You do want the option to spend more compute when the task is ambiguous, long, or easy to mess up.

  • two MoE models are open-weighted, including 235B total / 22B active and 30B total / 3B active variants
  • six dense models range from 0.6B to 32B
  • larger Qwen3 models support up to 128K context in the published specs
  • Qwen says the models support 119 languages and dialects

The release also calls out improved coding and agentic capabilities, plus stronger support for MCP. That is where Qwen3 becomes more than a benchmark entry. Open models that understand tool use and agent context are the ones that can move from notebooks into actual workflows.

Qwen recommends deployment through frameworks like SGLang and vLLM, and local usage through Ollama, LM Studio, MLX, llama.cpp, and KTransformers. Translation: the runway from download to experiment is short. Good. It should be.

The snarky version: closed labs keep inventing premium words for 'we burned more tokens.' Qwen3 gives builders a more practical framing: choose the model size, choose the reasoning behavior, and measure whether the result is worth the cost. Very rude to the magic show, honestly.

In short

Qwen3 open-weights a full range of dense and MoE models under Apache 2.0, with hybrid thinking modes that let builders trade speed for deeper reasoning when the task actually deserves it.