How to evaluate Alibaba’s Qwen3 series of large models?

Silicon Gamer

29/04/2025

updated 13/05/2025

1. Alibaba release Qwen3 Model 

On April 29, 2025, Alibaba unveiled its next-generation AI model, ​Qwen3, which it boldly positions as the new global leader in open-source large language models (LLMs). According to Alibaba Cloud, the flagship ​Qwen3-235B-A22B​ model outperforms top competitors like DeepSeek-R1, OpenAI’s o1/o3-mini, Grok-3, and Google’s Gemini 2.5-Pro across critical benchmarks for coding, mathematics, and general reasoning​.

​I’m used to every company always claiming they’re the world’s best when launching new products,however I’ll try it out myself and see how it actually performs.​​

 

2. What Makes Qwen3 Special?

Two major highlights:

 Hybrid Reasoning Model

  • Inspired by Claude 3.7, Qwen3 introduces adaptive reasoning—switching between reasoning and non-reasoning modes dynamically.

Massive MoE Architecture (235B Parameters)

  • Previous Qwen models maxed at 72B, the bigger the model, the better it works—until you hit the hardware wall.

  • Qwen3-235B-A22B (Mixture of Experts) narrows the gap, marking a significant upgrade.

2.1

 

 

3. Community Reactions

I checked the reactions from users on Reddit, and the discussion is quite lively with positive feedback. Some people mentioned that running Qwen3-30B-A3B MoE with a 4090 GPU is very slow under Ollama, while the speed in LM Studio is as expected. If you encounter the same issue, you can give it a try.

 

4. Download Qwen3 Models

Hugging Face Collection: 🔗 ​https://huggingface.co/collections/Qwen/qwen3-67dd247413f0e2e4f653967f

4.1

5. Experience

After trying it myself, I found that it is indeed an improvement over Qwen 2.5. However, it is still difficult to evaluate exactly how much progress has been made. I will provide more feedback after further use.

 

发表评论