China AI Model War 2026: Doubao, Qwen, DeepSeek vs GPT-5
Quick summary
ByteDance's Doubao has 155M weekly users. Alibaba's Qwen 3.5 beats GPT-5 on maths benchmarks. Chinese open-source models now make up 30% of global AI usage. Here's what developers need to know.
ByteDance released Doubao 2.0 on February 14, 2026, with 155 million weekly active users. Alibaba's Qwen 3.5 397B model beats GPT-5.2 on GPQA Diamond benchmarks. Chinese open-source models now account for 30% of global open-model usage on Hugging Face, up from 1.2% in late 2024.
What Is Happening in Chinese AI Right Now
Four major Chinese AI labs are releasing models simultaneously, each targeting a different angle of the market: ByteDance on consumer scale, Alibaba on open weights, DeepSeek on price floor, and Moonshot on agentic capabilities. The result is the most competitive AI model market outside the US.
All four are now genuinely competitive with Western frontier models on benchmarks, and all four are significantly cheaper per token through their APIs.
ByteDance Doubao 2.0: 155 Million Users, Agent Era
Doubao 2.0 (internally called Doubao-Seed-2.0) launched February 14, 2026. At launch, ByteDance reported 155 million weekly active users — making it one of the most-used AI applications in the world by user count. The framing at launch was "the agent era": Doubao 2.0 can decompose complex objectives into sub-tasks and execute them across tools autonomously.
Benchmark scores for Doubao-Seed-2.0:
- AIME 2025: 98.3
- Codeforces rating: 3020
- LiveCodeBench: 87.8
- SWE-Bench Verified: 76.5
ByteDance internally benchmarks Doubao-Seed-2.0 at parity with GPT-5.2 and Gemini 3 Pro on math, coding, and logical reasoning.
Pricing: Doubao 2.0 Pro costs $0.47 per million input tokens and $2.37 per million output tokens — roughly 3.7x cheaper on input and 5.9x cheaper on output than GPT-5.2. Lighter variants are cheaper: Doubao 1.5 Lite-32k runs at approximately $0.042 per million input tokens, about 70% below industry average.
Developer access caveat: Official API access to Doubao 2.0 requires a Chinese phone number for registration. International developers can access it through Volcano Engine (ByteDance's B2B cloud platform) via enterprise negotiation, or through third-party API aggregators like Novita AI.
Alibaba Qwen 3.5: Open Weights, Apache 2.0
Qwen 3.5 is the most developer-friendly Chinese model family. The entire range — from 0.8B to 397B parameters — is open-weight under Apache 2.0 license on Hugging Face. You can download and run them locally, fine-tune them, and use them commercially without restrictions.
The flagship Qwen3.5-397B-A17B is a mixture-of-experts model: 397B total parameters, 17B active per token. It has a 262,144-token native context window, extensible to over 1 million tokens, with native vision support and 201 language coverage.
Benchmark comparisons:
| Benchmark | Qwen3.5-397B | GPT-4o | Notes |
|---|---|---|---|
| MMLU-Pro | 82.5 | 88.7 | GPT-4o leads |
| GPQA Diamond | 88.4 | — | Best on public leaderboard |
| MathVision | 88.6 | — | Beats GPT-5.2 (83.0) |
| IFEval | 92.6 | — | Strong instruction following |
The 9B small model scores 66.1 on BFCL-V4 and 79.1 on TAU2-Bench agentic tasks — competitive with models many times its size, which matters for on-device and edge deployment.
After DeepSeek triggered a price war across Chinese AI labs, Alibaba cut Qwen API pricing from $1.10 to $0.07 per million tokens. That is a 93% price cut. The Qwen family has overtaken Meta's Llama as the most downloaded model series on Hugging Face.
DeepSeek: V3.2 is Out, R2 is Still Coming
The headline that dominated tech media in early 2026 was DeepSeek R2 — but R2 has not shipped. The delay has two causes: training instability on Huawei Ascend chips, and CEO Liang Wenfeng being dissatisfied with performance. DeepSeek returned to Nvidia hardware for critical training runs. As of early March 2026, an R2 release remains unconfirmed but is reportedly possible in March 2026.
What most people missed: DeepSeek V3.2 already shipped in December 2025 and is exceptional. The V3.2-Speciale variant achieved gold-medal performance at IMO 2025, IOI 2025, ICPC World Final 2025, and CMO 2025 simultaneously. It outperforms GPT-5 on reasoning tasks and matches Gemini 3.0 Pro. It is available as open weights on Hugging Face under MIT license.
DeepSeek API pricing remains the cheapest in the market: approximately $0.14 per million input tokens — 4x cheaper than Kimi K2 Thinking and 9x cheaper than Qwen3-Max. This is the price floor that forced every other Chinese lab to cut their prices.
Moonshot Kimi K2.5: 1 Trillion Parameters, 100 Agents
Moonshot AI released Kimi K2.5 on January 27, 2026. It is a 1-trillion parameter mixture-of-experts model with approximately 32B activated parameters per token. The standout feature is its agent swarm capability: K2.5 can direct up to 100 AI sub-agents in parallel, each using tools independently. Moonshot claims this reduces execution time for large-scale research tasks by up to 4.5x.
K2.5 is available as open weights on Hugging Face (moonshotai/Kimi-K2.5) and via API at platform.moonshot.ai with international access. Moonshot also launched Kimi Claw in February 2026 — a persistent browser-based agent environment with 5,000+ community skills and 40GB cloud storage, positioned as a developer platform for autonomous agent tasks.
The 30% Open-Source Market Share Shift
Chinese open-source models went from 1.2% of global open-model downloads in late 2024 to approximately 30% by early 2026. DeepSeek and Qwen drove the majority of this shift.
This has a direct impact on the self-hosted AI market. Developers running local models for privacy, cost, or latency reasons now have a genuine choice between Meta's Llama family and the Chinese alternatives. Qwen 3.5 9B running on a consumer GPU competes with models that required datacenter hardware eighteen months ago.
Which Chinese Models Can Developers Outside China Access?
| Model | Open Weights | International API | Notes |
|---|---|---|---|
| DeepSeek V3.2 | Yes (MIT) | Yes — api.deepseek.com | Cheapest option |
| Qwen 3.5 (all sizes) | Yes (Apache 2.0) | Yes — Alibaba Cloud International | Most permissive licence |
| Kimi K2.5 | Yes (HuggingFace) | Yes — platform.moonshot.ai | Best for agent tasks |
| Doubao 2.0 | No | Restricted — needs +86 phone | Use Novita AI as workaround |
For most developers outside China, DeepSeek V3.2 and Qwen 3.5 are the practical choices. Both have open weights, permissive licences, and working international APIs.
Key Takeaways
- 155 million weekly active users — Doubao 2.0 at launch, February 14, 2026
- 30% of global open-model downloads — Chinese models' Hugging Face share by early 2026, up from 1.2% in 2024
- $0.07 per million tokens — Qwen API price after DeepSeek triggered a 93% price cut
- 88.4 GPQA Diamond — Qwen3.5-397B, best score on the public benchmark leaderboard
- DeepSeek R2 still unshipped — V3.2 (already out, MIT licence) is the better story right now
- For developers: Qwen 3.5 (Apache 2.0, runs locally) and DeepSeek V3.2 (MIT, cheapest API) are the two Chinese models with no access restrictions and genuine frontier performance. Evaluate both before defaulting to GPT-4o or Claude for cost-sensitive workloads.
- What to watch: DeepSeek R2 launch — if it ships in March 2026 with the rumoured capability improvement, it will reset the price-performance benchmark again
Free Weekly Briefing
The AI & Dev Briefing
One honest email a week — what actually matters in AI and software engineering. No noise, no sponsored content. Read by developers across 30+ countries.
No spam. Unsubscribe anytime.
More on China
All posts →Huawei Ascend 910C: China Plans 600,000 AI Chips in 2026
Huawei plans to produce 600,000 Ascend 910C AI chips in 2026, nearly doubling 2025 output. China's AI companies are training models on a hardware stack completely separate from Nvidia and CUDA.
DeepSeek R2 Is Out: What Every Developer Needs to Know Right Now
DeepSeek R2 just dropped. It is multimodal, covers 100+ languages, and was trained on Nvidia Blackwell chips despite US export controls. Here is what changed from R1, what the benchmarks mean, and how to use it including running it locally.
India vs China AI Race 2026: Who's Winning? Humanoid Robots, Summits, and the Real Numbers
India hosted the world's largest AI summit; China's humanoid robots performed in front of a billion viewers. Both say they're winning the AI race. Here's the honest breakdown — India vs China AI 2026.
WAIC 2026 Shanghai: China's World Artificial Intelligence Conference — What to Expect
WAIC 2026 Shanghai (July): the World Artificial Intelligence Conference returns. What happened at WAIC 2025 — DeepSeek, Huawei CloudMatrix, 800+ companies — and what to expect from China's biggest AI event in 2026.
Free Tool
Will AI replace your job?
4 questions. Get a personalised developer risk score based on your stack, role, and what you actually build day to day.
Check Your AI Risk Score →Written by
Abhishek Gautam
Full Stack Developer & Software Engineer based in Delhi, India. Building web applications and SaaS products with React, Next.js, Node.js, and TypeScript. 8+ projects deployed across 7+ countries.