Back to blog
    Why Chinese Labs Now Dominate Open-Source AI
    industryopen-sourceqwendeepseekkimichinaanalysis

    Why Chinese Labs Now Dominate Open-Source AI

    By April 2026, Chinese labs hold the top five open-weight models on aggregate intelligence benchmarks. The pattern isn't an accident — it reflects strategic, structural, and economic differences between US and Chinese AI development that took years to play out.

    EErtas Team·

    If you'd told someone working on AI infrastructure in early 2024 that two years later the top five open-weight models on every major intelligence benchmark would all come from Chinese labs, they would have been skeptical. The dominant open-weight names then were Llama, Mistral, and Falcon — Meta, a French startup, and a UAE research institute. DeepSeek had just released V2; Qwen 2 was emerging; Kimi was still a Chinese-language Moonshot product mostly invisible to the English-speaking AI community.

    By April 2026, the picture has fundamentally changed. The current open-weight leaderboard:

    • DeepSeek V4 Pro (BenchLM 87)
    • Kimi K2.6 (BenchLM 86)
    • MiMo V2.5 Pro (Xiaomi, composite ~86)
    • GLM-5 / 5.1 (Z.ai, BenchLM 83)
    • Qwen 3.5-397B-A17B (Alibaba, BenchLM ~82)

    The top five — and indeed most of the top ten — are now exclusively Chinese-lab releases. Mistral Small 4, Hermes 4, Llama 4, and OpenAI's GPT-OSS round out the top tier of non-Chinese options, but the gap has widened consistently over 2025-2026.

    This isn't an accident or a one-quarter outlier. The pattern reflects strategic, structural, and economic differences between US and Chinese AI development that took years to play out. Understanding why is useful for anyone making model selection decisions in 2026 and beyond.

    The Strategic Choice: Open Weights as Geopolitical Lever

    The most important factor is also the simplest: Chinese AI companies treat open-weight releases as a strategic priority in ways US AI companies generally don't. The reasoning is geopolitical and economic. Chinese labs face structural disadvantages in the closed-API competition — they can't easily serve Western enterprise customers, they face export controls that limit access to frontier hardware, and they don't have the same brand recognition with global enterprise buyers that OpenAI and Anthropic do.

    Open weights flip the leverage. A Chinese lab that releases the top open-weight model becomes the default choice for any team that prefers self-hosted deployment, can't or won't use US-based proprietary APIs, or needs sovereignty over their AI infrastructure. The model gets deployed globally — including by US enterprises evaluating self-hosted alternatives to OpenAI and Anthropic — and the lab accumulates ecosystem influence that translates into long-term market position.

    US AI companies face the opposite calculus. OpenAI, Anthropic, and Google generate enormous revenue from closed APIs and have limited incentive to release competitive open-weight models that would cannibalize that revenue. The exceptions — Meta's Llama line, OpenAI's GPT-OSS release in August 2025 — are notable specifically because they're departures from the dominant US strategy, not the norm.

    This isn't a moral judgment in either direction. Both strategies make sense given the respective competitive positions. But it produces a structural pattern: Chinese labs invest more aggressively in open-weight releases, US labs invest more aggressively in closed APIs, and the open-weight leaderboard reflects that allocation of effort.

    The Economic Pattern: Lower Training Costs

    A second factor is training economics. DeepSeek's V3 paper from late 2024 made waves precisely because the model was reportedly trained for a fraction of what comparable Western models cost — around $5.6M according to DeepSeek, compared to estimates in the hundreds of millions for GPT-4-class training runs. The cost structure differences come from several sources:

    Engineering efficiency. Chinese labs have been forced to optimize hard around constraints that Western labs don't face. Export controls on advanced GPUs mean Chinese labs can't simply throw hardware at problems the way Western frontier labs can. Architectural innovations — DeepSeek's Multi-Head Latent Attention, Qwen's hybrid MoE designs, Kimi's expert routing optimization — are partly responses to hardware constraints that Western labs haven't had to engineer around.

    Training data curation. Chinese labs generally invest heavily in curated synthetic data generation rather than relying primarily on public web data. The cumulative effect is more efficient training: better signal per token, less repetition, higher data quality. Western labs do this too, but the marginal investment in Chinese labs has been higher relative to total training budget.

    Lower compute opportunity costs. A Western frontier lab running a $200M training run is forgoing $200M of compute that could otherwise serve paying API customers. Chinese labs face a different compute economy — domestic Chinese cloud and chip ecosystems, government-supported infrastructure investments, and lower opportunity costs for experimental training runs. This makes aggressive experimentation cheaper in expected-value terms.

    The result is that Chinese labs ship more model releases per unit of investment, iterate faster, and can afford to release intermediate checkpoints that Western labs would treat as too valuable to give away. Qwen, DeepSeek, and Kimi have all maintained release cadences in 2025-2026 that Western labs simply don't match in the open-weight category.

    The Structural Advantage: Domestic Compute Stack

    A third factor is the developing Chinese domestic AI compute stack. Through 2025-2026, Chinese chip development — Huawei Ascend, increasingly competitive domestic alternatives, and a maturing software stack around them — has reached the point where frontier-scale training is feasible without primary reliance on NVIDIA hardware.

    GLM-5, released by Z.ai in February 2026, was reportedly trained on 8× Huawei Ascend H20 chips. This isn't merely a curiosity — it represents the first frontier-scale open-weight model with documented training on non-NVIDIA infrastructure. The implications are real: a Chinese lab that can train and serve frontier models on domestic hardware has a path forward that US-export-control restrictions can't easily disrupt.

    For deployment teams outside China, this matters less directly. But it means the structural risk that Chinese open-weight releases would suddenly stop due to export controls or chip access has diminished. The Chinese open-weight ecosystem in 2026 is more resilient than it was in 2024, and the trajectory is toward more independence from Western compute supply chains, not less.

    The Talent Pattern: Different Career Paths

    The labor market for AI researchers in China is fundamentally different from the US one. In the US, top AI researchers face a competitive market dominated by frontier labs (OpenAI, Anthropic, Google DeepMind) plus an ecosystem of well-funded startups. Salary expectations are extremely high; researchers typically expect either substantial cash compensation or substantial equity upside.

    In China, the market is structured differently. Top AI talent flows through a smaller number of companies (Alibaba, Tencent, ByteDance, the major AI startups like DeepSeek, Moonshot, Z.ai) and the salary ceiling is lower in absolute terms — though purchasing power and lifestyle considerations make the comparison less clean than headline numbers suggest. The result is that Chinese AI labs can field large research teams at total compensation costs that Western frontier labs would find prohibitive for equivalent quality.

    Combined with culturally-different attitudes toward open publication (Chinese AI labs generally publish more aggressively than Western frontier labs, partly for international visibility), this produces a research ecosystem where the marginal cost of open-weight releases is lower than in the US. A Chinese lab releasing a frontier open-weight model is doing so with researchers who, on average, view open publication as a normal expected output of their work — not as a sacrifice of competitive advantage.

    The Implication: A Persistent Pattern

    The natural question is whether this is a temporary lead or a structural pattern. Our read is that it's structural, at least for the next several years. The combination of strategic alignment (open weights as geopolitical lever), economic structure (lower training costs and labor costs), infrastructure independence (developing domestic compute stack), and cultural patterns (more aggressive open publication) creates a sustained advantage that won't reverse in any single product cycle.

    US AI companies are unlikely to suddenly start releasing competitive open-weight models that cannibalize their closed-API revenue. The OpenAI GPT-OSS release in August 2025 was a notable departure but specifically positioned as a sub-frontier release rather than a competitive open-weight flagship. Meta's Llama 4 reception was mixed, and Behemoth has been paused. The structural conditions that produced the current Chinese-lab dominance haven't reversed.

    For teams making infrastructure decisions in 2026, this pattern has practical implications. Building a deployment strategy that depends on US-developed open-weight models being competitive is increasingly risky. The base case for the next 24 months is that Chinese-lab open-weight releases continue to dominate the leaderboard, with periodic competitive responses from Mistral (in Europe), Nous Research (with Hermes-style steerable fine-tunes), and occasional notable US releases like GPT-OSS, but no sustained shift in overall ecosystem position.

    What This Means for Production Deployments

    The practical answer for most teams: get comfortable with Chinese-lab open-weight models as a primary infrastructure choice. The licensing is broadly commercial-permissive (Apache 2.0 or equivalent for most current flagships), the deployment infrastructure (vLLM, Ollama, llama.cpp, TensorRT-LLM) supports them as first-class options, and the model quality is genuinely better than Western alternatives at every parameter scale where leaderboard data is available.

    There are legitimate reasons to prefer non-Chinese alternatives in specific contexts: regulatory constraints (some industries can't use Chinese-developed models for compliance reasons), data sovereignty preferences (European deployments often choose Mistral specifically for EU positioning), or supply-chain diversification strategies that intentionally include non-Chinese options. These are valid considerations and will continue to drive a meaningful slice of deployment decisions toward Mistral, Hermes 4, and similar non-Chinese options.

    But the default — the model you choose when no specific structural reason pushes you elsewhere — should probably be a Chinese-lab open-weight option in 2026. Qwen 3.6 for most production deployments, DeepSeek V4 for multi-GPU server deployments, Kimi K2.6 for long-horizon agentic workloads, MiMo V2.5 Pro for agentic coding. The structural conditions that produced this leadership aren't reversing soon, and aligning your infrastructure to that reality produces better outcomes than fighting it.

    Closing Thought

    There's a tendency in Western AI commentary to frame Chinese-lab dominance as alarming or as evidence of geopolitical concern. We don't see it that way. Open-weight releases — regardless of where they originate — make AI infrastructure available to a broader set of deployment teams than closed APIs alone would, and the resulting ecosystem is healthier and more resilient than one dominated by a few US-based proprietary providers.

    The fact that the strongest open-weight models in 2026 happen to come from Chinese labs is a result of strategic and economic dynamics specific to the current moment. It's not a permanent feature of the ecosystem. But for teams making deployment decisions today, the practical guidance is clear: the best open-weight options are Chinese-lab releases, the licensing supports commercial deployment, and aligning your infrastructure strategy to that reality produces better outcomes than waiting for US labs to close the gap. They probably won't, at least not on the open-weight axis specifically.

    Ship AI that runs on your users' devices.

    Early bird pricing starts at $14.50/mo — locked in for life. Plans for builders and agencies.

    Keep reading