PRC vs. Western Open-Source AI Models: What It Means for Agentic AI
Prepared for: U.S. Congressional Staff
Prepared by: EdgeRunner AI
Date: 28 April 2026
The United States maintains a lead in closed-frontier models (GPT-5, Claude 4.6, Gemini 3.0 Pro, Grok 4.2), but China now dominates the open-weight tier—the models most U.S. builders, sovereign-edge deployments, and agentic systems run.
Four key findings:
Bottom line: The U.S. is not losing the overall AI race, but it is losing the open-weight tier that determines control of sovereign edge, contested-environment, and agentic-builder AI. This gap has immediate national-security consequences for military, intelligence, and disconnected operations.

Pace (last 60 days): China released Minimax M2.5/2.7, Kimi 2.5/2.6, Qwen 3.5/3.6, GLM 5.1, DeepSeek v4, Xiaomi MiMo (3 variants), and more. U.S. releases: Google Gemma 4 and IBM Granite 4.1 only.
Chinese labs lead in parameter efficiency, long-context architectures (linear attention, n-gram caches, multi-token prediction), and auxiliary tools (OCR, vision, video). Most equivalent U.S. architectural R&D is now closed source.

Agentic systems—autonomous task completion, multi-step reasoning, tool use, and long-context operation—are the decisive use case.
No one has “won” yet. Remote Labor Index ~4%; most models score below 50% on Berkeley Function Calling Leaderboard v4.
U.S. closed lead is being pierced on agentic benchmarks. Xiaomi MiMo V2.5 Pro (open-weight) now tops Claw-Eval, ahead of GPT-5.4 and other U.S. frontier models.
Real-world usage confirms the shift. On OpenRouter: #1 model is Moonshot Kimi (Chinese); top three providers by volume are all Chinese (StepFun alone served 3.5T tokens). Cursor’s Composer 2 (world’s leading coding app) uses tuned Kimi K2.5.
Cost reality drives adoption. One EdgeRunner AI engineer runs a full agentic workload (~2 gigatokens/month) on consumer GPUs at $0 cost using Chinese open weight models. The Equivalent usage would cost $6,000/month on Claude Sonnet 4.5 for equivalent capability. Agentic workflows drive dramatically higher token usage which alters the cost/benefit baseline.
Ecosystem advantage: China is agentic-first. It ships open multi-agent frameworks (Manus/OpenManus), inference backends, curated agentic datasets, and post-training optimized for autonomous operation—layers the U.S. has not matched at open-source scale.


Many U.S. vendors integrating within the US Govt, DoD, & IC rely on Chinese open-weight models.

Identity contamination is already occurring: Anthropic Claude, Allen AI OLMo, and Liquid AI LFM2 have been observed identifying as Chinese models in benchmarks. Once in pre-training data, biases and vulnerabilities propagate downstream. Cloud-frontier wrappers (e.g., Palantir AIP, Scale Donovan) add value but do not solve air-gapped, contested, or edge requirements.

To close the open-weight gap and secure U.S. agentic sovereignty:
EdgeRunner AI’s Position
EdgeRunner builds air-gapped, on-device LLMs for warfighters using tuned U.S./NATO open weights. We currently choose between less-capable U.S. bases or more-capable Chinese bases that require costly isolation and auditing. A robust U.S. open-weight ecosystem would strengthen every U.S. company, program office, and operator in contested environments.