Chinese AI models now hold 61% of global token consumption on OpenRouter at prices up to 20 times lower than US rivals. For enterprise leaders deploying agentic AI at scale, the arithmetic is no longer theoretical.
The token is the new unit of enterprise AI spending. Not licenses, not seats, not annual contracts. Tokens. Every query your applications send to a large language model, every step an AI agent takes in an automated workflow, every line of code your development team generates through an AI assistant: all of it resolves into token counts at the end of the month. And right now, the price of a token varies by a factor of 20 depending on which model you choose.
That is the business reality buried inside a data shift that the AI industry has been tracking since February 2026. OpenRouter, the largest aggregation platform for large language model application programming interfaces with access to more than 400 models from over 60 providers, began reporting that Chinese-built models had crossed 61% of total token consumption. For the first time in any sustained period, US models were displaced from the top three positions simultaneously. MiniMax M2.5 claimed first place. Moonshot AI's Kimi K2.5 took second. Zhipu AI's GLM-5 landed third.
What drove that shift is not some sudden leap in Chinese model quality, though quality has improved. It is pricing, architecture, and a specific use case that has become the dominant driver of token consumption globally: agentic workflows.
The Numbers That Should Be in Your AI Budget Model
The pricing gap is blunt enough to put directly in front of a finance committee. MiniMax M2.5 charges $0.30 per million input tokens and $1.10 per million output tokens. DeepSeek V3.2 runs $0.26 per million input tokens and $0.38 per million output tokens. By comparison, Anthropic's Claude and OpenAI's premium offerings sit at $5 to $15 per million input tokens and $15 to $75 per million output tokens, depending on the specific model tier.
Those are not small differences. TeamDay.ai's March 2026 OpenRouter analysis estimated that a complex coding task at Claude Opus 4 pricing could run $50 to $100. The same task using DeepSeek V3.2 costs approximately $0.50. That is a 100-to-1 ratio. Multiply that across a development team of 50 engineers running AI-assisted coding tools daily, and it becomes a material budget line item within a single quarter.
"If your agent is burning through millions of tokens a day, even a small per-token price difference becomes a significant line item. That's a structural tailwind for Chinese labs, and it only grows as agentic adoption scales."
Will Liang, CEO, Amplify AI Group — via Nikkei Asia
Why Agentic Workflows Turned This Into an Urgent Problem
The token consumption story matters so much right now because of a structural shift in how AI is actually being used. Programming tasks grew from 11% of total OpenRouter token volume to more than 50% throughout 2025. Agentic workflows, in which models autonomously execute multi-step tasks including calling external tools, writing and running code, and iterating on their own outputs, now account for more than half of all output tokens on the platform.
That matters because agentic tasks are token-intensive by design. A single Claude Code session can consume 500,000 or more tokens. Tools like Cursor and Windsurf make dozens of separate model calls per task. When an enterprise moves from a handful of developers experimenting with AI-assisted coding to a deployment where AI agents are running continuously across software development, customer service, and back-office automation, the token bill compounds rapidly. At that scale, the pricing differential between a US frontier model and a Chinese alternative is not a performance tradeoff. It is an infrastructure cost question.
OpenRouter's chief operating officer Chris Clark put the pattern plainly: Chinese open-weight models are "disproportionately heavy in agentic flows run by US firms." This is not adoption by Chinese companies. American enterprises are choosing Chinese models to power their agentic automation, primarily because the economics work.
Where the Cost Advantage Comes From
Two structural factors drive the pricing gap, and both look durable.
The first is architecture. Chinese AI laboratories have widely adopted mixture-of-experts designs, which activate only a fraction of a model's total parameters for any given query. This dramatically reduces the compute required per token without sacrificing output quality on the tasks where these models perform well. The approach was partly born of constraint: US export controls on advanced semiconductors forced Chinese labs to extract more performance from less silicon. That pressure accelerated an architectural discipline that US labs, with easier access to compute, had less urgent reason to pursue.
The second is energy economics. China's government designated what it calls "computing-electricity synergy" a national priority in its 2026 policy planning, explicitly connecting energy policy to AI competitiveness. China's large-scale investment in renewable energy has driven down the effective cost of compute, and that flows directly into lower inference costs per token.
The Risk Side of the Ledger
The cost case is compelling. The risk case is equally serious, and most enterprises are only running one of those analyses. Routing enterprise data through Chinese model providers raises questions that procurement and legal teams are not currently equipped to answer at the speed technology teams are making these choices.
Data residency and sovereignty requirements vary by industry and by jurisdiction. Healthcare organizations operating under the Health Insurance Portability and Accountability Act, financial institutions under banking regulators, and government contractors under federal data handling rules face constraints that make the cost comparison secondary to the compliance question. For organizations without those restrictions, the calculus is different, but the due diligence is not.
There is also the question of consistency. Some of the token volume surge in Chinese models during February 2026 was amplified by promotional free access through developer tools like Kilo Code and Cline. Usage numbers inflated by promotional periods are not the same as sustained commercial adoption. Enterprise procurement requires evidence of reliability, not viral adoption spikes.
Andreessen Horowitz partner Martin Casado estimated that roughly 80% of startups using open-source AI stacks are running Chinese models. Startups optimizing for speed and cost operate under different constraints than enterprises managing regulated data at scale. The same model choice that makes sense for a ten-person development shop may be inappropriate for a Fortune 500 company's customer data pipelines.
What a Deliberate Strategy Looks Like
Technology teams are already making model selection decisions, usually without a policy framework to guide them. The practical response is not to pick a side in a geopolitical debate but to implement a tiered evaluation framework that maps model choice to data sensitivity.
Internal developer tooling that handles only proprietary code logic, with no customer data, sits in a different risk category than a customer-facing AI agent processing personally identifiable information. Background automation tasks that synthesize public data are different from processes that touch financial records. Organizations building agentic workflows now should be segmenting their token spend by data type rather than by model capability alone.
OpenRouter itself offers a practical demonstration of this approach: its auto-router functionality routes queries to cheaper models for simpler tasks and reserves premium models for complex ones. Enterprise technology teams can implement equivalent logic at the application layer, using cost-efficient models where data sensitivity and task complexity permit, and paying the premium where governance requires it.
If your enterprise's agentic AI deployments are consuming tens of millions of tokens per month, and your procurement team has not yet produced a data-sensitivity-tiered model sourcing policy, who in your organization is making the model selection decisions today, and what criteria are they actually using?
Wu, Zijing. "The Rise of China's Hottest New Commodity: AI Tokens." Financial Times, 26 Mar. 2026.
"February Sees Surge in AI Usage: China's AI Call Volume Overtakes US for First Time." National Business Daily / 36Kr, 26 Feb. 2026.
"Chinese AI Models Hit 61% Market Share on OpenRouter." Dataconomy, 25 Feb. 2026.
"Chinese AI Models Capture 61% of Global Token Usage: MiniMax and Moonshot Overtake US Rivals." Abit.ee, 25 Feb. 2026.
"Top AI Models on OpenRouter (March 2026)." TeamDay.ai, Mar. 2026.
"Everything Rises and BYD Capitalizes on Oil Crisis." Nikkei Asia #techAsia Newsletter, 2 Apr. 2026.
DeepSeek V3.2 Model Page. OpenRouter, 2026. openrouter.ai/deepseek/deepseek-v3.2.
