How Do Chinese AI Models Use Tokens to 'Export' Electricity?

比推2026-02-26 tarihinde yayınlandı2026-02-26 tarihinde güncellendi

Özet

The article explores how Chinese AI models are effectively "exporting" electricity globally through token consumption. In 2026, data from OpenRouter shows that Chinese models account for 61% of the top ten models' token usage, with MiniMax M2.5, Kimi K2.5, and Zhipu GLM-5 leading. This shift is driven by the rise of tools like OpenClaw, which exponentially increases token usage, and the significantly lower cost of Chinese models compared to Western alternatives (e.g., MiniMax at $0.3 per million input tokens vs. Claude Opus at $5). The core idea is that token consumption represents the export of computational power and electricity. When a developer overseas sends an API request, it is processed in Chinese data centers using China’s cheaper electricity (about 40% lower than the U.S.), without the power physically leaving the country. This creates a form of invisible, tariff-free "export" of energy. The trend mirrors Bitcoin mining, where China once dominated by converting cheap electricity into digital value, but AI tokens offer more utility and deeper integration into global workflows. However, challenges like data sovereignty and chip restrictions remain. Ultimately, the token economy is becoming a new frontier in U.S.-China tech competition, with Chinese models gaining influence by embedding into developers' systems worldwide.

Author: Black Lobster, Deep Tide TechFlow

Original Title: Token Exports: Selling China's Electricity to the World


In the summer of 1858, a copper-core cable was laid across the Atlantic seabed, connecting London and New York.

The significance of this event was never about transmission speed but about power dynamics. Whoever laid the submarine cables could control the flow of information. The British Empire, with its global telegraph network, held intelligence from colonies, cotton prices, and war news in its grasp.

The empire's strength was not just its fleet but also that cable.

Over 160 years later, this logic is replaying in an unexpected way.

In 2026, Chinese large models are quietly dominating the global developer market. The latest data from OpenRouter shows that Chinese models account for 61% of the token consumption among the top ten models on the platform, with the top three all from China. API requests from developers in San Francisco, Berlin, and Singapore are crossing the Pacific submarine cables to data centers in China, where computing power is consumed, electricity flows, and results are sent back.

The electricity never leaves China's power grid, but its value is delivered cross-border through tokens.

The Great Migration of AI Models

On February 24, 2026, OpenRouter released weekly data: the total token consumption of the top ten models was about 8.7 trillion, with Chinese models accounting for 5.3 trillion, or 61%. MiniMax M2.5 topped the list with 2.45 trillion tokens, followed by Kimi K2.5 and Zhipu GLM-5, all from China.

Latest data as of February 26

This is no coincidence; a fuse has been lit.

Earlier this year, OpenClaw emerged, an open-source tool that allows AI to truly work, directly controlling computers, executing commands, and handling complex workflows in parallel. Its GitHub stars surpassed 210,000 within weeks.

John, a finance professional, installed OpenClaw immediately, connected it to the Anthropic API, and began automatically monitoring stock information and generating trading signals. A few hours later, he stared at his account balance in disbelief: dozens of dollars, gone.

This is the new reality brought by OpenClaw. In the past, chatting with AI consumed a few thousand tokens per conversation, with negligible costs. With OpenClaw, AI runs dozens of subtasks in the background, repeatedly calling context and iterating cycles. Token consumption isn't linear; it's exponential. The bill accelerates like a car with the hood open, the fuel gauge dropping, unstoppable.

A "clever trick" soon spread among developer communities: using OAuth tokens to directly connect Anthropic or Google subscription accounts to OpenClaw, turning the "unlimited" monthly quota into free fuel for AI agents. This became a common method for many developers.

Official countermeasures followed.

On February 19, Anthropic updated its terms, explicitly prohibiting the use of Claude subscription credentials for third-party tools like OpenClaw. To access Claude functionality, users must use the API billing channel. Google went further, widely banning subscription accounts that accessed Antigravity and Gemini AI Ultra via OpenClaw.

"The world has suffered long under tyranny." John immediately switched to domestic large models.

On OpenRouter, the domestic model MiniMax M2.5 scores 80.2% on software engineering tasks, while Claude Opus 4.6 scores 80.8%, a negligible difference. But the prices are worlds apart: the former charges $0.3 per million input tokens, while the latter charges $5, a difference of about 17 times.

John switched over. His workflow continued to run, and his bill shrank by an order of magnitude. This migration is happening globally.

OpenRouter's COO, Chris Clark, put it bluntly: Chinese open-source models are capturing a significant market share because they account for an unusually high proportion of the agent workflows run by U.S. developers.

Electricity Exports

To understand the essence of token exports, one must first understand the cost structure of a token.

It seems light—one token is roughly equivalent to 0.75 English words. An ordinary conversation with AI consumes only a few thousand tokens. But when these tokens stack up in trillions, the underlying physical reality becomes heavy.

Breaking down the cost of a token, there are only two core components: computing power and electricity.

Computing power is the depreciation of GPUs. Buying an NVIDIA H100 costs about $30,000, and its lifespan, converted to each inference, is the depreciation cost. Electricity is the fuel for data centers to operate continuously. A GPU consumes about 700 watts at full load, and with cooling systems, the electricity bill for a large AI data center can easily exceed hundreds of millions of dollars annually.

Now, map this physical process.

A U.S. developer in San Francisco sends an API request. The request travels from California via Pacific submarine cables to a data center in China. GPU clusters start working, electricity flows from China's grid to those chips, inference is completed, and results are sent back. The entire process takes maybe a second or two.

Electricity never leaves China's power grid, but its value, through tokens, completes cross-border delivery.

Here's something miraculous that ordinary trade cannot achieve: tokens have no physical form, do not need to pass through customs, are not subject to tariffs, and are almost invisible in any current trade statistics. China exports a vast amount of computing power and electricity services, but in official commodity trade data, it is nearly invisible.

Tokens have become derivatives of electricity. Token exports are essentially electricity exports.

This is also thanks to China's relatively low electricity prices, which are about 40% lower than those in the U.S. This is a physical cost difference that competitors cannot easily replicate.

Additionally, Chinese AI models have algorithmic and "internal competition" advantages.

DeepSeek V3's MoE architecture activates only part of the parameters during inference, with independent tests showing its inference cost is about 36 times lower than GPT-4o. MiniMax M2.5 similarly activates only 10B of its 229B total parameters.

At the top layer is internal competition. Alibaba, ByteDance, Baidu, Tencent, Moon Dark Side, Zhipu, MiniMax... over a dozen companies are competing on the same track, driving prices below reasonable profit margins. Losing money to gain traction is already the industry norm.

Looking closely, this mirrors China's manufacturing exports: leveraging supply chain advantages and internal competition to drastically lower token prices.

From Bitcoin to Tokens

Before tokens, there was another form of electricity export.

Around 2015, power station managers in Sichuan, Yunnan, and Xinjiang began receiving strange visitors.

These people rented abandoned factories, filled them with dense rows of machines, and kept them running 24/7. The machines produced nothing but continuously solved a mathematical problem, occasionally mining a Bitcoin from this endless calculation.

This was the first-generation form of electricity exports: converting cheap hydropower and wind power, via mining machines' hash computations, into globally circulating digital assets, which were then cashed out into dollars on exchanges.

Electricity never crossed any border, but its value, embodied in Bitcoin, flowed to the global market.

During those years, China's computing power once accounted for over 70% of global Bitcoin mining power. China's hydropower and coal power, in this roundabout way, participated in a global redistribution of capital.

In 2021, this came to an abrupt halt. Regulatory crackdowns followed, miners scattered, and computing power migrated to Kazakhstan, Texas, and Canada.

But the logic itself never disappeared; it was just waiting for a new shell. Until ChatGPT emerged, large models began competing, and former Bitcoin mining farms transformed into AI data centers. Mining machines became computing GPUs, and the Bitcoin produced became tokens. The only constant was electricity.

Bitcoin exports and token exports are structurally similar in underlying logic, but tokens hold more commercial value today.

Bitcoin mining is pure mathematical computation, producing Bitcoin as a financial asset. Its value comes from scarcity and market consensus, unrelated to "what was computed." The computing power itself is unproductive, more like a byproduct of a trust mechanism.

Large model inference is different. GPUs consume electricity to produce real cognitive services: code, analysis, translation, creativity. The value of tokens comes directly from their utility to users. This is a deeper embedding; once a developer's workflow relies on a model, the cost of switching increases over time.

Of course, there is another key difference: Bitcoin mining was driven out of China, while token exports are actively chosen by global developers.

Token War

The submarine cable laid in 1858 represented the British Empire's sovereignty over the information highway. Whoever owned the infrastructure could define the rules of the game.

Token exports are also an undeclared war, full of resistance.

Data sovereignty is the first barrier. An API request from a U.S. developer processed by a Chinese data center means data physically flows through China. For individual developers and small applications, this isn't an issue. But for enterprise-sensitive data, financial information, and government compliance scenarios, this is a hard stop. This is why Chinese models have the highest penetration in development tools and personal applications but almost no presence in core enterprise systems.

The chip ban is the second barrier. China's AI development faces export controls on NVIDIA's high-end GPUs. MoE architecture and algorithmic optimization can only partially offset this disadvantage; the ceiling still exists.

But the current resistance is just the prologue; a larger battlefield is forming.

Tokens and AI models have become a new strategic dimension in the U.S.-China rivalry, comparable to semiconductors and the internet in the 20th century, or even closer to an older analogy: the space race.

In 1957, the Soviet Union launched Sputnik 1, shocking the U.S., which then launched the Apollo program, investing resources equivalent to hundreds of billions of dollars today to ensure it did not fall behind in the space race.

The logic of the AI race is strikingly similar but will be far more intense than the space race. Space is physical, something ordinary people don't feel. AI渗透的是经济的毛细血管,渗透到每一行代码、每一份合同、每一个政府决策系统的背后,都可能运行着某个国家的大模型. Whose model becomes the default infrastructure choice for global developers gains structural influence over the global digital economy.

This is what truly worries Washington about China's token exports.

When a developer's codebase, agent workflows, and product logic are built around a Chinese model's API, migration costs rise exponentially over time. Even if the U.S. legislates restrictions, developers will resist with their feet, just as no programmer can abandon GitHub today.

Today's token exports may be just the opening chapter of this long game. Chinese large models claim no intent to颠覆什么; they simply deliver services to every global developer with an API key at a lower price.

This time, those laying the cables are the engineering teams writing code in Hangzhou, Beijing, and Shanghai, and the GPU clusters running day and night in a southern province.

This rivalry has no countdown; it happens 24 hours a day, measured in tokens, on every developer's terminal.


Twitter:https://twitter.com/BitpushNewsCN

Bitpush TG Discussion Group:https://t.me/BitPushCommunity

Bitpush TG Subscription: https://t.me/bitpush

Original link:https://www.bitpush.news/articles/7614803

İlgili Sorular

QWhat is the core concept behind China's AI models 'exporting' power through tokens, as described in the article?

AThe core concept is that while the physical electricity never leaves China's grid, its value is delivered cross-border through AI-generated tokens. When global developers use Chinese AI models via API requests, the computation (and thus power consumption) occurs in Chinese data centers, but the resulting tokens—representing cognitive services like code, analysis, or translation—are transmitted back to users worldwide. This effectively 'exports' the value of China's cheap electricity without the power itself crossing any borders, bypassing traditional trade mechanisms like tariffs.

QAccording to the article, what event in early 2026 acted as a catalyst for the massive migration of global developers to Chinese AI models?

AThe catalyst was the release of 'OpenClaw,' an open-source tool that allows AI to directly control computers and execute complex workflows. This caused token consumption to become exponential rather than linear. When Anthropic and Google banned the use of subscription credentials for OpenClaw, developers sought cheaper alternatives. They migrated en masse to Chinese models like MiniMax M2.5, which offered comparable performance to models like Claude Opus but at a fraction of the cost (e.g., $0.3 per million input tokens vs. $5).

QHow does the article compare the historical export of Bitcoin's hashing power to the current export of AI tokens?

AThe article draws a parallel, stating both are forms of 'power export.' From around 2015, China used cheap hydropower and coal power to mine Bitcoin, converting electricity into a globally traded digital asset without the power physically leaving the grid. This was shut down in 2021. The logic now repeats with AI: GPU clusters in data centers consume Chinese electricity to produce valuable tokens (cognitive services) for global users. The key difference is that Bitcoin mining produced a financial asset based on scarcity, while AI tokens produce utility (e.g., code, analysis), creating deeper user dependency. Bitcoin mining was driven out of China, but token export is being actively chosen by global developers.

QWhat are the two main components of a token's cost structure, as explained in the article?

AThe two main cost components are computing power (算力) and electricity (电力). Computing power refers to the depreciation and amortization of GPUs (e.g., an Nvidia H100 costs about $30,000, and its lifespan cost per inference is the depreciation cost). Electricity is the fuel for data center operation, with a single GPU consuming about 700 watts at full load, plus cooling costs, leading to annual power bills for large AI data centers that can easily exceed hundreds of millions of dollars.

QWhat major challenges or 'walls' does the article identify for China's AI models in the global 'Token War'?

AThe article identifies two major 'walls' or challenges: 1. Data Sovereignty: API requests processed in Chinese data centers mean data flows through China physically. This is a significant concern for enterprises handling sensitive data, financial information, or government compliance, limiting Chinese model adoption in core systems despite high penetration in developer tools. 2. Chip Ban: Export controls on high-end Nvidia GPUs create a hardware disadvantage for Chinese AI development. While architectural optimizations like MoE (Mixture of Experts) can partially offset this, a ceiling on performance and capability remains.

İlgili Okumalar

The Essence of AI Layoffs: Why More AI Adoption Leads to More Corporate Anxiety?

The author, awaiting potential inclusion on an 8000-person layoff list, analyzes the true nature of recent "AI-driven" layoffs. They argue that while AI use, particularly tools like Claude for code generation, has skyrocketed and boosted developer output (e.g., 2-5x more code commits), this has not translated into proportional business growth or revenue. The core issue is a misalignment between increased "Input" (code) and tangible "Outcomes" (user value, revenue). AI acts as a costly B2B SaaS, inflating operational expenses without guaranteed returns. Two key problems emerge: 1) The friction that once filtered out bad ideas is gone, as AI allows cheap pursuit of even weak concepts. 2) Organizational "alignment tax"—the difficulty of coordinating across teams—becomes crippling when development velocity outpaces consensus-building. Thus, layoffs serve two immediate purposes: 1) To offset ballooning AI costs (Token consumption) and maintain cash flow, as rising input costs without outcome growth destroys unit economics. 2) To reduce organizational bloat and alignment friction by simply removing teams, thereby speeding up execution in the short term. Therefore, these layoffs are fundamentally caused by AI, even if AI doesn't directly replace roles. They represent a painful correction until companies learn to convert AI-driven productivity into real business outcomes and streamline organizational coordination to match the new pace of work. The cycle will continue until this learning curve is mastered.

marsbit31 dk önce

The Essence of AI Layoffs: Why More AI Adoption Leads to More Corporate Anxiety?

marsbit31 dk önce

Can the Solana Foundation and Google's Collaboration on Pay.sh Bridge the Payment Link Between Web2 and Web3 in the Agent Economy?

Solana Foundation, in collaboration with Google Cloud, has launched Pay.sh, a payment gateway designed to bridge the gap between AI agents and enterprise-grade service infrastructure. The initiative aims to solve a key bottleneck in the "agent economy": existing payment systems are ill-suited for autonomous AI agents. Traditional methods like credit cards require human verification, while newer on-chain protocols like x402 and MPP create a separate, Web3-native system that raises barriers for service providers. Pay.sh functions as a universal payment layer. It allows users to fund a Solana wallet via credit card or stablecoin, which then acts as an identity and payment proxy for AI agents. When an agent needs to access a paid API service (e.g., Google Cloud, Alibaba Cloud), Pay.sh handles the transaction seamlessly. It leverages the HTTP 402 status code ("Payment Required") to initiate payments, intelligently choosing between one-time transfers (x402-style) or session-based authorizations (MPC-style) based on the service's billing model. This spares agents from manual account registration and API key management. A key feature for service providers is low integration effort. They can adopt Pay.sh by providing a declarative configuration file, enabling features like tiered pricing, free tiers, and automatic revenue splitting to multiple addresses (e.g., for royalties, cloud costs). Providers can also list their APIs in a central Pay Skill Registry for agent discovery. The collaboration with Google Cloud provides crucial infrastructure for API proxying, traffic routing, and compliance logging, aiming to keep agent activities within regulated boundaries. By connecting Web2 services with Web3 payment rails, Pay.sh positions the Solana wallet as a foundational identity and payment tool for AI agents, potentially driving more transaction volume to the Solana ecosystem. However, the report notes challenges. The service registry currently lacks robust vetting, risking exposure to unauthorized or malicious third-party APIs. Pay.sh also inherits security and compatibility risks from its underlying payment protocols (x402, MPC). Furthermore, adoption may be hindered by varying regional data privacy and payment compliance regulations among API providers. Despite these hurdles, Pay.sh represents a significant step towards integrating Web2 and Web3 for autonomous agent commerce.

marsbit38 dk önce

Can the Solana Foundation and Google's Collaboration on Pay.sh Bridge the Payment Link Between Web2 and Web3 in the Agent Economy?

marsbit38 dk önce

Bitcoin's Bull-Bear Cycle Indicator Turns Positive for the First Time in 7 Months: End of Bear Market or False Breakout?

Bitcoin's "Bull-Bear Market Cycle Indicator" from CryptoQuant has turned positive for the first time since October 2025. This gauge, based on the P&L Index relative to its 365-day moving average, suggests a potential shift from a bear market phase. Concurrently, the Bull Score Index rose to a neutral reading of 50 in late April. The indicator's move into positive territory follows a roughly 35% price rebound from a low near $60,000 in February to above $81,000. The recovery over approximately three months was faster than the 12-month period observed during the 2022 bear market. However, analysts caution against premature optimism, citing a historical precedent from March 2022. Back then, the Bull Score Index briefly hit 50, but it proved to be a false signal as Bitcoin's price subsequently plunged further. Structural differences exist in the current cycle, including consistent inflows into spot Bitcoin ETFs and an increase in large holder addresses. Yet, some models, referencing the four-year halving cycle, suggest a potential deeper bottom near $50,000 might still be possible around late 2026. In summary, while on-chain data shows marked improvement and the worst panic may be over, market participants remain cautious. A convincing trend reversal confirmation likely requires Bitcoin to sustainably break above key resistance, such as the 200-day moving average near $82,000.

marsbit45 dk önce

Bitcoin's Bull-Bear Cycle Indicator Turns Positive for the First Time in 7 Months: End of Bear Market or False Breakout?

marsbit45 dk önce

How to Automate Any Workflow with Claude Skills (Complete Tutorial)

This is a comprehensive guide to mastering Claude Skills, a feature for creating permanent, reusable instruction sets that automate specific workflows. Unlike simple saved prompts, Skills function like trained employees, delivering consistent, high-quality outputs by defining the entire task process, standards, error handling, and output format. The guide is structured in four phases: **Phase 1: Installation (5 minutes).** Skills are folders containing a `SKILL.md` file. The user is instructed to find a relevant Skill online, install it, test it on a real task, and compare its performance to one-off prompts. **Phase 2: Building Your First Custom Skill.** Start by rigorously defining the Skill's purpose, trigger phrases, and providing a concrete example of perfect output. The `SKILL.md` file has two parts: a YAML frontmatter with a specific name/description/triggers, and a detailed, step-by-step workflow written in natural language with examples and quality standards. **Phase 3: Testing & Optimization for Production.** Test the Skill in three scenarios: 1) a standard, common task; 2) edge cases with missing or conflicting data; and 3) a pressure test with maximum complexity. Any failure indicates a needed instruction. Implement a weekly optimization cycle to continuously refine the Skill based on real usage. **Phase 4: Building a Complete Skill Library.** The goal is to create a team of Skills for all repetitive tasks. Examples are given for industries like real estate, marketing, finance, consulting, and e-commerce. The user should list their tasks, prioritize them, and build one new Skill per week, maintaining a master document to track their library. The conclusion emphasizes the compounding time savings: ten Skills saving 30 minutes each per week reclaims over 260 hours (6.5 work weeks) per year, fundamentally transforming one's work system.

marsbit1 saat önce

How to Automate Any Workflow with Claude Skills (Complete Tutorial)

marsbit1 saat önce

İşlemler

Spot
Futures
活动图片