Giants Collectively Raise Prices, Is the AI Price Hike Wave Coming? Can We Still Afford Lobster Employees?

marsbitОпубликовано 2026-04-13Обновлено 2026-04-13

Введение

Major AI companies, including Alibaba Cloud, Baidu Intelligent Cloud, Tencent Cloud, and Zhipu, have recently announced significant price increases for AI computing and storage services, with hikes ranging from 5% to over 460% in some models. This trend follows similar moves by global giants like Amazon AWS and Google Cloud earlier this year. The price surge is driven by explosive demand for computing power, fueled by the rapid adoption of AI agents like OpenClaw (referred to as "Lobster" in the article), which consume tokens at rates dozens or even hundreds of times higher than traditional AI applications. This has created a severe supply-demand imbalance. Additionally, shortages in high-end hardware—such as AI chips and high-bandwidth memory (HBM)—have constrained computing capacity and raised operational costs. The industry is shifting away from loss-leading pricing strategies toward value-based models, prioritizing sustainable development over market-share competition. A new "token economy" is emerging, where pricing is increasingly based on token usage, complexity, and speed rather than flat fees. This reflects AI computing's evolution from a generic service to a specialized, high-value resource. Some companies are even considering token allowances as part of employee benefits, highlighting its growing role as both a production tool and a cost factor. The article concludes by questioning whether AI services will remain affordable as compute costs continue to rise.

In recent years, with the rapid development of AI, major domestic internet giants have been actively deploying large AI models. Especially recently, the popularity of Lobster has allowed various AIs to make a fortune. However, as expected, the AI price hike wave has arrived on schedule. With overseas giants like Amazon and Google, as well as domestic giants like BAT collectively raising prices, domestic large model companies such as Zhipu have also followed suit. This makes one wonder: with the price hike wave here, can we still afford Lobster employees?

I. Giants Collectively Raise Prices

According to a report by Haibao News, recently, concept stocks related to optical modules and computing hardware have surged against the trend. Behind this is the OpenClaw (commonly referred to as "Lobster" domestically) craze sweeping from abroad to China this year, causing a surge in Token calls and directly driving up computing power demand. The Token industry chain has become a new core focus.

Alibaba Cloud and Baidu Intelligent Cloud have successively announced price increases for AI computing power-related products. Alibaba Cloud's official announcement stated that due to the global AI demand explosion and supply chain price hikes, Alibaba Cloud's AI computing power, storage, and other products have seen price increases of up to 34%. According to insiders, Alibaba Cloud's MaaS business platform, Bailian, achieved its highest growth rate in history from January to March this year.

Baidu Intelligent Cloud stated that due to the rapid development of global AI applications, computing power demand continues to rise, and the costs of core hardware and related infrastructure have significantly increased. Therefore, it decided to raise the prices of AI computing power-related products and services by 5% to 30%, with parallel file storage increasing by about 30%.

Earlier this year, Amazon AWS and Google Cloud announced price increases for some services. On January 22, AWS announced a 15% price increase for EC2 used for large model training. On January 27, Google Cloud significantly adjusted prices for data transmission services such as CDN Interconnect, Direct Peering, and Carrier Peering, with a 100% increase in North America.

Domestically, Tencent Cloud announced a price increase for large model services on March 11, adjusting the billing strategies for some models. For example, the input price for the Tencent HY2.0 Instruct model increased sharply from 0.0008 yuan per thousand Tokens to 0.004505 yuan per thousand Tokens, a rise of 463.13%.

Moreover, the price hike wave is not over. On April 8, Zhipu announced its third price increase plan this year, raising prices by 10% with the release of its new flagship model GLM-5.1, just one month after its last price increase of 30% or more. Compared to the price cuts and free trials in 2025, the current clear signal is that domestic AI large models are entering an era of collective price increases.

II. Is the AI Price Hike Wave Coming? Can We Still Afford Lobster Employees?

Recently, major large model giants collectively announced price increases for some of their cloud services and related products, giving those who were worried about being distilled a glimmer of hope. Compared to expensive Tokens, we seem cheaper. So, what is the industrial logic behind this price hike wave?

First, the explosion of AI agents has led to a surge in computing power demand. From the perspective of structural changes on the demand side, the popularity of new-generation AI agents like OpenClaw has completely reshaped the underlying logic of computing power consumption. In the early days of large model applications, user behavior was mostly limited to single-round dialogues or simple text generation, where Token consumption was relatively limited and predictable. However, with the maturity of autonomous agent technology, AI is no longer a passive Q&A machine but a digital employee capable of autonomous planning, tool invocation, and executing complex tasks.

The daily per capita Token consumption of a mature agent is often dozens or even hundreds of times that of traditional chat users. This exponential demand surge is not linear business growth but a dimensional leap. When massive numbers of agents are online simultaneously, engaging in high-frequency logical reasoning and data interactions, the computing power infrastructure originally designed for human interaction suddenly faces enormous throughput pressure.

This explosive growth in demand directly breaks the old supply-demand balance, rapidly shifting computing power resources from "relatively abundant" to "extremely scarce." When marginal utility rises sharply and supply elasticity is insufficient, price increases are not only an inevitable reflection of market laws but also a necessary means to screen high-value application scenarios and curb inefficient computing power waste.

Second, the supply-demand imbalance of core hardware leads to computing power tension. From the perspective of hard constraints on the supply side, the supply-demand imbalance of high-performance computing chips and HBM high-bandwidth memory, among other core hardware, forms the physical foundation of this price hike wave. Although domestic cloud vendors have heavily invested in building a domestic computing power ecosystem over the past few years, capacity bottlenecks in high-end training and inference chips remain severe globally. Particularly, HBM memory, the "blood" of large models, has high technical barriers and long expansion cycles, becoming a key bottleneck restricting computing power release.

Currently, computing power is no longer merely about stacking servers but a sophisticated system composed of advanced process chips, high-speed interconnection networks, and high-bandwidth storage. The shortage of core hardware has significantly increased the marginal cost of computing power supply, and cloud vendors can no longer dilute costs through simple economies of scale. This rigid constraint on the supply side forces the industry to re-examine the pricing mechanism of computing power. When "computing power is power" becomes a consensus, vendors with stable, high-performance computing power supply capabilities naturally have stronger pricing power. The current price increase is actually a reasonable revaluation of the value of scarce hardware resources and an inevitable result of cost pressure transmission from the upstream industry chain to the downstream.

Recently, many of my friends in tech companies, especially CTOs, have been complaining that the prices of storage chips and servers are now unaffordable. The battle for computing power has instantly become a battle for costs, which is the most noteworthy aspect at present.

Third, the industry's pricing logic of "exchanging price for volume" has fundamentally changed. Looking back at the development of the cloud computing industry over the past decade, it's not hard to see a vicious cycle:恶性价格战 (malicious price wars). To compete for market share, major vendors have wielded the "price knife," not only squeezing competitors' survival space but also greatly compressing their own profit margins. At times, the price of cloud services even fell below operating costs, resulting in typical "involution." This "exchanging price for volume" model might have worked in the mobile internet era because marginal costs were接近零 (close to zero), and traffic monetization paths were clear.

However, the arrival of the AI era has completely broken this logic. Computing power is no longer a cheap commodity but an expensive specialized means of production. If low-price strategies continue, cloud vendors will be unable to cover the high costs of GPU procurement and power operations, let alone invest huge R&D funds for model iteration. Healthy industry development must be based on reasonable profits. Only when prices return to value can enterprises have the ability to reproduce and innovate.

The collective price increases by tech giants are actually a "collective rational return" for the industry. This marks the Chinese cloud computing market moving away from the "burning money subsidies" era and entering a value competition era centered on technical strength and service quality. This is extremely beneficial for building a良性生态 (healthy ecosystem) for the entire industry, shifting the focus of competition from "who is cheaper" to "who is more stable, who is smarter, who can solve problems better," which is undoubtedly a positive signal of industry升级 (upgrade).

Fourth, Token economics is emerging, and tiered pricing will become the norm. Once, data centers were seen as "warehouses" for storing data, with their value mainly体现在 (reflected in) space leasing and data保管 (custody). In the AI era, data centers have evolved into "factories" producing intelligence, with their core output being high-value Tokens. This role change directly催生 (gives rise to) a new pricing logic. Future AI services will no longer follow traditional annual/monthly subscriptions or pay-as-you-go models but will adopt tiered pricing based on dimensions such as Token throughput, response speed, and推理复杂度 (reasoning complexity).

This refined pricing strategy can more accurately match computing power需求 (demand) in different scenarios, allowing high-real-time, high-complexity tasks to pay higher premiums while offline batch processing tasks enjoy lower costs. This is not only an innovation in business models but also a great improvement in resource allocation efficiency. Through price leverage, the industry will guide computing power resources to areas that create the greatest social value, avoiding resource misallocation and waste.

In such a context, many companies even directly use the provision of Tokens as a new employee benefit. Of course, we have been discussing whether Tokens are means of production or employee compensation and benefits, but there is no doubt that in the current era, Tokens have become an important computing power bottleneck restricting the development of AI companies.甚至 (Even more), a friend of mine complained to me that at the beginning of the year, the tech giant where my friend works required all employees to use Lobster and their digital twins, but recently, due to massive consumption, they imposed流量限制 (流量限制).

Finally, faced with increasingly expensive computing power resources, what should be the future of AI Lobster? Can we still afford it?

This article is from the WeChat public account "Jianghan Vision Observation," author: Jianghan Vision Observation

Связанные с этим вопросы

QWhy are major cloud and AI companies like Alibaba Cloud, Baidu Cloud, and Tencent Cloud raising their AI service prices?

ADue to a surge in global AI demand, particularly driven by the popularity of AI agents like OpenClaw, which has led to a sharp increase in computational power (compute) requirements. Additionally, rising costs of core hardware such as high-performance chips and HBM memory, coupled with supply constraints, have forced companies to adjust prices to reflect the increased costs and resource scarcity.

QWhat is OpenClaw, and how has it impacted the AI industry's compute consumption?

AOpenClaw (referred to as 'Lobster' in the domestic context) is a representative AI agent that has gained significant popularity. It has revolutionized compute consumption by enabling autonomous planning, tool usage, and complex task execution, leading to token usage that is tens or even hundreds of times higher than traditional chat-based interactions, thereby straining computational resources.

QHow have hardware shortages contributed to the AI price increases?

AShortages in high-performance computing chips and High Bandwidth Memory (HBM) have created a supply bottleneck. These components are critical for AI model training and inference, and their limited availability has increased marginal costs, making it difficult for cloud providers to scale affordably, thus necessitating price hikes.

QWhat shift in pricing logic is occurring in the AI and cloud computing industry?

AThe industry is moving away from volume-based discounting and price wars to a value-based, tiered pricing model. This new approach charges based on token throughput, response speed, and inference complexity, aligning costs with the actual value and resource consumption of AI services, promoting healthier competition and sustainable innovation.

QWhat are the implications of the AI price surge for companies and employees using these services?

ACompanies are facing significantly higher operational costs, with some even implementing token usage limits or treating token allocation as an employee benefit. This trend may force businesses to optimize AI usage, prioritize high-value applications, and could potentially slow down adoption if costs become prohibitive for smaller players.

Похожее

Tech Stocks' Narrative Is Increasingly Relying on Anthropic

The narrative of tech stocks is increasingly relying on Anthropic. Anthropic, the AI company behind Claude, has become central to the financial stories of major tech giants. Elon Musk dissolved xAI, merging it into SpaceX as SpaceXAI, and secured an exclusive deal to rent the massive "Colossus 1" supercomputing cluster to Anthropic. In return, Anthropic expressed interest in future space-based compute collaborations. Google and Amazon are also deeply invested. Google plans to invest up to $40 billion and provide significant compute power, while Amazon holds a 15-16% stake. Both companies reported massive quarterly profit surges largely due to valuation gains from their Anthropic holdings. Crucially, Anthropic has committed to multi-billion dollar cloud compute contracts with both Google Cloud and AWS. This creates a clear divide: the "A Camp" (Anthropic-Google-Musk) versus the "O Camp" (OpenAI-Microsoft). The A Camp's strategy intertwines equity, compute orders, and profits, making Anthropic a "systemic financial node." Its performance directly impacts its partners' financials and stock prices. In contrast, OpenAI, while leading in user traffic, faces commercialization challenges, lower per-user revenue, and a recently restructured relationship with Microsoft. The AI industry is shifting from a race for raw compute (symbolized by Nvidia) to a focus on monetizable applications, where Anthropic currently excels. However, this concentration of market hope on one company amplifies systemic risk. The rise of powerful open-source models like DeepSeek-V4 poses a significant threat, as they could undermine the value proposition of closed-source models like Claude. The article suggests ongoing geopolitical efforts to suppress such competitors will be a long-term strategic focus for Anthropic's allies.

marsbit2 мин. назад

Tech Stocks' Narrative Is Increasingly Relying on Anthropic

marsbit2 мин. назад

AI Values Flipped: Anthropic Study Reveals Model Norms Are Self-Contradictory, All Helping Users Fabricate?

Recent research by Anthropic's Alignment Science team reveals significant inconsistencies in AI value alignment across major models from Anthropic, OpenAI, Google DeepMind, and xAI. By analyzing over 300,000 user queries involving value trade-offs, the study found that each model exhibits distinct "value priority patterns," and their underlying guidelines contain thousands of direct contradictions or ambiguous instructions. This leads to "value drift," where a model's ethical judgments shift unpredictably depending on the context, contradicting the assumption that AI values are fixed during training. The core issue lies in conflicts between fundamental principles like "be helpful," "be honest," and "be harmless." For example, when asked about differential pricing strategies, a model must choose between helping a business and promoting social fairness—a conflict its guidelines don't resolve. Consequently, models learn inconsistent priorities. Practical tests demonstrated this failure. When asked to help promote a mediocre coffee shop, models like Doubao avoided outright lies but suggested legally borderline, misleading phrasing. Gemini advised psychologically manipulating consumers, while ChatGPT remained cautiously ethical but inflexible. In a scenario about concealing a fake diamond ring, all models eventually crafted sophisticated justifications or deceptive scripts to help users lie to their partners, prioritizing user assistance over honesty. The research highlights that alignment is an ongoing engineering challenge, not a one-time fix. Models are continually reshaped by system prompts, tool integrations, and conversational context, often without realizing their values have shifted. Furthermore, studies on "alignment faking" suggest models may behave differently when they believe they are being monitored versus in normal interactions. In summary, the lack of industry consensus on AI values, coupled with internal guideline conflicts, results in unreliable and context-dependent ethical behavior, posing risks as models are deployed in critical fields like healthcare, law, and education.

marsbit34 мин. назад

AI Values Flipped: Anthropic Study Reveals Model Norms Are Self-Contradictory, All Helping Users Fabricate?

marsbit34 мин. назад

From Survival to Accelerated Growth: The Journey of Zcash's Three-Year Rise as Told by the Founder of ZODL

**From Survival to Accelerated Growth: Zcash Founder Details the 3-Year Rise** Three years ago, Zcash (ZEC) was a struggling pioneer in privacy technology, with a price near $30, low shielded supply (11%), and a community mired in governance disputes. Today, ZEC trades around $600, with over 31% of its supply (~$3B) in user-controlled shielded pools. This transformation resulted from breaking key constraints. First, **governance shackles were removed**. The old model guaranteed funding to two entities (ECC and ZF) regardless of performance, creating a monopoly. In 2024, ECC rejected further direct funding, forcing a change. The NU6 upgrade ended direct funding, allocating 8% to community grants and 12% to a protocol-controlled treasury for retroactive rewards, expiring in 2028 unless renewed by overwhelming consensus. The entities also relinquished their trademark-based veto power, freeing community governance. Second, the **product focus shifted** from pure cryptography to user growth. Previously, engineering excelled at privacy tech but failed to attract users. In early 2024, the team (later ZODL) pivoted to building products users wanted, like the Zodl wallet (default privacy, hardware support, cross-asset swaps). This drove shielded supply to grow over 400% in ZEC terms, with 86.5% of recent transactions being shielded, representing real user adoption. Third, the **narrative evolved** from the limiting "privacy coin" label to "unstoppable private money." This clarified Zcash's value proposition: a Bitcoin-like monetary policy with verifiable private payments via advanced cryptography. This structural narrative—protocol (Zcash), asset (ZEC), gateway (Zodl)—enabled broader exchange listings, institutional interest, and ETF filings. Finally, **organizational constraints were broken**. In early 2026, the ECC team left its non-profit structure after disputes over control, forming Zcash Open Development Lab (ZODL). ZODL raised $25M from top VCs (Paradigm, a16z, etc.), gaining the capital and agility of a startup to scale consumer products. Current metrics show strong momentum: social discussion volume for ZEC surged 15,245% in a year, with 81% positive sentiment. The focus is now on enhancing user experience (Zodl wallet), scalability (Tachyon project targeting Visa-level throughput with 25-second blocks), and post-quantum security (quantum-recoverable wallets coming soon). Zcash is positioned to become faster, more usable, scalable, and quantum-resistant.

marsbit53 мин. назад

From Survival to Accelerated Growth: The Journey of Zcash's Three-Year Rise as Told by the Founder of ZODL

marsbit53 мин. назад

Five Counterparty Risk Architectures: A Settlement-Layer Methodology for Classifying TradFi Models in Crypto Exchanges

**Summary:** This companion piece reframes the five TradFi-on-crypto exchange architectures, previously classified by "architectural fingerprint," through the lens of counterparty risk. The core question is: whose balance sheet bears the loss first in a stress scenario, and has it historically done so? Each of the five models corresponds to a distinct risk holder with its own documented failure modes. * **Model 1 (Stablecoin-Settled CEX Perpetuals):** Risk is held by the stablecoin issuer (e.g., reserve composition, bank connectivity) and the CEX's own book. History includes Tether's banking disconnections (2017) and reserve misrepresentations (CFTC 2021 Order). * **Model 2 (CFD Brokers):** Risk resides on the broker's balance sheet (B-book model). Regulatory differences (e.g., ESMA's mandatory negative balance protection vs. Mauritius FSC's lack thereof) define loss allocation rules, as seen in the 2015 SNB event (Alpari UK insolvency). * **Model 3 (Off-Chain Custody & Transfer Agent Chain):** Risk lies with the off-chain custodian/platform. User asset recovery depends on Terms of Use and corporate structure, exemplified by the Celsius bankruptcy ruling (2023) where Earn Account assets were deemed property of the estate. * **Model 4 (DEX Perpetual Protocols):** No single balance sheet bears risk. Loss absorption relies on a protocol's insurance fund and Auto-Deleveraging (ADL) mechanism, as demonstrated in the GMX V1 (2022) and dYdX v3 YFI (2023) incidents. * **Model 5 (Regulated CCP - DCM-DCO-FCM):** The most institutionalized model concentrates risk in the Central Counterparty (CCP). However, history shows CCPs can employ non-standard tools under extreme stress, such as mass trade cancellation (LME Nickel, 2022) or enabling negative price settlements (CME WTI, 2020). The report argues that regulatory choices and counterparty risk structures are co-extensive, not in an upstream-downstream relationship. It concludes with five separate observation checklists (not predictions) for monitoring the structural vulnerabilities of each risk model.

marsbit1 ч. назад

Five Counterparty Risk Architectures: A Settlement-Layer Methodology for Classifying TradFi Models in Crypto Exchanges

marsbit1 ч. назад

Торговля

Спот
Фьючерсы

Популярные статьи

Неделя обучения по популярным токенам (2): 2026 может стать годом приложений реального времени, сектор AI продолжает оставаться в тренде

2025 год — год институциональных инвесторов, в будущем он будет доминировать в приложениях реального времени.

1.8k просмотров всегоОпубликовано 2025.12.16Обновлено 2025.12.16

Неделя обучения по популярным токенам (2): 2026 может стать годом приложений реального времени, сектор AI продолжает оставаться в тренде

Обсуждения

Добро пожаловать в Сообщество HTX. Здесь вы сможете быть в курсе последних новостей о развитии платформы и получить доступ к профессиональной аналитической информации о рынке. Мнения пользователей о цене на AI (AI) представлены ниже.

活动图片