Behind OpenAI's $110 Billion Financing: The Game Between Amazon and Microsoft

Odaily星球日报Опубліковано о 2026-02-28Востаннє оновлено о 2026-02-28

Анотація

OpenAI has secured $110 billion in new funding at a $730 billion pre-money valuation, with major investments from Amazon ($50 billion), Nvidia ($30 billion), and SoftBank ($30 billion). Notably, CEO Sam Altman’s public acknowledgment included Microsoft—an existing investor and partner—immediately after Amazon, signaling strategic alignment shifts. A key insight lies in two technical distinctions: “Stateless API” and “Stateful Runtime Environment.” Stateless API, the current mainstream model, handles single-request tasks without retaining context, but faces commoditization and margin pressure. In contrast, Stateful Runtime Environment supports persistent, autonomous AI agents capable of complex multi-step workflows, representing the future of enterprise AI adoption. Microsoft’s existing agreement ensures Azure remains the exclusive cloud provider for OpenAI’s Stateless API, securing present-day revenue streams. Meanwhile, Amazon’s expanded $100 billion partnership with OpenAI focuses on co-developing Stateful Runtime Environments via AWS Bedrock, positioning AWS as the infrastructure backbone for next-generation AI agents. This dual-cloud strategy enhances OpenAI’s leverage, reducing dependency on Microsoft while pitting Amazon and Microsoft in competition for future AI dominance. OpenAI gains negotiating power by diversifying its infrastructure partnerships and aligning each cloud giant with distinct—yet complementary—AI futures.

Original | Odaily Planet Daily (@OdailyChina)

Author | Azuma (@azuma_eth)

On the evening of February 27, OpenAI announced the completion of its latest funding round of $110 billion at a pre-money valuation of $730 billion.

The funding for this round came from three giants: Amazon invested $50 billion (an initial investment of $15 billion, with the remaining $35 billion to be delivered over the coming months upon meeting specific conditions), NVIDIA invested $30 billion (which will be recouped through the procurement of a total of 5 GW of computing power), and SoftBank also invested $30 billion.

After the funding was completed, OpenAI founder Sam Altman expressed his thanks to the three investors on his personal X account. Notably, Sam Altman's order of thanks was Amazon, Microsoft, NVIDIA, SoftBank — the name of Microsoft, the "old" shareholder and important partner who did not contribute in this round, was mentioned immediately after Amazon, which committed the most funding.

Aakash Gupta, an overseas blogger who has long followed the AI sector, pointed out that while most people are focused on the astronomical figure of $110 billion, the most critical piece of information in Sam Altman's statement lies in two overlooked technical terms: "Stateless API" and "Stateful Runtime Environment," which were attributed to Microsoft and Amazon, respectively.

Behind the Technical Terms: The Present and Future of AI

The core difference between Stateless API and Stateful Runtime Environment lies in the two words: "Stateless" and "Stateful".

The "Stateless" nature of a Stateless API means that the server does not maintain persistent state across requests — one call completes one inference; you ask a question, the AI answers, and after the lifecycle of this request ends, the system does not retain context or continue running. The "Stateful" nature of a Runtime Environment, on the other hand, implies a persistent execution environment — the Agent has historical memory, can persist, collaborate across tasks, and perform tasks over the long term.

Stateless API is currently the mainstream form of LLM commercialization. Industries such as finance, retail, manufacturing, and healthcare mostly integrate AI in this form to embed it into existing systems (e.g., various Q&A assistants, document summarization, search enhancement, etc.). The advantage of this model is that enterprises can quickly overlay AI capabilities into their existing architecture with minimal friction, achieving functional optimization without significant restructuring of organization and processes. However, as model capabilities converge, computing power costs continue to decrease, and price competition intensifies, the token-based billing of Stateless APIs is more likely to become standardized and commoditized, with marginal profits facing ongoing compression.

In contrast, Stateful Runtime Environment currently has limited commercial scale, but it represents not just simple "functional optimization" but a shift in business paradigm — it is not only able to answer questions but can also be seen as a digital workforce to concretely execute tasks. This means that the budget it touches will extend from mere interface call fees to automation, process management, and even part of the labor costs. Because of this, the market's expectations for Stateful Runtime Environment are far higher than its current scale.

Aakash Gupta also stated on this matter that in 2026 and 2027, almost all enterprise roadmaps will revolve around "autonomous agent workflows," not one-off API calls. Companies that invest heavily in AI in the future will increasingly tend to purchase systems that can run sustainably, collaborate across tools, and maintain context over the long term.

To put it in the simplest terms: Stateless API represents the present, while Stateful Runtime Environment represents the future.

What Did Microsoft and Amazon Take, Respectively?

On the day the funding was completed, Microsoft and Amazon each announced their latest cooperation agreements with OpenAI.

Microsoft stated in its announcement that the terms of the partnership jointly announced by Microsoft and OpenAI in October 2025 would remain unchanged (the terms include OpenAI purchasing $250 billion worth of Azure services). Azure remains the exclusive cloud provider for OpenAI's Stateless API; any Stateless API calls to OpenAI models generated through cooperation with any third party (including Amazon) will be hosted on Azure; OpenAI's first-party products, including Frontier, will also continue to be hosted on Azure.

Amazon stated in its announcement that AWS and OpenAI will jointly build a Stateful Runtime Environment powered by OpenAI models and provide services to AWS customers through Amazon Bedrock, helping enterprises build generative AI applications and Agents at production scale; AWS will also become the exclusive third-party cloud distribution service provider for OpenAI Frontier; the existing $38 billion multi-year cooperation agreement between AWS and OpenAI will be expanded to $100 billion, with a term of 8 years. OpenAI will consume 2 GW of Trainium computing power through AWS infrastructure to support the needs of Stateful Runtime Environment, Frontier, and other advanced workloads; OpenAI and Amazon will also develop customized models that can be used to support Amazon's customer-facing applications.

Comparing the two announcements, the current situation becomes clear.

Microsoft, with its $250 billion agreement and exclusive service rights, has locked in the current traffic engine. As long as OpenAI's Stateless API is called, Azure will bill behind the scenes — regardless of who the customer is or where the channel is, the traffic will ultimately return to Azure. This is a highly certain cash flow, but the problem lies in the trend of shrinking profit margins for Stateless APIs; call volume may continue to grow, but actual profits may not necessarily remain stable in the long run.

On the other hand, Amazon, with its $50 billion in real money and a $100 billion expanded agreement, has secured the underlying hosting rights for the AI Agent era for AWS. Once Agents become the core carriers of enterprise productivity, the resources that are truly consumed over the long term — computing power, storage, scheduling systems, workflow orchestration, and cross-tool collaboration — will all be deposited on AWS's runtime environment.

One controls the current cash flow; the other bets on the future structure of productivity.

OpenAI's Diversified Bet

Before the future truly arrives, no one knows whether Microsoft's or Amazon's choice is correct. But what is certain is that under these two clearly defined cooperation agreements with explicit interest divisions, OpenAI's initiative is significantly increasing.

Over the past few years, OpenAI has been highly dependent on Microsoft for cloud infrastructure. Microsoft is not only a major shareholder holding 27% of the shares but also the controller of the infrastructure. This binding brought effective early-stage resource advantages to OpenAI, but it also meant that the balance of bargaining power would naturally tilt toward Microsoft. With Amazon's strong entry, it and Microsoft are bound to engage in direct competition to seize the future service rights of OpenAI.

For OpenAI, this is a typical diversified betting strategy — not deeply binding with any single cloud service provider, not letting future growth be entirely constrained by one party, and using future business as a bargaining chip to obtain better terms.

Neither Microsoft nor Amazon can afford to give up OpenAI at present. When neither can leave the table, bargaining power naturally returns to OpenAI's hands.

Пов'язані питання

QWhat is the core difference between Stateless API and Stateful Runtime Environment in the context of AI?

AStateless API refers to a server-side service that does not retain persistent state across requests—each call completes one inference, and the system does not retain context or continue running after the request lifecycle ends. Stateful Runtime Environment, on the other hand, involves a persistent execution environment where AI agents have historical memory, can exist continuously, collaborate across tasks, and perform long-term tasks.

QWhich companies invested in OpenAI's latest $110 billion funding round, and what were their respective contributions?

AAmazon invested $50 billion ($15 billion initially, with the remaining $35 billion to be delivered over the coming months upon meeting specific conditions), Nvidia invested $30 billion (which will be recouped through the procurement of a total of 5 GW of computing power), and SoftBank also invested $30 billion.

QWhat did Microsoft secure in its agreement with OpenAI regarding Stateless API?

AMicrosoft secured Azure as the exclusive cloud provider for OpenAI's Stateless API. Any Stateless API calls to OpenAI models generated through cooperation with any third party, including Amazon, will be hosted on Azure. OpenAI's first-party products, including Frontier, will also continue to be hosted on Azure.

QWhat did Amazon obtain through its expanded agreement with OpenAI, particularly in relation to Stateful Runtime Environment?

AAmazon obtained a partnership with OpenAI to jointly build a Stateful Runtime Environment powered by OpenAI models, which will be provided to AWS customers via Amazon Bedrock. AWS also became the exclusive third-party cloud distribution service provider for OpenAI Frontier. The existing $38 billion multi-year cooperation agreement was expanded to $100 billion over 8 years, with OpenAI consuming 2 GW of Trainium computing power through AWS infrastructure to support Stateful Runtime Environment, Frontier, and other advanced workloads.

QHow does OpenAI's strategy of engaging with both Microsoft and Amazon reflect its approach to future growth and bargaining power?

AOpenAI's strategy reflects a分散押注 (diversified betting) approach. By not deeply binding with any single cloud service provider and avoiding complete dependence on one party for future growth, OpenAI uses its future business as a bargaining chip to secure more favorable terms. As both Microsoft and Amazon are unwilling to withdraw from the partnership, bargaining power naturally returns to OpenAI.

Пов'язані матеріали

OpenAI Post-Training Engineer Weng Jiayi Proposes a New Paradigm Hypothesis for Agentic AI

OpenAI engineer Weng Jiayi's "Heuristic Learning" experiments propose a new paradigm for Agentic AI, suggesting that intelligent agents can improve not just by training neural networks, but also by autonomously writing and refining code based on environmental feedback. In the experiment, a coding agent (powered by Codex) was tasked with developing and maintaining a programmatic strategy for the Atari game Breakout. Starting from a basic prompt, the agent iteratively wrote code, ran the game, analyzed logs and video replays to identify failures, and then modified the code. Through this engineering loop of "code-run-debug-update," it evolved a pure Python heuristic strategy that achieved a perfect score of 864 in Breakout and performed competitively with deep reinforcement learning (RL) algorithms in MuJoCo control tasks like Ant and HalfCheetah. This approach, termed Heuristic Learning (HL), contrasts with Deep RL. In HL, experience is captured in readable, modifiable code, tests, logs, and configurations—a software system—rather than being encoded solely into opaque neural network weights. This offers potential advantages in explainability, auditability for safety-critical applications, easier integration of regression tests to combat catastrophic forgetting, and more efficient sample use in early learning stages, as demonstrated in broader tests on 57 Atari games. However, the blog acknowledges clear limitations. Programmatic strategies struggle with tasks requiring long-horizon planning or complex perception (e.g., Montezuma's Revenge), areas where neural networks excel. The future vision is a hybrid architecture: specialized neural networks for fast perception (System 1), HL systems for rules, safety, and local recovery (also System 1), and LLM agents providing high-level feedback and learning from the HL system's data (System 2). The core proposition is that in the era of capable coding agents, a significant portion of an AI's learned experience could be maintained as an auditable, evolving software system.

marsbit52 хв тому

OpenAI Post-Training Engineer Weng Jiayi Proposes a New Paradigm Hypothesis for Agentic AI

marsbit52 хв тому

Your Claude Will Dream Tonight, Don't Disturb It

This article explores the recent phenomenon of AI companies increasingly using anthropomorphic language—like "thinking," "memory," "hallucination," and now "dreaming"—to describe machine learning processes. Focusing on Anthropic's newly announced "Dreaming" feature for its Claude Agent platform, the piece explains that this function is essentially an automated, offline batch processing of an agent's operational logs. It analyzes past task sessions to identify patterns, optimize future actions, and consolidate learnings into a persistent memory system, akin to a form of reinforcement learning and self-correction. The article draws parallels to similar features in other AI agent systems like Hermes Agent and OpenClaw, which also implement mechanisms for reviewing historical data, extracting reusable "skills," and strengthening long-term memory. It notes a key difference from human dreaming: these AI "dreams" still consume computational resources and user tokens. Further context is provided by discussing the technical challenges of managing AI "memory" or context, highlighting the computational expense of large context windows and innovations like Subquadratic's new model claiming drastically longer contexts. The core critique argues that this strategic use of human-centric vocabulary does more than market products; it subtly reshapes user perception. By framing algorithms with terms associated with consciousness, companies blur the line between tool and autonomous entity. This linguistic shift can influence user expectations, tolerance for errors, and even perceptions of responsibility when systems fail, potentially diverting scrutiny from the companies and engineers behind the technology. The article concludes by speculating that terms like "daydreaming" for predictive task simulation might be next, continuing this trend of embedding the idea of an "inner life" into computational processes.

marsbit54 хв тому

Your Claude Will Dream Tonight, Don't Disturb It

marsbit54 хв тому

Торгівля

Спот
Ф'ючерси
活动图片