The Next Earthquake in AI: Why the Real Danger Isn't the SaaS Killer, but the Computing Power Revolution?

marsbitPublished on 2026-02-11Last updated on 2026-02-11

Abstract

The next seismic shift in AI is not the threat of "SaaS killers" but a fundamental revolution in computing power. While many focus on how AI applications like Claude Cowork are disrupting traditional software, the real transformation is happening beneath the surface—in the infrastructure that powers AI. Two converging technological paths are challenging NVIDIA’s GPU dominance: 1. **Algorithmic Efficiency**: DeepSeek’s Mixture-of-Experts (MoE) architecture allows massive models (e.g., DeepSeek-V2 with 236B parameters) to activate only a small fraction of "experts" (9%) during computation, achieving GPT-4-level performance at 10% of the computational cost. This decouples AI capability from sheer compute power. 2. **Specialized Hardware**: Inference-optimized chips from companies like Cerebras and Groq integrate memory directly onto the chip, eliminating data transfer delays. This "zero-latency" design drastically improves speed and efficiency, prompting even OpenAI to sign a $10B deal with Cerebras. Together, these advances could cause a cost collapse: training costs may drop by 90%, and inference costs could fall by an order of magnitude. The total cost of running world-class AI may plummet to 10-15% of current GPU-based solutions. This paradigm shift threatens NVIDIA’s valuation, built on the assumption of perpetual GPU dominance. If the market realizes that GPUs are no longer the only—or best—option, the foundation of NVIDIA’s trillions in market cap could crumble. The ...

Written by: Bruce

Lately, the entire tech and investment communities have been fixated on the same thing: how AI applications are "killing" traditional SaaS. Since @AnthropicAI's Claude Cowork demonstrated how easily it can help you write emails, create PPTs, and analyze Excel spreadsheets, a panic about "software is dead" has begun to spread. This is indeed frightening, but if your gaze stops here, you might be missing the real seismic shift.

It's like we're all looking up at a drone dogfight in the sky, but no one notices that the entire continental plate beneath our feet is quietly shifting. The real storm is hidden beneath the surface, in a corner most people can't see: the computing power foundation that supports the entire AI world is undergoing a "silent revolution."

And this revolution might end the grand party hosted by AI's shovel seller: NVIDIA @nvidia, much sooner than anyone imagined.

Two Revolutionary Paths Converging

This revolution isn't a single event, but rather the convergence of two seemingly independent technological paths. They are like two armies closing in, forming a pincer movement against NVIDIA's GPU hegemony.

The first path is the algorithm slimming revolution.

Have you ever thought about whether a super brain really needs to mobilize all its cells when thinking about a problem? Obviously not. DeepSeek figured this out with their Mixture of Experts (MoE) architecture.

You can think of it like a company with hundreds of experts in different fields. But every time you need to solve a problem, you only call upon the two or three most relevant experts, rather than having everyone brainstorm together. This is the cleverness of MoE: it allows a massive model to activate only a small portion of its "experts" during each computation, drastically saving computing power.

What's the result? The DeepSeek-V2 model nominally has 236 billion "experts" (parameters), but only needs to activate 21 billion of them for each task—less than 9% of the total. Yet its performance is comparable to GPT-4, which requires 100% full operation. What does this mean? AI capability is decoupling from the computing power it consumes!

In the past, we all assumed that the stronger the AI, the more GPUs it would need. Now, DeepSeek shows us that through clever algorithms, the same effect can be achieved at one-tenth the cost. This directly puts a huge question mark on the essential nature of NVIDIA GPUs.

The second path is the hardware "lane change" revolution.

AI work is divided into two phases: training and inference. Training is like going to school, requiring reading countless books (data); here, GPUs with their "brute force" parallel computing are indeed useful. But inference is like our daily use of AI, where response speed is more critical.

GPUs have an inherent weakness in inference: their memory (HBM) is external, and data transfer back and forth causes latency. It's like a chef whose ingredients are in a fridge in the next room; every time they cook, they have to run over to get them—no matter how fast, it can't be instant. Companies like Cerebras and Groq started from scratch, designing dedicated inference chips that solder the memory (SRAM) directly onto the chip, putting the ingredients right at hand, achieving "zero latency" access.

The market has voted with real money. OpenAI, while complaining about NVIDIA's GPU inference performance, turned around and signed a $10 billion deal with Cerebras specifically to rent their inference services. NVIDIA itself panicked, spending $20 billion to acquire Groq, precisely to not fall behind in this new race.

When the Two Paths Converge: A Cost Avalanche

Now, let's put these two things together: run an algorithmically "slimmed-down" DeepSeek model on a hardware platform with "zero latency" like a Cerebras chip.

What happens?

A cost avalanche.

First, the slimmed-down model is small enough to be loaded entirely into the chip's built-in memory at once. Second, without the external memory bottleneck, the AI's response speed becomes astonishingly fast. The final result: training costs drop by 90% due to the MoE architecture, and inference costs drop by another order of magnitude due to specialized hardware and sparse computation. In total, the cost of owning and operating a world-class AI could be just 10%-15% of the traditional GPU solution.

This isn't an improvement; it's a paradigm shift.

The Carpet is Being Pulled from Under NVIDIA's Throne

Now you should understand why this is more fatal than the "Cowork panic."

NVIDIA's multi-trillion dollar valuation today is built on a simple story: AI is the future, and the future of AI depends on my GPUs. But now, the foundation of that story is being shaken.

In the training market, even if NVIDIA maintains its monopoly, if customers can do the work with one-tenth the GPUs, the overall size of this market could shrink significantly.

In the inference market, a cake ten times larger than training, NVIDIA not only lacks an absolute advantage but is also facing a siege from various players like Google and Cerebras. Even its biggest customer, OpenAI, is defecting.

Once Wall Street realizes that NVIDIA's "shovels" are no longer the only, or even the best, option, what will happen to the valuation built on the expectation of "permanent monopoly"? I think we all know.

Therefore, the biggest black swan in the next six months might not be which AI application has killed what, but a seemingly insignificant piece of tech news: for example, a new paper on the efficiency of MoE algorithms, or a report showing a significant increase in market share for dedicated inference chips, quietly announcing a new phase in the computing power war.

When the "shovel seller's" shovels are no longer the only choice, his golden age may well be over.

Related Questions

QWhat is the core argument of the article regarding the next major shift in AI?

AThe article argues that the next major shift in AI is not the threat of AI applications killing traditional SaaS, but rather a 'silent revolution' in the computational power (compute) that underpins the entire AI world. This revolution, driven by algorithmic efficiency and new hardware, threatens to disrupt the dominance of companies like NVIDIA.

QWhat are the two technological paths converging to challenge NVIDIA's GPU dominance?

AThe two converging paths are: 1) The algorithmic 'slimming revolution,' exemplified by architectures like Mixture of Experts (MoE) from DeepSeek, which drastically reduces the computational power needed for a given level of performance. 2) The hardware 'lane-changing revolution,' with companies like Cerebras and Groq designing specialized inference chips that eliminate memory bottlenecks, offering vastly faster and more efficient processing than traditional GPUs.

QHow does the Mixture of Experts (MoE) architecture, as used in DeepSeek-V2, achieve its efficiency?

AThe MoE architecture works like a company of experts. Instead of activating the entire massive model for every task, it only activates the most relevant small subset of 'experts' (a fraction of the total parameters). For example, DeepSeek-V2 has 236 billion parameters but only activates 21 billion (less than 9%) for a given task, achieving performance comparable to models that require 100% activation, thus decoupling AI capability from compute consumption.

QWhat specific market action is cited as evidence of the shift away from NVIDIA's GPUs for AI inference?

AThe article cites OpenAI's actions as key evidence: while complaining about the inefficiency of NVIDIA GPUs for inference, OpenAI signed a $10 billion deal to rent inference services from Cerebras, a company specializing in dedicated inference chips. Additionally, NVIDIA's own response—spending $20 billion to acquire Groq—is presented as a move to avoid falling behind in this new hardware paradigm.

QWhat is the potential financial impact on NVIDIA if the described compute revolution succeeds?

AThe article suggests a potential 'cost avalanche' where the total cost of owning and operating a world-class AI could drop to just 10-15% of the traditional GPU-based solution. This would severely challenge NVIDIA's business model, which is built on the premise that AI's future is dependent on its GPUs. If the market realizes NVIDIA's 'shovels' are no longer the only or best option, the 'permanent monopoly' expectation underpinning its multi-trillion dollar valuation could collapse.

Related Reads

Warsh Hearing Concludes: What Are the Notable Signals for the Crypto Industry?

The Senate Banking Committee held a confirmation hearing for Judy Shelton, a Federal Reserve nominee, who faced intense questioning regarding her ability to maintain the central bank's independence amid pressure from President Trump to lower interest rates. Shelton denied any pre-arranged commitments on rate cuts and emphasized her independence, though Democrats remained skeptical, citing contradictions with Trump's public statements. Shelton characterized post-pandemic inflation as a major policy failure and called for a "regime change" in the Fed’s approach, including reforms to inflation measurement and communication strategies. She criticized the current practice of Fed officials frequently signaling future rate moves and did not commit to maintaining post-meeting press conferences, suggesting potential reductions in transparency. Regarding crypto markets, Shelton’s extensive investments in digital asset companies—including Solana, DeFi, and blockchain infrastructure—were noted, though she has pledged to divest these holdings due to ethics rules. Her familiarity with the crypto industry and deregulatory leanings may signal a more open, though cautious, stance toward digital assets. However, concerns were raised about potential conflicts of interest, especially given Trump family involvement in crypto-financial ventures. The timing of her confirmation remains uncertain, pending a Justice Department investigation into current Chair Powell. Shelton’s potential leadership could lead to a more hawkish, productivity-focused Fed with tighter policy communication—factors that may significantly influence liquidity conditions and macro narratives for crypto markets.

marsbit4h ago

Warsh Hearing Concludes: What Are the Notable Signals for the Crypto Industry?

marsbit4h ago

Trading

Spot
Futures

Hot Articles

Discussions

Welcome to the HTX Community. Here, you can stay informed about the latest platform developments and gain access to professional market insights. Users' opinions on the price of AI (AI) are presented below.

活动图片