Meta: Can Afford Trillion-Dollar Computing Power, But Can't Keep Key People

marsbit2026-02-28 tarihinde yayınlandı2026-02-28 tarihinde güncellendi

Özet

Meta's AI Ambition: A $135 Billion Bet on Chips, But Losing Key Talent In July 2025, Meta recruited top AI infrastructure engineer Ruoming Pang from Apple with a compensation package worth over $200 million. However, just seven months later, he left for OpenAI, forfeiting much of his unvested equity. This high-profile departure is part of a broader trend of key talent leaving Meta's AI division, including Chief AI Scientist Yann LeCun and other senior figures. The exodus is largely attributed to the fallout from the Llama 4 model's release in April 2025. The model was later revealed to have been benchmarked unethically, using different model versions to optimize scores on different tests, severely damaging trust within the developer community. This scandal led CEO Mark Zuckerberg to lose confidence in the team, resulting in a major reorganization. He appointed 28-year-old Scale AI CEO Alexandr Wang as Chief AI Officer, who now oversees the new Meta Superintelligent Lab (MSL). The planned flagship model, Llama 4 Behemoth, was indefinitely delayed. Compounding these software issues, Meta also canceled its most advanced in-house AI training chip project, a critical part of its plan to reduce reliance on Nvidia. This failure has triggered a panic-buying spree. In February 2026, Meta announced a capital expenditure budget of $115-$135 billion, nearly double the previous year's. Within ten days, it signed massive, multi-year chip deals: a multi-billion dollar agreement with Nvid...

Written by: Ada, Shenchao TechFlow

Ruoming Pang hadn't even warmed his seat at Meta before he left.

In July 2025, Mark Zuckerberg snatched this most sought-after Chinese engineer in the AI infrastructure field from Apple's grasp with a multi-year compensation package totaling over $200 million. Pang was placed into Meta's Superintelligence Lab, responsible for building the infrastructure for the next generation of AI models.

Seven months later, OpenAI poached him.

According to The Information, OpenAI waged a recruitment campaign for Pang that lasted several months. Although Pang had told colleagues he was "very happy working at Meta," he ultimately chose to leave. Bloomberg reported that his compensation package at Meta was tied to milestones, and leaving early meant forfeiting the majority of unvested equity.

$200 million couldn't buy 7 months of loyalty.

This isn't a simple story of a job change.

One Person's Departure, A Signal for Many

Pang wasn't the first to leave.

Last week, Mat Velloso, Product Lead for the Developer Platform at Meta's Superintelligence Lab, also announced his departure. He had left Google DeepMind to join Meta in July last year and stayed for less than 8 months. Going further back, in November 2025, Turing Award winner and Chief AI Scientist Yann LeCun, who had been at Meta for 12 years, announced his departure to start a company, working on the "world model" he had long championed. Geoffrey Hinton's key disciple and Meta's VP of Generative AI Research, Russ Salakhutdinov, also recently announced his exit.

To understand the AI brain drain at Meta, one must first understand how damaging the Llama 4 incident was.

In April 2025, Meta proudly launched the Llama 4 series Scout and Maverick models. The official paper data was spectacular, claiming to comprehensively suppress GPT-4.5 and Claude Sonnet 3.7 in core benchmark tests like MATH-500 and GPQA Diamond.

However, this flagship model, which carried Meta's ambitions, quickly revealed its true colors in third-party independent blind tests by the open-source community, showing a cliff-like gap between its actual generalization and reasoning capabilities and its advertised performance. Faced with strong community skepticism, Chief AI Scientist Yann LeCun eventually admitted that the team had "used different model versions to run different test sets to optimize the final score" during the testing phase.

In the rigorous world of AI academia and engineering, this crosses an unforgivable red line. In other words, the team trained Llama 4 to be a "small-town exam crammer" only good at past exam papers, not a true "top student" with cutting-edge intelligence. For math tests, they showed the math whiz version; for programming tests, they showed the programming whiz version. Each individual test looked strong, but it wasn't the same model.

In AI academia, this is called "cherry-picking"; in the world of exam-oriented education, it's called "having a ringer take the test."

For Meta, which had always prided itself as the "beacon of open source," this scandal directly destroyed the most valuable asset of trust in its developer ecosystem. The direct cost was that Zuckerberg "completely lost confidence" in the engineering integrity of the original GenAI team, triggering a subsequent series of parachuting in executives and sidelining the core infrastructure department.

He spent $14.3 to $15 billion to acquire a 49% stake in the data annotation company Scale AI, parachuting the 28-year-old Scale AI CEO Alexandr Wang into the role of Meta's Chief AI Officer and establishing the Meta Superintelligence Lab (MSL). Turing Award winner LeCun had to report to this 28-year-old in the new structure. In October, Meta cut about 600 positions in MSL, including members of the FAIR research department created by LeCun.

And the flagship model Llama 4 Behemoth, originally planned for release in the summer of 2025, was repeatedly delayed, pushed from summer to fall, and eventually shelved indefinitely.

Meta instead began developing the next-generation text model codenamed "Avocado" and an image/video model codenamed "Mango." According to reports, Avocado's goal is to compete with GPT-5 and Gemini 3 Ultra. Originally scheduled for delivery by the end of 2025, it was postponed to Q1 2026 due to failed performance tests and training optimization. Meta is considering releasing it as a closed-source model, abandoning the open-source tradition of the Llama series.

Meta made two fatal mistakes with its AI models. First, benchmark fraud, which directly destroyed the trust of the community. Second, forcing a basic research department like FAIR, which requires a decade of grinding, into a product organization chasing quarterly KPIs. These two things combined are the root cause of the current brain drain.

In-House Chips: Another Broken Leg

People are fleeing, and chips are also a problem.

According to The Information, Meta last week canceled its project to develop its most advanced in-house AI training chip.

Meta's in-house chip plan is called MTIA (Meta Training and Inference Accelerator). The company's initial roadmap was ambitious: MTIA v4 codenamed "Santa Barbara," v5 "Olympus," and v6 "Universal Core" were planned for delivery between 2026 and 2028. Olympus was designed as Meta's first chip based on a 2nm chiplet architecture, aiming to cover both high-end model training and real-time inference, ultimately replacing Nvidia's role in Meta's training clusters.

Now, this most advanced training chip has been canceled.

Meta hasn't made no progress; MTIA has had some success on the inference side. The inference chip MTIA v3, codenamed "Iris," has been deployed on a large scale in Meta's data centers, mainly for Facebook Reels and Instagram's recommendation systems, reportedly reducing the total cost of ownership by 40% to 44%. But inference and training are two different things. Inference is running the model; training is teaching the model. Meta can make its own inference chips, but it can't build training chips that can compete head-on with Nvidia.

This isn't the first time historically. In 2022, Meta attempted to develop its own inference chip. After failing in a small-scale deployment, it directly abandoned the project and placed a large order with Nvidia.

The setback in developing its own chips directly accelerated Meta's external buying spree.

$135 Billion in Panic Buying

In January 2026, Meta announced its capital expenditure budget for the year was $115 to $135 billion, almost double last year's $72.2 billion. The bulk of this money will be spent on chips.

Within 10 days, three major deals landed:

February 17: Meta signed a multi-year, cross-generational strategic cooperation agreement with Nvidia. Meta will deploy "millions" of Nvidia Blackwell and next-generation Vera Rubin GPUs, plus standalone Grace CPUs. Analysts estimate the deal is worth tens of billions of dollars. Meta became the first supercomputing customer globally to deploy Nvidia's standalone Grace CPUs on a large scale.

February 24: Meta signed a multi-year chip deal with AMD worth $60 to $100 billion. Meta will purchase AMD's latest MI450 series GPUs and sixth-generation EPYC CPUs. As part of the deal, AMD issued warrants to Meta for up to 160 million shares of common stock, representing about 10% of AMD, at a price of $0.01 per share, vesting in batches based on delivery milestones.

February 26: According to The Information, Meta signed a multi-year deal with Google worth billions of dollars to rent Google Cloud's TPU chips to train and run its next-generation large language model. Simultaneously, the two sides are also discussing Meta directly purchasing TPUs starting in 2027 for deployment in its own data centers.

A social media company placed orders potentially totaling over a hundred billion dollars with three chip suppliers in 10 days.

This isn't diversification. This is panic buying.

The Three Layers of Logic Behind the Computing Power Panic

Why is Meta in such a hurry?

First, in-house chips are no longer an option. Canceling the most advanced training chip project means Meta must rely on external purchases to meet its AI training needs for the foreseeable future. MTIA chips on the inference end can handle mature businesses like recommendation systems, but training cutting-edge models like Avocado, which aims to rival GPT-5, requires Nvidia or hardware of equivalent grade.

Second, competitors won't wait. OpenAI has secured massive resources from Microsoft, SoftBank, to UAE sovereign wealth funds. Anthropic has locked in supplies of 1 million TPU and Trainium chips each from Google and Amazon. Google's Gemini 3 was trained entirely on TPUs. If Meta can't secure enough computing power, it won't even keep its ticket to the race.

Third, and perhaps most fundamentally, Zuckerberg needs to use "purchasing power" to compensate for the lack of "R&D power." The Llama 4 debacle, the loss of core talent, and the setback in developing its own chips—these three things combined have made Meta's AI narrative fragile in front of Wall Street. Signing major deals with Nvidia, AMD, and Google at least sends a signal: We have money, we are buying, we haven't given up.

Meta's current strategy is: if you can't figure out the software, smash the hardware; if you can't keep the people, buy the chips. But the AI race isn't a game you can win by writing checks. Computing power is a necessary condition, not a sufficient one. Without a top-tier model team and a clear technical roadmap, even the most chips are just expensive inventory sitting in warehouses.

The Buyer's Dilemma

Looking back at Meta's three deals in February, an interesting detail was missed by most.

Meta is buying current Blackwell and future Vera Rubin from Nvidia; the deal with AMD is for MI450 and future MI455X; it's renting current Ironwood TPUs from Google, planning to buy them directly next year.

Three suppliers, three completely different hardware architectures and software ecosystems.

This means Meta will have to juggle between three截然不同的底层生态: Nvidia's CUDA, AMD's ROCm, and Google's XLA/JAX. A multi-vendor strategy can分散供应链风险 and压低硬件采购溢价, but it will bring exponentially increasing engineering complexity.

This is Meta's most fatal soft spot right now. Making a trillion-parameter model train efficiently on these three sets of hardware with completely different underlying programming models requires not just engineers who understand CUDA, but architects who can build cross-platform training frameworks from scratch.

There are probably no more than 100 such people in the world. Ruoming Pang was one of them.

Spending $100 billion to buy the world's most complex hardware combination, while simultaneously losing the brains that can驾驭 these hardware—this is the most surreal picture in Zuckerberg's grand gamble.

Zuckerberg's Gamble

Zooming out a bit, Zuckerberg's operational path in AI over the past 18 months is strikingly similar to his rhythm when he went All In on the metaverse:

See the trend, invest heavily, recruit aggressively, encounter setbacks, pivot strategy abruptly, invest heavily again.

2021 to 2023 was the metaverse, resulting in annual losses of tens of billions, and the stock price eventually falling from $380 to $88. 2024 to 2026 is AI,同样是不计代价地砸钱 (investing regardless of cost), frequent organizational reshuffles, and the same narrative of "trust me, I have a vision."

The difference is that this time, the AI trend is indeed much more substantial than the metaverse. And Meta has money to burn; its advertising business generates ample cash flow. In Q4 2025, Meta's revenue was $59.9 billion, a year-on-year increase of 24%.

The problem is: Money can buy chips, buy computing power, even people sitting in seats, but it can't buy the people who stay.

Pang chose OpenAI, Russ Salakhutdinov chose to leave, LeCun chose to start his own company.

Zuckerberg's current bet is that by buying enough chips, building enough large data centers, and spending enough money, he will eventually find or cultivate the people who can use these resources.

This bet might work. Meta is, after all, one of the richest tech companies in the world, with over $100 billion in operating cash flow as its sturdiest moat. From OpenAI to Anthropic, from Google to other competitors, Meta is continuously poaching people. According to a Qbitai report, nearly 40% of the 44 people in Meta's Superintelligence team are from OpenAI.

But the残酷之处 (cruelty) of the AI race is that computing power reserves, talent lists, and model performance are all public. The Llama 4 benchmark fraud incident proved that in this industry, you can't maintain a lead with PPTs and PR.

The market ultimately only recognizes one thing: Is your model good enough?

Position on the Food Chain

As the AI arms race enters 2026, the food chain排序 (ranking) has become初步清晰 (initially clear):

At the top are OpenAI and Google. OpenAI has the strongest models, the largest user base, and the most aggressive financing. Google has complete vertical integration with its own chips, its own models, and its own cloud infrastructure. Anthropic follows closely behind, securing a spot in the first tier with the product strength of its Claude model and dual-line computing power supply from Google and Amazon.

Meta? It has spent the most money, signed the most chip contracts, done the most frequent organizational reshuffles, but so far, it has not produced a cutting-edge model that can convince the market.

Meta's AI story is somewhat like Yahoo in 2005. At that time, Yahoo was also one of the richest internet companies, also疯狂收购和砸钱 (acquiring and spending frantically), but it just couldn't make a search engine as good as Google's. Money isn't omnipotent. Zuckerberg needs to figure out what Meta really wants to do in AI, not just buy whatever is hot.

Of course, it's still too early to write Meta's obituary. 3.58 billion monthly active users, $59.9 billion in quarterly revenue, the world's largest social dataset—these are assets that any competitor would find difficult to replicate.

If the next-generation model codenamed Avocado can be delivered as scheduled in 2026 and return to the first tier, all of Zuckerberg's spending and restructuring will be packaged as "the strategic courage to turn the tide." But if it underperforms again, then this $135 billion will only have bought warehouses of silicon wafers that heat up when powered on.

After all, the AI arms race in Silicon Valley has no shortage of super buyers waving checks. What's lacking are the people who know how to use this computing power to forge the future.

İlgili Sorular

QWhy did Pang Ruoming leave Meta after only 7 months, and what does his departure signify about Meta's AI talent retention?

APang Ruoming left Meta to join OpenAI after a months-long recruitment effort, despite initially expressing satisfaction with his role. His departure, along with other key figures like Yann LeCun and Russ Salakhutdinov, signals a deeper issue within Meta's AI division, including loss of trust from the Llama 4 benchmark scandal, organizational restructuring, and a shift away from foundational research towards product-driven goals, leading to talent disillusionment.

QWhat was the consequences of Meta's Llama 4 benchmark manipulation scandal?

AThe Llama 4 benchmark manipulation, where Meta used different model versions to optimize test scores, severely damaged its credibility in the AI community. This led to a loss of trust, internal restructuring including the appointment of a new Chief AI Officer, layoffs, the indefinite postponement of Llama 4 Behemoth, and a strategic pivot towards developing new models like 'Avocado' and 'Mango', potentially as closed-source projects.

QWhy did Meta cancel its most advanced in-house AI training chip project, and how does this impact its strategy?

AMeta canceled its advanced MTIA training chip project (codenamed 'Olympus') due to development challenges, as it could not create a chip competitive with NVIDIA's for training cutting-edge models. This failure forced Meta into a 'panic buying' spree, signing multi-year, multi-billion dollar deals with NVIDIA, AMD, and Google Cloud for GPUs, CPUs, and TPUs to secure the necessary compute power for AI training, revealing a reliance on external suppliers.

QWhat is the fundamental dilemma behind Meta's strategy of spending over $100 billion on compute hardware while losing key AI talent?

AMeta's dilemma is that while it can use its massive cash flow to purchase immense compute power (e.g., deals with NVIDIA, AMD, Google), it is simultaneously losing the rare, top-tier AI engineers and architects needed to effectively utilize this complex, multi-architecture hardware. This creates a situation where expensive hardware may become underutilized 'inventory' without the human expertise to build and train state-of-the-art models on it.

QHow does Meta's current position in the AI 'food chain' compare to its competitors like OpenAI and Google?

AMeta is not in the top tier of the AI food chain. OpenAI and Google lead with the strongest models, vast user bases, and vertical integration (e.g., Google's TPUs). Anthropic is also a strong contender. Despite spending the most money on compute and undergoing frequent reorganizations, Meta has yet to deliver a frontier model that convincingly competes with these leaders, placing it in a position where it must prove itself with its upcoming 'Avocado' model to be considered a top player.

İlgili Okumalar

You Bet on the News, the Pros Read the Rules: The True Cognitive Gap in Losing Money on Polymarket

The article explains that the key to profiting on Polymarket, a prediction market platform, lies not just predicting real-world events correctly, but in meticulously understanding the specific rules that govern how each market will be resolved. It illustrates this with examples, such as a market on Venezuela's 2026 leader, where the official rules defining "officially holds" the office overruled the intuitive answer of who was in practical control. Other examples include debates over the definition of a "token" or what constitutes an "agreement." The core argument is that a "reality vs. rules" gap creates pricing discrepancies that savvy traders ("车头" or "whales") exploit. The platform has a formal dispute resolution process managed by UMA token holders to settle ambiguous outcomes. This process involves proposal submission, a challenge window, a discussion period, and a final vote. However, the article highlights a critical flaw in this system compared to a traditional court: the lack of separation between the arbiters (UMA voters) and the interested parties (traders with financial stakes in the outcome). This conflict of interest undermines the discussion phase, leads to herd mentality, and results in opaque final decisions without explanatory rulings. Consequently, the system lacks a body of precedent, making it difficult for users to learn from past disputes. The ultimate takeaway is that success on Polymarket requires a lawyer-like scrutiny of the rules to identify and capitalize on the cognitive gap between how events appear and how they are contractually defined for settlement.

marsbit15 dk önce

You Bet on the News, the Pros Read the Rules: The True Cognitive Gap in Losing Money on Polymarket

marsbit15 dk önce

Will the Fed Still Cut Interest Rates? Tonight's Data Is Crucial

The core debate surrounding the Federal Reserve's potential interest rate cuts is intensifying amid geopolitical conflict and rebounding inflation. The key question is whether high energy prices will cause persistent inflation or weaken consumer demand enough to force the Fed to cut rates. Citigroup presents a bullish case for cuts, arguing that oil supply disruptions from the Strait of Hormuz are temporary and will not lead to lasting inflationary pressure. They point to receding bond yields and oil prices as evidence the market is pricing in a short-lived shock. Citi's data also shows tightening financial conditions, a stabilizing labor market, and healthy tax returns, supporting their view that the path to lower rates remains open. Conversely, Deutsche Bank offers a starkly contrasting, more hawkish outlook. They argue the Fed's current policy is already neutral and expect rates to remain unchanged indefinitely. Their view is based on stalled disinflation progress and a shift toward more hawkish rhetoric from key Fed officials like Waller, who cited risks from prolonged Middle East conflict and tariffs. Other officials, including Williams and Hammack, signaled rates would likely stay on hold for a "considerable time." The market pricing has shifted dramatically, now forecasting zero cuts in 2026. The imminent release of the March retail sales "control group" data is highlighted as a critical test. This metric, which excludes gas station sales, will reveal if high gasoline prices are eroding consumer spending in other areas. A weak reading could support the case for imminent rate cuts, while a strong one would bolster the argument for the Fed to hold steady. This data is pivotal for determining the near-term policy path.

marsbit36 dk önce

Will the Fed Still Cut Interest Rates? Tonight's Data Is Crucial

marsbit36 dk önce

The Second Half of Macro Influencer Fu Peng's Career

Fu Peng, a prominent Chinese macroeconomist and former chief economist of Northeast Securities, has joined Hong Kong-based digital asset management firm Bitfire Group (formerly New Huo Group) as its chief economist. This move, announced in April 2026, triggered an 11% surge in Bitfire's stock price. Fu, known for his accessible macroeconomic commentary and large social media following, will focus on integrating digital assets into global asset allocation frameworks, particularly combining FICC (fixed income, currencies, and commodities) with cryptocurrencies for institutional clients. His career includes roles at Lehman Brothers and Solomon International, with significant influence gained through public communication. However, in late 2024, Fu faced temporary social media bans after a controversial private speech at HSBC on China's economic challenges, though he denied regulatory sanctions. He later left Northeast Securities citing health reasons. Bitfire, a licensed virtual asset manager serving high-net-worth clients, seeks to build trust and attract traditional capital through Fu’s expertise and credibility. The partnership represents a strategic shift for both: Fu enters the crypto sector after a traditional finance peak, while Bitfire aims to leverage his macro framework for institutional adoption. Outcomes remain uncertain regarding capital inflows and compatibility within corporate structure.

marsbit1 saat önce

The Second Half of Macro Influencer Fu Peng's Career

marsbit1 saat önce

İşlemler

Spot
Futures

Popüler Makaleler

PEOPLE Nasıl Satın Alınır

HTX.com’a hoş geldiniz! ConstitutionDAO (PEOPLE) satın alma işlemlerini basit ve kullanışlı bir hâle getirdik. Adım adım açıkladığımız rehberimizi takip ederek kripto yolculuğunuza başlayın. 1. Adım: HTX Hesabınızı OluşturunHTX'te ücretsiz bir hesap açmak için e-posta adresinizi veya telefon numaranızı kullanın. Sorunsuzca kaydolun ve tüm özelliklerin kilidini açın. Hesabımı Aç2. Adım: Kripto Satın Al Bölümüne Gidin ve Ödeme Yönteminizi SeçinKredi/Banka Kartı: Visa veya Mastercard'ınızı kullanarak anında ConstitutionDAO (PEOPLE) satın alın.Bakiye: Sorunsuz bir şekilde işlem yapmak için HTX hesap bakiyenizdeki fonları kullanın.Üçüncü Taraflar: Kullanımı kolaylaştırmak için Google Pay ve Apple Pay gibi popüler ödeme yöntemlerini ekledik.P2P: HTX'teki diğer kullanıcılarla doğrudan işlem yapın.Borsa Dışı (OTC): Yatırımcılar için kişiye özel hizmetler ve rekabetçi döviz kurları sunuyoruz.3. Adım: ConstitutionDAO (PEOPLE) Varlıklarınızı SaklayınConstitutionDAO (PEOPLE) satın aldıktan sonra HTX hesabınızda saklayın. Alternatif olarak, blok zinciri transferi yoluyla başka bir yere gönderebilir veya diğer kripto para birimlerini takas etmek için kullanabilirsiniz.4. Adım: ConstitutionDAO (PEOPLE) Varlıklarınızla İşlem YapınHTX'in spot piyasasında ConstitutionDAO (PEOPLE) ile kolayca işlemler yapın.Hesabınıza erişin, işlem çiftinizi seçin, işlemlerinizi gerçekleştirin ve gerçek zamanlı olarak izleyin. Hem yeni başlayanlar hem de deneyimli yatırımcılar için kullanıcı dostu bir deneyim sunuyoruz.

373 Toplam GörüntülenmeYayınlanma 2024.12.12Güncellenme 2025.03.21

PEOPLE Nasıl Satın Alınır

Tartışmalar

HTX Topluluğuna hoş geldiniz. Burada, en son platform gelişmeleri hakkında bilgi sahibi olabilir ve profesyonel piyasa görüşlerine erişebilirsiniz. Kullanıcıların PEOPLE (PEOPLE) fiyatı hakkındaki görüşleri aşağıda sunulmaktadır.

活动图片