450亿美元估值,梁文锋「下山」

marsbitPublicado em 2026-05-08Última atualização em 2026-05-08

文 | 节点财经,作者 | 零度 

450亿美元。

2026年5月初,这个数字像一记重锤,打破创投圈近三年的沉闷。

两周前,圈子里还在传阿里和腾讯要掏200亿美元给DeepSeek“续费”。谁知风向转瞬即逝,根据《金融时报》的信源,国家大基金(国家集成电路产业投资基金)正打算亲自下场主导这一轮融资,估值直接定在了450亿美元。

半个月,估值翻倍。

在算力吃紧、融资寒冬、甚至AI退潮论都快被说烂了的今天,DeepSeek以一己之力,宣告了谁才是中国AI圈当之无愧的“头牌”。但《节点财经》认为,这背后,绝不只是几家巨头在账面上玩数字游戏,而是一次关于中国AI底层逻辑的推倒重来。

梁文锋的“下山”

要聊DeepSeek,绕不开那个叫梁文锋的男人。

在很长一段时间里,梁文锋在圈内像个隐士。背靠幻方量化这个“提款机”,他早期的态度硬得像块石头:不拿外部融资,不搞花里胡哨的发布会,甚至不怎么爱搭理投资人。

这种“老子有钱,只管闷头搞技术”的底气,让DeepSeek避开了浮躁,硬生生靠着算法优化,在国际上杀出了一条血路。

那么,为什么一直拒绝融资的梁文锋,在2026年选择了拥抱资本?

《节点财经》认为,一个直接原因是,AI已经从“脑力活”变成了“重工业”。

早期的DeepSeek靠的是灵气,是用精妙的算法在英伟达的芯片间跳舞。但到了2026年,大模型竞争已经进入了L4级智能体(MasterAgent)的博弈期,这不仅是算法的对决,更是电力、算力、甚至国家级资源调度权的对决。幻方再有钱,也只是一个顶级的“私家作坊”,它给不了DeepSeek那种能够撬动整个半导体产业链的“通行证”。

梁文锋的“下山”,本质上是一种战略性妥协。他意识到,在现在的国际环境下,想做一个“世外高人”已经不可能了。不入局,就会在算力封锁的铁幕下因为缺氧而窒息。

更重要的一层原因是人才。

近期像郭达雅(跳槽字节)、罗福莉(跳槽小米)等一些早期核心技术骨干相继离职。这在圈内引起了不小的震动,甚至有人觉得DeepSeek要“青黄不接”了。

但如果你去看DeepSeek V4发布时的作者名单,数据其实很硬:其核心的研究工程团队大约270人,在研发期间离职的只有10人左右。不到4%的离职率,放在任何一家大模型公司都是奇迹——要知道,OpenAI前两年的核心人才流失率可是超过了25%。

之所以大家觉得“离职潮”凶猛,是因为DeepSeek早期的团队太精简、光芒太盛。当这几十个“天才少年”里的任何一个被大厂用几倍的高薪、更显赫的头衔挖走时,外界都会觉得是地动山摇。

但另一个层面看,人才的争夺竞争确实在加剧。

为什么非得现在融资?其实就是为了给人才一个“确定的价格”。以前DeepSeek不融资,员工手里攥着的期权只是“纸面富贵”,看不见摸不着。现在450亿美元的估值一砸下来,大家突然发现,自己手里的股份是真的值钱了。在被大厂高薪诱惑时,员工会算一笔账:是去大厂拿现钱,还是留在DeepSeek等这个450亿美元的盘子再翻几倍?

另外就是对抗大厂的“钞能力”,字节、腾讯、小米这些巨头挖人,动不动就是股权加奖金。梁文锋需要这轮融资,把DeepSeek从一个“极客实验室”正式变成一个拥有强大资本背书的“超级独角兽”,在人才市场上抢回主动权。

大基金被传“入局”

根据《金融时报》消息显示,国家集成电路产业投资基金(简称“大基金”)正在与DeepSeek洽谈其首轮融资事宜。

大基金下场投一家大模型公司,这在圈子里绝对算是破天荒的头一遭。

截至目前,大基金共成立了三期基金。一期是"战略布局",解决了"有没有"的问题;二期是"精准排雷",着力解决"会不会被卡脖子"的问题,全力攻克设备、材料等上游核心环节。 2024年5月大基金三期基金成立,规模超一二期之和,开始向AI芯片侧扩张。

整体看,以前大基金的钱都是往芯片制造、光刻机、蚀刻机这些硬核硬件里砸的,讲究的是“强基”。现在突然转过头来,要把DeepSeek推向450亿美元的高位,这背后其实是一场精准的战略资源置换。

《节点财经》看来,大基金之所以投他,归根结底是因为DeepSeek手里握着三张能改变这盘大棋走势的“牌”:

首先,它是国产算力的“首席调音师”。

大基金投了那么多年半导体,最大的心结就是国产芯片“好造不好用”。

华为昇腾、寒武纪、壁仞……这些国产AI芯片虽然硬件指标上去了,但在生态和适配上一直被英伟达压着打。很多模型厂商为了图省事,首选还是CUDA生态。

DeepSeek不一样,它是典型的“硬骨头”。它最擅长的事情就是通过极致的算法优化,让原本在软件层面略显生涩的国产芯片,跑出远超标称的实战性能。大基金投DeepSeek,其实是投一个能把国产半导体产业链“带飞”的超级引擎。 只要DeepSeek在国产算力平台上跑通了L4级智能体(MasterAgent),就等于给国产芯片发了一张通向全球市场的“合格证”。

其次,它是打破竞争的“刺头”。

在大模型这个领域,如果跟着OpenAI的屁股后面堆算力、砸美金,那永远只能做个追随者。

DeepSeek之所以被国家队看重,是因为它证明了另一种可能:用更少的钱、更少的卡,办成更大的事。

这种“算法换算力”的非对称打法,非常符合咱们现在被极限施压下的生存逻辑。大基金主导融资,是要把这种“低功耗、高智力”的技术路径确立为国产AI的主流方向,避免大家在算力红海里无效内耗。

第三,它是生态闭环的“最后一块拼图”。

此前咱们的AI产业是各玩各的:大基金投芯片,阿里腾讯搞应用,创业公司做模型。

大基金这次领投,就像是伸出一只手,把散落一地的珍珠串成了项链。

底层: 大基金保障国产芯片的优先级。

中间: DeepSeek提供全栈自研的核心大脑。

顶层: 互联网巨头的场景给模型喂数据、找出口。

这种“国家队保底+极客带头+巨头落地”的组合拳,才是大基金真正想看到的全产业链安全闭环。

所以,大基金投的不是一个会聊天的机器人,而是一个能让中国AI在“算力荒漠”里强行开辟出一片绿洲的软硬一体化指挥官。

450亿美元的估值,买的是一个能让国产半导体产业真正“活”起来的机会。这种机会,在眼下的博弈中,确实是不折不扣的“无价之宝”。

450亿美元背后:不仅是中美较量,更是一场“生态缝合”

很多人喜欢把DeepSeek的崛起解读为中美之间的“博弈”。但《节点财经》看来,这种观点有点浅了。

在2026年这个节点,美国手里攥着B200芯片的霸权,试图把中国AI锁死在“算力贫矿区”。而DeepSeek的站位,恰恰是中国给出的最强硬回应。大基金之所以主导这轮融资,看中的绝不是DeepSeek能写几首诗,而是它的能力。

DeepSeek正在完成一次史无前例的“生态缝合”:

向上,它在“养”算力: 在国产芯片还在艰难爬坡的关键期,DeepSeek用顶级算法优化,硬生生让国产芯片跑出了超越预期的实战效果。既然买不到最顶级的卡,那就用最顶级的逻辑去榨干每一颗国产芯片的残余价值。这不仅是省钱,这是在为国产算力产业“续命”。

向下,它在“造”闭环: 此前,中国的AI生态是断裂的——芯片是芯片,模型是模型,应用是应用。而这轮如果由大基金领投、互联网大厂跟进融资,实际上是完成了一个全链路的合龙:大基金保底层,DeepSeek出脑子,互联网巨头出场景。

这种打法,让原本散落在各处的国产资源,第一次有了一个统一的、能打硬仗的“指挥部”。

在这个450亿美元的局里,有一种很微妙的智慧。

《节点财经》看来,互联网巨头的参与,不再是往日的“跑马圈地”,而更像是一种生态上的互补。大厂手里有亿级的用户入口和企业级场景,那是DeepSeek这种“极客实验室”最缺的实战训练场。而国家队的入场,则给这个充满了不确定性的产业打了一剂最强的强心针。

对于梁文锋来说,这轮融资是一场“成人礼”。他告别了那个纯粹的实验室时代,踏入了最复杂的利益博弈中心。这种转变或许少了点极客的清高,但多了一份在大国博弈中生存下去的坚韧。

450亿美元贵吗?如果只是买一个模型,那贵得离谱;但如果买的是一个在极限封锁下依然能自我进化的AI生态系统,这个价格其实承载的是未来的国运折现。

在2026年的风暴中心,DeepSeek已经拿到了最重的一张筹码。至于未来它能不能真正成为那座抗衡硅谷的“数字长城”,我们要看的,不仅是它的算法深度,更是它在这场资本巨浪中整合资源的能力。

毕竟,在AI这个名为“无限博弈”的游戏里,活下去并变得不可替代,才是唯一的真理。

Leituras Relacionadas

Gensyn AI: Don't Let AI Repeat the Mistakes of the Internet

In recent months, the rapid growth of the AI industry has attracted significant talent from the crypto sector. A persistent question among researchers intersecting both fields is whether blockchain can become a foundational part of AI infrastructure. While many previous AI and Crypto projects focused on application layers (like AI Agents, on-chain reasoning, data markets, and compute rentals), few achieved viable commercial models. Gensyn differentiates itself by targeting the most critical and expensive layer of AI: model training. Gensyn aims to organize globally distributed GPU resources into an open AI training network. Developers can submit training tasks, nodes provide computational power, and the network verifies results while distributing incentives. The core issue addressed is not decentralization for its own sake, but the increasing centralization of compute power among tech giants. In the era of large models, access to GPUs (like the H100) has become a decisive bottleneck, dictating the pace of AI development. Major AI companies are heavily dependent on large cloud providers for compute resources. Gensyn's approach is significant for several reasons: 1) It operates at the core infrastructure layer (model training), the most resource-intensive and technically demanding part of the AI value chain. 2) It proposes a more open, collaborative model for compute, potentially increasing resource utilization by dynamically pooling idle GPUs, similar to early cloud computing logic. 3) Its technical moat lies in solving complex challenges like verifying training results, ensuring node honesty, and maintaining reliability in a distributed environment—making it more of a deep-tech infrastructure company. 4) It targets a validated, high-growth market with genuine demand, rather than pursuing blockchain integration without purpose. Ultimately, the boundaries between Crypto and AI are blurring. AI requires global resource coordination, incentive mechanisms, and collaborative systems—areas where crypto-native solutions excel. Gensyn represents a step toward making advanced training capabilities more accessible and collaborative, moving beyond a niche controlled by a few giants. If successful, it could evolve into a fundamental piece of AI infrastructure, where the most enduring value in the AI era is often created.

marsbitHá 11h

Gensyn AI: Don't Let AI Repeat the Mistakes of the Internet

marsbitHá 11h

Why is China's AI Developing So Fast? The Answer Lies Inside the Labs

A US researcher's visit to China's top AI labs reveals distinct cultural and organizational factors driving China's rapid AI development. While talent, data, and compute are similar to the West, Chinese labs excel through a pragmatic, execution-focused culture: less emphasis on individual stardom and conceptual debate, and more on teamwork, engineering optimization, and mastering the full tech stack. A key advantage is the integration of young students and researchers who approach model-building with fresh perspectives and low ego, prioritizing collective progress over personal credit. This contrasts with the US culture of self-promotion and "star scientist" narratives. Chinese labs also exhibit a strong "build, don't buy" mentality, preferring to develop core capabilities—like data pipelines and environments—in-house rather than relying on external services. The ecosystem feels more collaborative than tribal, with mutual respect among labs. While government support exists, its scale is unclear, and technical decisions appear driven by labs, not state mandates. Chinese companies across sectors, from platforms to consumer tech, are building their own foundational models to control their tech destiny, reflecting a broader cultural drive for technological sovereignty. Demand for AI is emerging, with spending patterns potentially mirroring cloud infrastructure more than traditional SaaS. Despite challenges like a less mature data industry and GPU shortages, Chinese labs are propelled by vast talent, rapid iteration, and deep integration with the open-source community. The competition is evolving beyond a pure model race into a contest of organizational execution, developer ecosystems, and industrial pragmatism.

marsbitHá 13h

Why is China's AI Developing So Fast? The Answer Lies Inside the Labs

marsbitHá 13h

3 Years, 5 Times: The Rebirth of a Century-Old Glass Factory

Corning, a 175-year-old glass company, is experiencing a dramatic revival as a key player in AI infrastructure, driven by surging demand for high-performance optical fiber in data centers. AI data centers require vastly more fiber than traditional ones—5 to 10 times as much per rack—to handle high-speed data transmission between GPUs. This structural demand shift, coupled with supply constraints from the lengthy expansion cycle for fiber preforms, has created a significant supply-demand gap. Nvidia has invested in Corning, along with Lumentum and Coherent, in a $4.5 billion total commitment to secure the optical supply chain for AI. Corning's competitive edge lies in its expertise in producing ultra-low-loss, high-density, and bend-resistant specialty fiber, which is critical for 800G+ and future 1.6T data rates. Its deep involvement in co-packaged optics (CPO) with partners like Nvidia further solidifies its position. While not the largest fiber manufacturer globally, Corning's revenue from enterprise/data center clients now exceeds 40% of its optical communications sales, and it has secured multi-year supply agreements with major hyperscalers including Meta and Nvidia. Financially, Corning's optical communications revenue has surged, doubling from $1.3 billion in 2023 to over $3 billion in 2025. Its stock price has risen nearly 6-fold since late 2023. Key future catalysts include the rollout of Nvidia's CPO products and the scale of undisclosed customer agreements. However, risks include high current valuations and potential disruption from next-generation technologies like hollow-core fiber. The company's long-term bet on light over electricity, maintained even through the telecom bubble crash, is now being validated by the AI boom.

marsbitHá 13h

3 Years, 5 Times: The Rebirth of a Century-Old Glass Factory

marsbitHá 13h

Trading

Spot
Futuros
活动图片