豆包的价格战,打出了定价权

marsbitPublicado a 2026-05-07Actualizado a 2026-05-07

文 | Next 趋势,作者 | 方远,编辑 | 小雨

2024年5月,火山引擎总裁谭待在大会上宣布,豆包大模型的定价大幅低于行业价格,仅为0.0008元/千tokens,引得台下惊呼。

两年后,豆包披露了新的付费订阅计划,推出三档付费版本,最高档连续包月价格为500元。

曾经在价格战中最激进的玩家,率先推出了新的收费模式。

而竞争对手们的第一反应并非是跟进,而是强调自身仍然免费。

 低价清场

2026年5月4日,豆包在App Store最新的版本页面中新增了订阅选项,推出三档付费版本,其中标准版定价68元,最高档位达500元。

消息一经发出,直接登上微博热搜第一。

评论区有人说收费就卸载,有人称豆包在割韭菜,还有人担心免费版会遭到严重降智。

两年前,谭待在火山引擎FORCE原动力大会上宣布,豆包通用模型pro-32k版,推理输入价格为0.0008元/千tokens,仅为行业均价的一百五十分之一。

彼时台下的人算了一笔账,按这个价格计算,1块钱能买到125万tokens的输入量,相当于三本《三国演义》。

这次定价,直接掀了行业的桌子。

阿里云将通义千问旗舰模型Qwen-Long的输入价格从0.02元下调至0.0005元,降幅达97%。

百度则更干脆,两款主力模型Speed和Lite直接免费。

价格战随之展开,其伤害也迅速在行业内蔓延。

创业公司的生存空间被严重挤压,要么向下深耕垂直行业,要么向上构建应用生态,通用层的中间地带已不复存在。

百度智能云事业群总裁沈抖曾表示:“国内大模型行业的恶意价格战,导致整体创收与国外相比差了多个数量级。”

如今,行业发生了剧烈变化。

国联民生证券测算显示,中国整体日均Token消耗从2024年初的1000亿级,飙升至2026年2月份的180万亿级。

需求爆发下,智谱AI、腾讯云等相继发布涨价通知,部分产品涨幅超400%。

价格战的底层逻辑本是烧钱换市场,但当市场规模到顶时,烧钱就不是投资,而是浪费。

谭待曾在朋友圈发文回应大模型价格战一事,强调豆包1.5Pro的预训练远低于国内其他模型,在当前的价格下仍有可观的毛利。他在此前接受采访时也曾强调:火山引擎从不做赔钱换市场的事。

若情况属实,那过去两年的价格战,豆包就不是在赔本赚吆喝,而是凭借成本优势在碾压对手。

 走不掉的三亿人

QuestMobile 2026年一季报公布的数据显示,豆包的月活跃用户数达到3.45亿,排名第二的千问为1.66亿,两者相差超过1.8亿。

同时,豆包单季度的新增用户量达到了1.01亿,占了全行业新增用户的80%。

两年时间,豆包月活从2600万到3.45亿,翻了13倍,远超所有竞品。

当豆包几乎成了整个品类的代名词时,它的定价便等同于是在制定新的行业规则。

但值得注意的是,3.45亿的用户量,并不都是直接的收费对象。

豆包的目标不是让每个人都付费,而是在庞大的用户规模下,即使只有一小部分人选择付费,也足以支撑其发展需求。

除了用户规模庞大,豆包的用户留存同样表现亮眼。

相关数据显示,豆包的30日留存率达44.5%,远高于千问的23.5%和元宝的30.1%;月人均使用次数豆包为54.8次,是千问19.8次的三倍;活跃率方面,豆包也以33.5%领先于千问的17.1%。

每一项都是碾压级的差距。

高留存率意味着用户在面临收费时的转换成本极高,用户虽然喊着收费就卸载,但真卸载之后,又该用什么替代呢?

截至2026年3月,豆包的日均Token使用量已突破120万亿,较2024年5月增长了1000倍。

但1000倍的用量增长背后是1000倍的算力账单。

字节2025年资本开支超1500亿元,其中约900亿元投向AI算力;2026年计划投入1600亿元,其中850亿元将用于AI芯片采购。

高额支出带来的是2025年净利润同比下滑超70%。

对此,抖音集团副总裁李亮澄清,若按经营利润率口径计算,实际只是小幅下滑,但他也坦言,抖音电商增速放缓与新兴业务相关投入加大确实带来了压力。

豆包的免费模式并非不可持续,而是每多运营一天,都会挤压核心业务的利润空间。

2026年一季度,全球AI应用月活跃用户数突破27亿,其中中国贡献了超50%的新增量,国内AI应用月活规模已达8.51亿。

可以说,当前具备AI使用条件的用户,基本都已成为AI应用的使用者。

当客已经获完了,免费就从投资变成了成本。

 规则重写 

豆包宣布收费后,竞品的反应更为激烈。

DeepSeek率先宣布其V4-Pro模型API限时享2.5折优惠;阿里千问推出补贴活动,以请用户喝奶茶的形式吸引关注;元宝与文心一言则直接打出了免费使用的旗号。

豆包一张价目表,把所有竞品都推入了两难境地:收费,无异于将用户拱手让给对手;但维持免费,又等于承认自己仍处于用户教育阶段。

豆包标准版68元/月的价格不是拍脑袋定的。

腾讯研究院2025年的调研数据显示,约四分之三的AI用户已付费或愿意付费,而在已付费用户中,超55%的人月度支出低于100元。68元的定价,精准契合了大多数用户的心理承受区间。

这套定价有没有先例?有。

Kimi已为整个行业完成了一次验证。K2.5大模型上线不足一个月,累计收入便超过2025年全年水平,全球付费用户环比增长4倍。而Stripe的数据更为惊人,Kimi个人订阅用户的订单数环比激增8280%。

月之暗面估值冲到100-120亿美元,创下中国公司从成立到十角兽的最快纪录。

中国AI用户不是不付费,是之前没有产品配得上那个价格。

而豆包3.45亿的月活跃用户,即便仅1%完成转化,其规模也将远超Kimi的全部付费用户。

更早的先例可追溯至十年前。2016年,微信支付率先对提现业务收取0.1%的手续费。彼时,支付宝在移动支付领域仍处于领先地位,外界普遍认为,微信支付尚未取得绝对优势便率先收费,无异于主动让出了竞争先机。

结果没有用户因此放弃微信支付。微信支付的核心场景,如发红包、AA付款、转账还款等,具有极强的黏性,只能在其生态内完成。

用户的退出成本极高,高到收费根本无法构成威胁。

如今,豆包重演了同一幕。用户量最大的人第一个收费,定义的不是自己的利润,是整个品类的价格基准。

摩根士丹利已经在算账了。按付费转化率0.3%至3%、月活3.45亿至5.25亿测算,豆包年化订阅收入区间为1亿至15亿美元,中性情景下对应收入约为4.26亿至6.84亿美元。

不过,华尔街的乐观建立在一个前提上:国内C端工具产品的年均续费率能做到30%就已是行业顶级水平,而海外同类产品的续费率普遍在60%以上。豆包能不能把首月冲动付费变成长期留存,还没有答案。

同一时间,大洋彼岸的OpenAI走了另一条路。

ChatGPT的周活跃用户达5亿,付费用户仅2500万,付费率约5%,也就是说每20个用户中,仅有1个愿意付费。

OpenAI的解决方案是推出低价版ChatGPT Go,定价8美元/月,同时植入广告,设定2026年广告收入目标为25亿美元。通过广告收入补贴免费用户,以此将付费门槛降至最低。

豆包则选择了一条相反的路径:它不依赖广告来稀释用户体验,而是直接让重度用户为高阶功能付费。

两条路径的分歧背后,其实都源于纯免费模式难以为继的判断。只不过,一个选择用广告作为兜底手段,另一个则选择用订阅服务来支撑运营。

摩根士丹利给出了更明确的定性:中国消费者AI使用习惯的培育阶段已基本完成,行业正从用户补贴转向商业可持续性发展。

若豆包的付费模式得到市场验证,通义、元宝等仍维持免费策略的竞争对手,将面临跟进收费还是坚守免费的艰难抉择。

两年前,豆包以0.0008元/千tokens的定价打破了行业原有的定价体系;两年后,它又通过68元至500元的价格区间重新构建了这一体系。

那个曾打破规则的参与者,如今成了重新确立规则的主导者。

第一个开出价目表的人,定义的并非自身利润率,而是让整个行业从此有了明确的价格基准。

Lecturas Relacionadas

Gensyn AI: Don't Let AI Repeat the Mistakes of the Internet

In recent months, the rapid growth of the AI industry has attracted significant talent from the crypto sector. A persistent question among researchers intersecting both fields is whether blockchain can become a foundational part of AI infrastructure. While many previous AI and Crypto projects focused on application layers (like AI Agents, on-chain reasoning, data markets, and compute rentals), few achieved viable commercial models. Gensyn differentiates itself by targeting the most critical and expensive layer of AI: model training. Gensyn aims to organize globally distributed GPU resources into an open AI training network. Developers can submit training tasks, nodes provide computational power, and the network verifies results while distributing incentives. The core issue addressed is not decentralization for its own sake, but the increasing centralization of compute power among tech giants. In the era of large models, access to GPUs (like the H100) has become a decisive bottleneck, dictating the pace of AI development. Major AI companies are heavily dependent on large cloud providers for compute resources. Gensyn's approach is significant for several reasons: 1) It operates at the core infrastructure layer (model training), the most resource-intensive and technically demanding part of the AI value chain. 2) It proposes a more open, collaborative model for compute, potentially increasing resource utilization by dynamically pooling idle GPUs, similar to early cloud computing logic. 3) Its technical moat lies in solving complex challenges like verifying training results, ensuring node honesty, and maintaining reliability in a distributed environment—making it more of a deep-tech infrastructure company. 4) It targets a validated, high-growth market with genuine demand, rather than pursuing blockchain integration without purpose. Ultimately, the boundaries between Crypto and AI are blurring. AI requires global resource coordination, incentive mechanisms, and collaborative systems—areas where crypto-native solutions excel. Gensyn represents a step toward making advanced training capabilities more accessible and collaborative, moving beyond a niche controlled by a few giants. If successful, it could evolve into a fundamental piece of AI infrastructure, where the most enduring value in the AI era is often created.

marsbitHace 9 hora(s)

Gensyn AI: Don't Let AI Repeat the Mistakes of the Internet

marsbitHace 9 hora(s)

Why is China's AI Developing So Fast? The Answer Lies Inside the Labs

A US researcher's visit to China's top AI labs reveals distinct cultural and organizational factors driving China's rapid AI development. While talent, data, and compute are similar to the West, Chinese labs excel through a pragmatic, execution-focused culture: less emphasis on individual stardom and conceptual debate, and more on teamwork, engineering optimization, and mastering the full tech stack. A key advantage is the integration of young students and researchers who approach model-building with fresh perspectives and low ego, prioritizing collective progress over personal credit. This contrasts with the US culture of self-promotion and "star scientist" narratives. Chinese labs also exhibit a strong "build, don't buy" mentality, preferring to develop core capabilities—like data pipelines and environments—in-house rather than relying on external services. The ecosystem feels more collaborative than tribal, with mutual respect among labs. While government support exists, its scale is unclear, and technical decisions appear driven by labs, not state mandates. Chinese companies across sectors, from platforms to consumer tech, are building their own foundational models to control their tech destiny, reflecting a broader cultural drive for technological sovereignty. Demand for AI is emerging, with spending patterns potentially mirroring cloud infrastructure more than traditional SaaS. Despite challenges like a less mature data industry and GPU shortages, Chinese labs are propelled by vast talent, rapid iteration, and deep integration with the open-source community. The competition is evolving beyond a pure model race into a contest of organizational execution, developer ecosystems, and industrial pragmatism.

marsbitHace 10 hora(s)

Why is China's AI Developing So Fast? The Answer Lies Inside the Labs

marsbitHace 10 hora(s)

3 Years, 5 Times: The Rebirth of a Century-Old Glass Factory

Corning, a 175-year-old glass company, is experiencing a dramatic revival as a key player in AI infrastructure, driven by surging demand for high-performance optical fiber in data centers. AI data centers require vastly more fiber than traditional ones—5 to 10 times as much per rack—to handle high-speed data transmission between GPUs. This structural demand shift, coupled with supply constraints from the lengthy expansion cycle for fiber preforms, has created a significant supply-demand gap. Nvidia has invested in Corning, along with Lumentum and Coherent, in a $4.5 billion total commitment to secure the optical supply chain for AI. Corning's competitive edge lies in its expertise in producing ultra-low-loss, high-density, and bend-resistant specialty fiber, which is critical for 800G+ and future 1.6T data rates. Its deep involvement in co-packaged optics (CPO) with partners like Nvidia further solidifies its position. While not the largest fiber manufacturer globally, Corning's revenue from enterprise/data center clients now exceeds 40% of its optical communications sales, and it has secured multi-year supply agreements with major hyperscalers including Meta and Nvidia. Financially, Corning's optical communications revenue has surged, doubling from $1.3 billion in 2023 to over $3 billion in 2025. Its stock price has risen nearly 6-fold since late 2023. Key future catalysts include the rollout of Nvidia's CPO products and the scale of undisclosed customer agreements. However, risks include high current valuations and potential disruption from next-generation technologies like hollow-core fiber. The company's long-term bet on light over electricity, maintained even through the telecom bubble crash, is now being validated by the AI boom.

marsbitHace 11 hora(s)

3 Years, 5 Times: The Rebirth of a Century-Old Glass Factory

marsbitHace 11 hora(s)

Trading

Spot
Futuros
活动图片