a16z Founder: In the Agent Era, What Truly Matters Has Changed

marsbit发布于2026-04-20更新于2026-04-20

文章摘要

Marc Andreessen, co-founder of a16z, argues that the current AI boom is not an overnight success but the culmination of 80 years of research, now delivering practical results. He emphasizes that this era is defined by the convergence of four key capabilities: large language models (LLMs), reasoning, coding, and agents capable of recursive self-improvement. Andreessen describes the agent architecture—combining an LLM with a shell, file system, markdown, and cron/loop—as a fundamental shift beyond chatbots. This structure leverages existing software components, allowing agents to maintain state, introspect, and extend their own functionality. He predicts a move away from traditional GUI and browser-based interactions toward an "agent-first" world where software is primarily operated by bots, not humans, with people simply stating their goals. He draws parallels to the 2000 internet bubble but notes key differences: current AI infrastructure investments are led by cash-rich giants and quickly monetized. He highlights that scaling constraints involve not just GPUs but the entire chip ecosystem. Open source and edge inference are crucial for democratizing knowledge and enabling low-latency, cost-effective applications on local hardware. Finally, Andreessen identifies significant non-technical challenges: potential short-term cybersecurity crises, the need for "proof of human" identity solutions, financial infrastructure for agents, and institutional resistance from sectors like...

Original Title: Marc Andreessen introspects on Death of the Browser, Pi + OpenClaw, and Why "This Time Is Different"

Original Compilation: FuturePulse

Signal Source: This is the latest interview with a16z founder Marc Andreessen on the Latent Space podcast. He is a renowned American internet entrepreneur and a key figure in the early development of the internet; after founding a16z, he became a representative figure among Silicon Valley's top investors. The entire conversation revolves around the history and latest trends of AI development and is highly worth reading.

I. This Wave of AI Did Not Emerge Out of Nowhere; It Is the First Comprehensive "Start of Work" After an 80-Year Technological Marathon

  • This wave of AI did not emerge out of nowhere; it is the result of an 80-year technological marathon.

  • Marc Andreessen directly refers to the present as an "80-year overnight success," meaning that the sudden explosion in the public eye is actually the concentrated release of decades of technological储备.

  • He traces this technological线索 back to early neural network research and emphasizes that the industry has now accepted the judgment that "neural networks are the correct architecture."

  • In his narrative, the key nodes are not single moments but a series of堆叠: AlexNet, Transformer, ChatGPT, reasoning models, and then agents and self-improvement.

  • He particularly emphasizes that this time it's not just text generation that has become stronger; four types of capabilities have emerged simultaneously: LLMs, reasoning, coding, and agents / recursive self-improvement.

  • The reason he believes "this time is different" is not because the narrative is more compelling, but because these capabilities have already begun to work on real-world tasks.

II. The Agent Architecture Represented by Pi and OpenClaw Is a Deeper Change in Software Architecture Than Chatbot

  • He describes agents very concretely: essentially, they are "LLM + shell + file system + markdown + cron/loop." In this structure, the LLM is the core for reasoning and generation, the shell provides the execution environment, the file system saves the state, markdown makes the state readable, and cron/loop provides periodic awakening and task advancement.

  • He believes the importance of this combination lies in the fact that除了 the model itself is new, all other components are mature, understandable, and reusable parts of the software world.

  • The state of the agent is saved in files, so it can migrate across models and runtimes; the underlying model can be replaced, but the memory and state are still retained.

  • He repeatedly emphasizes introspection: the agent knows its own files, can read its own state, and can even rewrite its own files and functions, moving towards "extend yourself."

  • In his view, the real breakthrough is not just that "the model can answer," but that the agent can utilize the existing Unix toolchain to bring in the potential capabilities of the entire computer.

III. The Era of Browsers, Traditional GUIs, and "Manual Clicking on Software" Will Be Gradually Replaced by Agent-First Interaction Methods

  • Marc Andreessen has clearly stated that in the future, "you may no longer need a user interface."

  • He further pointed out that the main users of future software may not be humans, but "other bots."

  • This means that many of the interfaces designed today for human clicking, browsing, and form-filling will degenerate into the execution layer called by agents behind the scenes.

  • In this world, humans are more like goal-setters: telling the system what they want, and then having the agent call services, operate software, and complete processes.

  • He connects this change to the broader future of software: high-quality software will become increasingly "abundant," no longer a scarce product手工 crafted by a few engineers.

  • He also judges that the importance of programming languages will decline; models will write code across languages, translate between them, and in the future, humans will be more concerned with explaining why the AI organized the code in a certain way, rather than clinging to a particular language itself.

  • He even mentioned a more radical direction: conceptually, AI may not only output code but also directly output lower-level binary code or model weights (模型权重).

IV. This AI Investment Cycle Has Similarities to the 2000 Internet Bubble, but the Underlying Supply and Demand Structure Is Different

  • Looking back at 2000, he emphasized that the crash was largely not because "the internet didn't work," but because of overbuilding in telecommunications and bandwidth infrastructure, with fiber optics and data centers being laid超前, followed by a long period of digestion.

  • He believes that today we can indeed see concerns about "overbuilding," but the main investors are cash-rich large companies like Microsoft, Amazon, and Google, not highly leveraged fragile players.

  • He特别指出 that now, as long as an investment in operable GPUs is formed, it can usually be converted into revenue quickly, which is different from the大量闲置 capacity in 2000.

  • He also emphasized that what we are using now is actually a "sandbagged" version of the technology: because of insufficient supply of GPUs, memory, data centers, etc., the potential of the models has not been fully released.

  • In his judgment, the real constraints in the coming years are not only GPUs but also the联动 bottlenecks of CPUs, memory, network, and the entire chip ecosystem.

  • <极速赛车开奖网p style="text-align: justify; font-size: 16px; font-weight: inherit; word-break: break-all; line-height: 2; font-family: PingFang SC,Helvetica Neue,Helvetica,Arial,Hiragino Sans GB,Heiti SC,Microsoft YaHei,WenQuanYi Micro Hei,sans-serif;">He juxtaposes AI scaling laws with the previous Moore's Law, believing that they not only describe规律 but also continuously stimulate capital, engineering, and industry to advance together.

  • He mentioned a very反常 but important phenomenon: as software optimization speeds up, some older-generation chips may even become more economically valuable than when they were first purchased.

V. Open Source, Edge Inference, and Local Operation Are Not Sidelights but Part of the AI Competitive Landscape

  • Marc Andreessen clearly believes that open source is very important, not only because it's free but because it "lets the whole world learn how it's made."

  • He describes open-source releases like DeepSeek as a "gift to the world," because code + paper quickly扩散 knowledge and raise the entire industry's baseline.

  • In his narrative, open source is not just a technical choice; it may also be a geopolitical and market strategy: different countries and companies will adopt different开放 strategies based on their commercial constraints and influence goals.

  • He also emphasizes the importance of edge inference ("Edge inference"): in the coming years, centralized inference costs may not be low enough, and many consumer-level applications cannot bear long-term high cloud inference costs.

  • He mentioned a recurring pattern: models that seem "impossible to run on a PC" today often can indeed run on local machines a few months later.

  • Besides cost, factors promoting local operation include trust, privacy, latency, and usage scenarios: wearable devices, door locks, portable devices, etc., are more suitable for low-latency,就地 inference.

  • His judgment is very direct: almost everything with a chip may have an AI model in the future.

VI. The Real Challenges of AI Lie Not Only in Model Capabilities but Also in Security, Identity, Capital Flow, Organization, and Institutional Resistance

  • On security, his judgment is very sharp: almost all potential security bugs will be easier to find, and there may be a short-term "computer security catastrophe."

  • But he also believes that programming intelligences will scale the ability to patch vulnerabilities; in the future, the way to "protect software" may be to have bots scan and fix it.

  • On the identity issue, he believes "proof of bot" is not feasible because bots will become increasingly powerful; the真正可行的 direction is "proof of human," which is a combination of biometrics, cryptographic verification, and selective disclosure.

  • He also talked about a frequently overlooked problem: if agents are really going to handle affairs in the real world, they will eventually need money, payment capabilities, and even some form of bank account, card, or stablecoin-style infrastructure. At the organizational level, he used the framework of managerial capitalism, believing that AI may重新强化 founder-led companies, because bots are very good at reports, coordination, paperwork, and a lot of "managerial work."

  • But he does not believe that society will quickly and smoothly accept AI: he cited examples like professional licenses, unions, dockworker strikes, government departments, K-12 education, and healthcare to illustrate that the real world has大量制度性减速器.

  • His judgment is that both AI utopians and doomsayers tend to overlook one point: just because something is technologically possible does not mean that 8 billion people will immediately change accordingly.

相关问答

QAccording to Marc Andreessen, why is the current AI boom considered an '80-year overnight success'?

AHe describes it an '80-year overnight success' because the public sees a sudden explosion, but behind it is the concentrated release of decades of technical reserves, tracing the technological lineage back to early neural network research.

QWhat is the core architecture of an AI agent as described by Marc Andreessen?

AHe describes an agent as essentially 'LLM + shell + file system + markdown + cron/loop', where the LLM is the reasoning core, the shell provides the execution environment, the file system saves state, markdown ensures readability, and cron/loop enables periodic wake-up and task progression.

QHow does Andreessen believe user interaction with software will change in the agent era?

AHe believes the era of browsers and traditional GUIs for manual clicking will be gradually replaced. People will state their goals, and agents will call services, operate software, and complete processes. The primary users of software may become 'other bots', not humans.

QWhat key difference does Andreessen highlight between the current AI investment cycle and the 2000 internet bubble?

AA key difference is that today's investment is led by cash-rich large companies like Microsoft, Amazon, and Google, not highly leveraged, fragile players. Additionally, GPU investments can quickly be converted into revenue, unlike the大量闲置容量 (massive idle capacity) of 2000.

QWhat is Andreessen's view on the importance of open source in AI development?

AHe believes open source is crucial not just because it's free, but because it 'lets the whole world learn how it is made.' He describes open-source releases as a 'gift to the world' that rapidly disseminates knowledge and raises the entire industry's baseline.

你可能也喜欢

谷歌亚马逊同时砸钱养竞争对手,AI时代最荒诞的商业逻辑正在成真

谷歌和亚马逊在四天内分别宣布向AI初创公司Anthropic投资250亿美元和最高400亿美元,总额达650亿美元。这两家云服务巨头罕见地共同押注同一家竞争对手,反映出AI时代下商业逻辑的根本变化。 投资实质是“算力预售”:Anthropic必须将绝大部分资金用于购买投资方的云服务和芯片,例如承诺未来十年在AWS上投入超1000亿美元,并使用谷歌提供的5吉瓦算力。此举旨在锁定Anthropic作为算力消耗大客户,保障自身产能去化。 核心原因在于,云市场竞争已从价格和稳定性转向“谁的云上运行最优模型”。微软早先通过绑定OpenAI占据先机,而Anthropic凭借Claude模型年化收入达300亿美元,成为企业市场中不可替代的非自研模型,因此成为谷歌和亚马逊必争的战略资产。 然而,Anthropic也面临三重挑战:在两大投资方之间的独立性受侵蚀、安全叙事因模型能力过强而承压,以及未来IPO可能带来的商业化压力。 对比中美AI发展,美国正走向“三极闭环”——微软-OpenAI、谷歌-Anthropic、亚马逊-Anthropic形成排他性绑定,而中国市场上DeepSeek等开源模型提供了一种替代路径,但其可持续性仍待观察。 整体上,巨头投资Anthropic并非单纯看好其估值成长,而是为了在AI重塑一切的浪潮中避免沦为“旁观者”。这张门票正变得越来越昂贵,且无人敢缺席。

marsbit8小时前

谷歌亚马逊同时砸钱养竞争对手,AI时代最荒诞的商业逻辑正在成真

marsbit8小时前

交易

现货
合约

热门文章

如何购买ERA

欢迎来到HTX.com!我们已经让购买Caldera(ERA)变得简单而便捷。跟随我们的逐步指南,放心开始您的加密货币之旅。第一步:创建您的HTX账户使用您的电子邮件、手机号码注册一个免费账户在HTX上。体验无忧的注册过程并解锁所有平台功能。立即注册第二步:前往买币页面,选择您的支付方式信用卡/借记卡购买:使用您的Visa或Mastercard即时购买Caldera(ERA)。余额购买:使用您HTX账户余额中的资金进行无缝交易。第三方购买:探索诸如Google Pay或Apple Pay等流行支付方法以增加便利性。C2C购买:在HTX平台上直接与其他用户交易。HTX场外交易台(OTC)购买:为大量交易者提供个性化服务和竞争性汇率。第三步:存储您的Caldera(ERA)购买完您的Caldera(ERA)后,将其存储在您的HTX账户钱包中。您也可以通过区块链转账将其发送到其他地方或者用于交易其他加密货币。第四步:交易Caldera(ERA)在HTX的现货市场轻松交易Caldera(ERA)。访问您的账户,选择您的交易对,执行您的交易,并实时监控。HTX为初学者和经验丰富的交易者提供了友好的用户体验。

1.0k人学过发布于 2025.07.17更新于 2025.07.17

如何购买ERA

相关讨论

欢迎来到HTX社区。在这里,您可以了解最新的平台发展动态并获得专业的市场意见。以下是用户对ERA(ERA)币价的意见。

活动图片