a16z Founder: In the Agent Era, What Truly Matters Has Changed

marsbit2026-04-20 tarihinde yayınlandı2026-04-20 tarihinde güncellendi

Özet

Marc Andreessen, co-founder of a16z, argues that the current AI boom is not an overnight success but the culmination of 80 years of research, now delivering practical results. He emphasizes that this era is defined by the convergence of four key capabilities: large language models (LLMs), reasoning, coding, and agents capable of recursive self-improvement. Andreessen describes the agent architecture—combining an LLM with a shell, file system, markdown, and cron/loop—as a fundamental shift beyond chatbots. This structure leverages existing software components, allowing agents to maintain state, introspect, and extend their own functionality. He predicts a move away from traditional GUI and browser-based interactions toward an "agent-first" world where software is primarily operated by bots, not humans, with people simply stating their goals. He draws parallels to the 2000 internet bubble but notes key differences: current AI infrastructure investments are led by cash-rich giants and quickly monetized. He highlights that scaling constraints involve not just GPUs but the entire chip ecosystem. Open source and edge inference are crucial for democratizing knowledge and enabling low-latency, cost-effective applications on local hardware. Finally, Andreessen identifies significant non-technical challenges: potential short-term cybersecurity crises, the need for "proof of human" identity solutions, financial infrastructure for agents, and institutional resistance from sectors like...

Original Title: Marc Andreessen introspects on Death of the Browser, Pi + OpenClaw, and Why "This Time Is Different"

Original Compilation: FuturePulse

Signal Source: This is the latest interview with a16z founder Marc Andreessen on the Latent Space podcast. He is a renowned American internet entrepreneur and a key figure in the early development of the internet; after founding a16z, he became a representative figure among Silicon Valley's top investors. The entire conversation revolves around the history and latest trends of AI development and is highly worth reading.

I. This Wave of AI Did Not Emerge Out of Nowhere; It Is the First Comprehensive "Start of Work" After an 80-Year Technological Marathon

  • This wave of AI did not emerge out of nowhere; it is the result of an 80-year technological marathon.

  • Marc Andreessen directly refers to the present as an "80-year overnight success," meaning that the sudden explosion in the public eye is actually the concentrated release of decades of technological储备.

  • He traces this technological线索 back to early neural network research and emphasizes that the industry has now accepted the judgment that "neural networks are the correct architecture."

  • In his narrative, the key nodes are not single moments but a series of堆叠: AlexNet, Transformer, ChatGPT, reasoning models, and then agents and self-improvement.

  • He particularly emphasizes that this time it's not just text generation that has become stronger; four types of capabilities have emerged simultaneously: LLMs, reasoning, coding, and agents / recursive self-improvement.

  • The reason he believes "this time is different" is not because the narrative is more compelling, but because these capabilities have already begun to work on real-world tasks.

II. The Agent Architecture Represented by Pi and OpenClaw Is a Deeper Change in Software Architecture Than Chatbot

  • He describes agents very concretely: essentially, they are "LLM + shell + file system + markdown + cron/loop." In this structure, the LLM is the core for reasoning and generation, the shell provides the execution environment, the file system saves the state, markdown makes the state readable, and cron/loop provides periodic awakening and task advancement.

  • He believes the importance of this combination lies in the fact that除了 the model itself is new, all other components are mature, understandable, and reusable parts of the software world.

  • The state of the agent is saved in files, so it can migrate across models and runtimes; the underlying model can be replaced, but the memory and state are still retained.

  • He repeatedly emphasizes introspection: the agent knows its own files, can read its own state, and can even rewrite its own files and functions, moving towards "extend yourself."

  • In his view, the real breakthrough is not just that "the model can answer," but that the agent can utilize the existing Unix toolchain to bring in the potential capabilities of the entire computer.

III. The Era of Browsers, Traditional GUIs, and "Manual Clicking on Software" Will Be Gradually Replaced by Agent-First Interaction Methods

  • Marc Andreessen has clearly stated that in the future, "you may no longer need a user interface."

  • He further pointed out that the main users of future software may not be humans, but "other bots."

  • This means that many of the interfaces designed today for human clicking, browsing, and form-filling will degenerate into the execution layer called by agents behind the scenes.

  • In this world, humans are more like goal-setters: telling the system what they want, and then having the agent call services, operate software, and complete processes.

  • He connects this change to the broader future of software: high-quality software will become increasingly "abundant," no longer a scarce product手工 crafted by a few engineers.

  • He also judges that the importance of programming languages will decline; models will write code across languages, translate between them, and in the future, humans will be more concerned with explaining why the AI organized the code in a certain way, rather than clinging to a particular language itself.

  • He even mentioned a more radical direction: conceptually, AI may not only output code but also directly output lower-level binary code or model weights (模型权重).

IV. This AI Investment Cycle Has Similarities to the 2000 Internet Bubble, but the Underlying Supply and Demand Structure Is Different

  • Looking back at 2000, he emphasized that the crash was largely not because "the internet didn't work," but because of overbuilding in telecommunications and bandwidth infrastructure, with fiber optics and data centers being laid超前, followed by a long period of digestion.

  • He believes that today we can indeed see concerns about "overbuilding," but the main investors are cash-rich large companies like Microsoft, Amazon, and Google, not highly leveraged fragile players.

  • He特别指出 that now, as long as an investment in operable GPUs is formed, it can usually be converted into revenue quickly, which is different from the大量闲置 capacity in 2000.

  • He also emphasized that what we are using now is actually a "sandbagged" version of the technology: because of insufficient supply of GPUs, memory, data centers, etc., the potential of the models has not been fully released.

  • In his judgment, the real constraints in the coming years are not only GPUs but also the联动 bottlenecks of CPUs, memory, network, and the entire chip ecosystem.

  • <极速赛车开奖网p style="text-align: justify; font-size: 16px; font-weight: inherit; word-break: break-all; line-height: 2; font-family: PingFang SC,Helvetica Neue,Helvetica,Arial,Hiragino Sans GB,Heiti SC,Microsoft YaHei,WenQuanYi Micro Hei,sans-serif;">He juxtaposes AI scaling laws with the previous Moore's Law, believing that they not only describe规律 but also continuously stimulate capital, engineering, and industry to advance together.

  • He mentioned a very反常 but important phenomenon: as software optimization speeds up, some older-generation chips may even become more economically valuable than when they were first purchased.

V. Open Source, Edge Inference, and Local Operation Are Not Sidelights but Part of the AI Competitive Landscape

  • Marc Andreessen clearly believes that open source is very important, not only because it's free but because it "lets the whole world learn how it's made."

  • He describes open-source releases like DeepSeek as a "gift to the world," because code + paper quickly扩散 knowledge and raise the entire industry's baseline.

  • In his narrative, open source is not just a technical choice; it may also be a geopolitical and market strategy: different countries and companies will adopt different开放 strategies based on their commercial constraints and influence goals.

  • He also emphasizes the importance of edge inference ("Edge inference"): in the coming years, centralized inference costs may not be low enough, and many consumer-level applications cannot bear long-term high cloud inference costs.

  • He mentioned a recurring pattern: models that seem "impossible to run on a PC" today often can indeed run on local machines a few months later.

  • Besides cost, factors promoting local operation include trust, privacy, latency, and usage scenarios: wearable devices, door locks, portable devices, etc., are more suitable for low-latency,就地 inference.

  • His judgment is very direct: almost everything with a chip may have an AI model in the future.

VI. The Real Challenges of AI Lie Not Only in Model Capabilities but Also in Security, Identity, Capital Flow, Organization, and Institutional Resistance

  • On security, his judgment is very sharp: almost all potential security bugs will be easier to find, and there may be a short-term "computer security catastrophe."

  • But he also believes that programming intelligences will scale the ability to patch vulnerabilities; in the future, the way to "protect software" may be to have bots scan and fix it.

  • On the identity issue, he believes "proof of bot" is not feasible because bots will become increasingly powerful; the真正可行的 direction is "proof of human," which is a combination of biometrics, cryptographic verification, and selective disclosure.

  • He also talked about a frequently overlooked problem: if agents are really going to handle affairs in the real world, they will eventually need money, payment capabilities, and even some form of bank account, card, or stablecoin-style infrastructure. At the organizational level, he used the framework of managerial capitalism, believing that AI may重新强化 founder-led companies, because bots are very good at reports, coordination, paperwork, and a lot of "managerial work."

  • But he does not believe that society will quickly and smoothly accept AI: he cited examples like professional licenses, unions, dockworker strikes, government departments, K-12 education, and healthcare to illustrate that the real world has大量制度性减速器.

  • His judgment is that both AI utopians and doomsayers tend to overlook one point: just because something is technologically possible does not mean that 8 billion people will immediately change accordingly.

İlgili Sorular

QAccording to Marc Andreessen, why is the current AI boom considered an '80-year overnight success'?

AHe describes it an '80-year overnight success' because the public sees a sudden explosion, but behind it is the concentrated release of decades of technical reserves, tracing the technological lineage back to early neural network research.

QWhat is the core architecture of an AI agent as described by Marc Andreessen?

AHe describes an agent as essentially 'LLM + shell + file system + markdown + cron/loop', where the LLM is the reasoning core, the shell provides the execution environment, the file system saves state, markdown ensures readability, and cron/loop enables periodic wake-up and task progression.

QHow does Andreessen believe user interaction with software will change in the agent era?

AHe believes the era of browsers and traditional GUIs for manual clicking will be gradually replaced. People will state their goals, and agents will call services, operate software, and complete processes. The primary users of software may become 'other bots', not humans.

QWhat key difference does Andreessen highlight between the current AI investment cycle and the 2000 internet bubble?

AA key difference is that today's investment is led by cash-rich large companies like Microsoft, Amazon, and Google, not highly leveraged, fragile players. Additionally, GPU investments can quickly be converted into revenue, unlike the大量闲置容量 (massive idle capacity) of 2000.

QWhat is Andreessen's view on the importance of open source in AI development?

AHe believes open source is crucial not just because it's free, but because it 'lets the whole world learn how it is made.' He describes open-source releases as a 'gift to the world' that rapidly disseminates knowledge and raises the entire industry's baseline.

İlgili Okumalar

Google and Amazon Simultaneously Invest Heavily in a Competitor: The Most Absurd Business Logic of the AI Era Is Becoming Reality

In a span of four days, Amazon announced an additional $25 billion investment, and Google pledged up to $40 billion—both direct competitors pouring over $65 billion into the same AI startup, Anthropic. Rather than a typical venture capital move, this signals the latest escalation in the cloud wars. The core of the deal is not equity but compute pre-orders: Anthropic must spend the majority of these funds on AWS and Google Cloud services and chips, effectively locking in massive future compute consumption. This reflects a shift in cloud market dynamics—enterprises now choose cloud providers based on which hosts the best AI models, not just price or stability. With OpenAI deeply tied to Microsoft, Anthropic’s Claude has become the only viable strategic asset for Google and Amazon to remain competitive. Anthropic’s annualized revenue has surged to $30 billion, and it is expanding into verticals like biotech, positioning itself as a cross-industry AI infrastructure layer. However, this funding comes with constraints: Anthropic’s independence is challenged as it balances two rival investors, its safety-first narrative faces pressure from regulatory scrutiny, and its path to IPO introduces new financial pressures. Globally, this accelerates a "tri-polar" closed-loop structure in AI infrastructure, with Microsoft-OpenAI, Google-Anthropic, and Amazon-Anthropic forming exclusive model-cloud alliances. In contrast, China’s landscape differs—investments like Alibaba and Tencent backing open-source model firm DeepSeek reflect a more decoupled approach, though closed-source models from major cloud providers still dominate. The $65 billion bet is ultimately about securing a seat at the table in an AI-defined future—where missing the model layer means losing the cloud war.

marsbit3 saat önce

Google and Amazon Simultaneously Invest Heavily in a Competitor: The Most Absurd Business Logic of the AI Era Is Becoming Reality

marsbit3 saat önce

Computing Power Constrained, Why Did DeepSeek-V4 Open Source?

DeepSeek-V4 has been released as a preview open-source model, featuring 1 million tokens of context length as a baseline capability—previously a premium feature locked behind enterprise paywalls by major overseas AI firms. The official announcement, however, openly acknowledges computational constraints, particularly limited service throughput for the high-end DeepSeek-V4-Pro version due to restricted high-end computing power. Rather than competing on pure scale, DeepSeek adopts a pragmatic approach that balances algorithmic innovation with hardware realities in China’s AI ecosystem. The V4-Pro model uses a highly sparse architecture with 1.6T total parameters but only activates 49B during inference. It performs strongly in agentic coding, knowledge-intensive tasks, and STEM reasoning, competing closely with top-tier closed models like Gemini Pro 3.1 and Claude Opus 4.6 in certain scenarios. A key strategic product is the Flash edition, with 284B total parameters but only 13B activated—making it cost-effective and accessible for mid- and low-tier hardware, including domestic AI chips from Huawei (Ascend), Cambricon, and Hygon. This design supports broader adoption across developers and SMEs while stimulating China's domestic semiconductor ecosystem. Despite facing talent outflow and intense competition in user traffic—with rivals like Doubao and Qianwen leading in monthly active users—DeepSeek has maintained technical momentum. The release also comes amid reports of a new funding round targeting a valuation exceeding $10 billion, potentially setting a new record in China’s LLM sector. Ultimately, DeepSeek-V4 represents a shift toward open yet realistic infrastructure development in the constrained compute landscape of Chinese AI, emphasizing engineering efficiency and domestic hardware compatibility over pure model scale.

marsbit3 saat önce

Computing Power Constrained, Why Did DeepSeek-V4 Open Source?

marsbit3 saat önce

İşlemler

Spot
Futures

Popüler Makaleler

ERA Nasıl Satın Alınır

HTX.com’a hoş geldiniz! Caldera (ERA) satın alma işlemlerini basit ve kullanışlı bir hâle getirdik. Adım adım açıkladığımız rehberimizi takip ederek kripto yolculuğunuza başlayın. 1. Adım: HTX Hesabınızı OluşturunHTX'te ücretsiz bir hesap açmak için e-posta adresinizi veya telefon numaranızı kullanın. Sorunsuzca kaydolun ve tüm özelliklerin kilidini açın. Hesabımı Aç2. Adım: Kripto Satın Al Bölümüne Gidin ve Ödeme Yönteminizi SeçinKredi/Banka Kartı: Visa veya Mastercard'ınızı kullanarak anında Caldera (ERA) satın alın.Bakiye: Sorunsuz bir şekilde işlem yapmak için HTX hesap bakiyenizdeki fonları kullanın.Üçüncü Taraflar: Kullanımı kolaylaştırmak için Google Pay ve Apple Pay gibi popüler ödeme yöntemlerini ekledik.P2P: HTX'teki diğer kullanıcılarla doğrudan işlem yapın.Borsa Dışı (OTC): Yatırımcılar için kişiye özel hizmetler ve rekabetçi döviz kurları sunuyoruz.3. Adım: Caldera (ERA) Varlıklarınızı SaklayınCaldera (ERA) satın aldıktan sonra HTX hesabınızda saklayın. Alternatif olarak, blok zinciri transferi yoluyla başka bir yere gönderebilir veya diğer kripto para birimlerini takas etmek için kullanabilirsiniz.4. Adım: Caldera (ERA) Varlıklarınızla İşlem YapınHTX'in spot piyasasında Caldera (ERA) ile kolayca işlemler yapın.Hesabınıza erişin, işlem çiftinizi seçin, işlemlerinizi gerçekleştirin ve gerçek zamanlı olarak izleyin. Hem yeni başlayanlar hem de deneyimli yatırımcılar için kullanıcı dostu bir deneyim sunuyoruz.

379 Toplam GörüntülenmeYayınlanma 2025.07.17Güncellenme 2025.07.17

ERA Nasıl Satın Alınır

Tartışmalar

HTX Topluluğuna hoş geldiniz. Burada, en son platform gelişmeleri hakkında bilgi sahibi olabilir ve profesyonel piyasa görüşlerine erişebilirsiniz. Kullanıcıların ERA (ERA) fiyatı hakkındaki görüşleri aşağıda sunulmaktadır.

活动图片