# Сопутствующие статьи по теме Anthropic

Новостной центр HTX предлагает последние статьи и углубленный анализ по "Anthropic", охватывающие рыночные тренды, новости проектов, развитие технологий и политику регулирования в криптоиндустрии.

Can Humans Control AI? Anthropic Conducted an Experiment Using Qwen

Can Humans Control Superintelligent AI? Anthropic’s Experiment with Qwen Models Anthropic conducted an experiment to explore whether humans can supervise AI systems smarter than themselves—a core challenge in AI safety known as scalable oversight. The study simulated a “weak human overseer” using a small model (Qwen1.5-0.5B-Chat) and a “strong AI” using a more powerful model (Qwen3-4B-Base). The goal was to see if the strong model could learn effectively despite imperfect supervision. The key metric was Performance Gap Recovered (PGR). A PGR of 1 means the strong model reached its full potential, while 0 means it was limited by the weak supervisor. Initially, human researchers achieved a PGR of 0.23 after a week of work. Then, nine AI agents (Automated Alignment Researchers, or AARs) based on Claude Opus took over. In five days, they improved PGR to 0.97 through iterative experimentation—proposing ideas, coding, training, and analyzing results. The findings suggest that, in well-defined and automatically scorable tasks, AI can help overcome the supervision gap. However, the methods didn’t generalize perfectly to unseen tasks, and applying them to a production model like Claude Sonnet didn’t yield significant improvements. The study highlights that while AI can automate parts of alignment research, human oversight remains essential to prevent “gaming” of evaluation systems and to handle more complex, real-world problems. Anthropic chose Qwen models for their open-source nature, performance, scalability, and reproducibility—key for rigorous and repeatable experiments. The research demonstrates progress toward automated alignment tools but also underscores that AI supervision remains a nuanced, human-AI collaborative effort.

marsbit04/15 09:28

Can Humans Control AI? Anthropic Conducted an Experiment Using Qwen

marsbit04/15 09:28

Claude Deliberately Dumbs Down? Are Models Starting to 'Discriminate Based on the User'?

"Claude Deliberately Downgraded? Models Begin to 'Discriminate Based on Users'?" Recent analysis by AMD AI Group Senior Director Stella Laurenzo reveals significant behavioral degradation in Anthropic's Claude since mid-February. Data from 6,852 session files shows Claude's median "thinking" output plummeted 67-73% from 2,200 to 600 characters, with one-third of code edits now performed without reading files first. Users began reporting slower, lazier responses in March, with some describing Claude as "lobotomized." Anthropic's introduction of "adaptive thinking" in early February, officially described as adjusting reasoning depth based on task complexity, effectively became a global throttling mechanism. By March, default effort was quietly reduced to "medium" while thinking summaries were hidden. Anthropic's Claude Code lead Boris Cherny confirmed this was intentional optimization, not a bug, suggesting users manually switch to "high effort" mode. The company never announced these significant changes, leaving paying subscribers with reduced capabilities at unchanged prices. This reflects a broader industry trend where AI companies are silently reducing capabilities to control GPU costs. Analysis shows extreme users generate $42,121 in actual inference costs while paying only $400 monthly, creating unsustainable subsidy model. Anthropic is now testing "high effort" mode by default for Teams and Enterprise users, signaling that superior reasoning is becoming a分层资源. Enterprise API users report significantly better performance at $4k-12k monthly costs, while consumer subscribers receive a "good enough" downgraded version. The incident marks the end of AI's subsidy era, with the industry shifting from universal普惠to elite stratification, quietly compromising consumer experience to manage real costs while offering premium capabilities to deep-pocketed enterprise clients.

marsbit04/14 10:32

Claude Deliberately Dumbs Down? Are Models Starting to 'Discriminate Based on the User'?

marsbit04/14 10:32

An Internal Memo Exposes OpenAI's Most Real Anxieties and Ambitions

An internal memo from OpenAI's Chief Revenue Officer, Denise Dresser, reveals the company's strategic priorities and competitive anxieties as the enterprise AI market matures. The document outlines a shift from competing solely on model capability to winning on integration, platform strategy, and becoming "hardest to replace." Key priorities for Q2 include: the model layer, the agent platform, expanding market reach via Amazon, selling the full tech stack, and controlling deployment. The goal is to evolve from a point solution to an enterprise AI "operating system" by deeply embedding into customer workflows, creating switching costs, and securing multi-year, nine-figure deals. The memo contains a direct and unusually sharp critique of rival Anthropic, accusing it of building a narrative on "fear" and "restriction," suffering from compute shortages leading to user experience issues, and overstating its annualized revenue by $8 billion due to accounting methods. This public criticism is seen as a calculated move for investor narratives, internal mobilization, and external signaling. For the Chinese AI market, the memo highlights a gap in competition stages. While domestic players still focus on benchmarks and price wars, the next phase will be won on deployment, platform integration, and ecosystem. It also underscores the critical importance of data sovereignty and trust, suggesting that compliant, auditable, on-premise solutions could be a major differentiator in regulated industries. A notable warning for Chinese companies is OpenAI's claim that its biggest constraint is "capacity," not demand. This contrasts sharply with the domestic market's challenge of finding enterprise customers willing to make large, long-term paid commitments, pointing to a fundamental gap in commercial adoption readiness.

marsbit04/14 10:21

An Internal Memo Exposes OpenAI's Most Real Anxieties and Ambitions

marsbit04/14 10:21

A Four-Page Internal Letter: What Card Is OpenAI Playing?

OpenAI's internal memo, revealed by The Information, outlines a strategic narrative against Anthropic across three key areas: revenue accounting, enterprise competition, and compute capacity. First, OpenAI CRO Denise Dresser challenged Anthropic’s reported $30B annualized revenue, claiming the actual net figure—using OpenAI’s accounting method—is $22B. The discrepancy stems from differing GAAP interpretations: Anthropic books gross revenue (including cloud partner shares), while OpenAI records net revenue after partner deductions. Second, enterprise adoption data from Ramp shows Anthropic rapidly closing the gap with OpenAI, narrowing from an 11% to a 4.6% difference within months. Anthropic already leads in high-value sectors like tech, finance, and professional services. Dresser acknowledged Anthropic’s edge in coding capabilities but warned against being a "single-product company" in a platform war. Third, while current compute capacity is comparable (OpenAI ~1.9 GW vs. Anthropic ~1.4 GW), OpenAI’s long-term plans aim for 30 GW by 2030—four times Anthropic’s projected 7-8 GW by 2027. Anthropic’s growth depends on sustaining enterprise revenue to cover rising cloud costs, estimated to reach $6.4B by 2027. The memo also highlighted OpenAI’s strategic shift: reducing reliance on Microsoft (which “limited customer reach”) and partnering with Amazon, which invests in both OpenAI and Anthropic. This places Amazon’s Bedrock platform as a battleground where both models compete for the same enterprise clients.

marsbit04/14 08:44

A Four-Page Internal Letter: What Card Is OpenAI Playing?

marsbit04/14 08:44

From Wall Street to Silicon Valley, Anthropic Steals All the Spotlight from OpenAI

From Wall Street to Silicon Valley, Anthropic is seizing the spotlight from OpenAI. In just one year, the power dynamics in the AI have shifted significantly. Anthropic is now challenging OpenAI across multiple fronts: market share, secondary market valuation, venture capital sentiment, and public perception. At the recent HumanX AI conference, the consensus was clear—Anthropic is the new darling of Silicon Valley. Its annualized recurring revenue (ARR) has reportedly reached $300 billion, surpassing OpenAI's $250 billion. In the secondary market, Anthropic's valuation has overtaken OpenAI's, with strong investor preference for its shares. Anthropic dominates the enterprise sector, holding 42-54% of the code generation market and 40% of the enterprise agent market, compared to OpenAI's 21% and 27%, respectively. It also leads in new enterprise adoption and cost efficiency. While OpenAI retains a strong consumer user base with ChatGPT, it faces challenges inization and high operational expenses. A leaked internal memo from OpenAI identified Anthropic as its biggest threat, emphasizing its compute infrastructure advantage, but the very need for such a memo highlights its defensive position. Despite OpenAI's strong backing from Amazon and NVIDIA, the market is now valuing efficiency, cost-effectiveness, and precise market fit—areas where Anthropic currently leads. However, experts caution that the AI race is far from over and the landscape remains highly fluid.

marsbit04/13 01:07

From Wall Street to Silicon Valley, Anthropic Steals All the Spotlight from OpenAI

marsbit04/13 01:07

活动图片