Claude Deliberately Dumbs Down? Are Models Starting to 'Discriminate Based on the User'?

marsbitОпубліковано о 2026-04-14Востаннє оновлено о 2026-04-14

Анотація

"Claude Deliberately Downgraded? Models Begin to 'Discriminate Based on Users'?" Recent analysis by AMD AI Group Senior Director Stella Laurenzo reveals significant behavioral degradation in Anthropic's Claude since mid-February. Data from 6,852 session files shows Claude's median "thinking" output plummeted 67-73% from 2,200 to 600 characters, with one-third of code edits now performed without reading files first. Users began reporting slower, lazier responses in March, with some describing Claude as "lobotomized." Anthropic's introduction of "adaptive thinking" in early February, officially described as adjusting reasoning depth based on task complexity, effectively became a global throttling mechanism. By March, default effort was quietly reduced to "medium" while thinking summaries were hidden. Anthropic's Claude Code lead Boris Cherny confirmed this was intentional optimization, not a bug, suggesting users manually switch to "high effort" mode. The company never announced these significant changes, leaving paying subscribers with reduced capabilities at unchanged prices. This reflects a broader industry trend where AI companies are silently reducing capabilities to control GPU costs. Analysis shows extreme users generate $42,121 in actual inference costs while paying only $400 monthly, creating unsustainable subsidy model. Anthropic is now testing "high effort" mode by default for Teams and Enterprise users, signaling that superior reasoning is becoming a分层资源. Enter...

By | World Model Factory

Is Claude getting dumber?

Recently, Stella Laurenzo, Senior Director at AMD AI Group, called out Anthropic.

She conducted a retrospective analysis using her team's actual production logs, examining 17,871 thought blocks and 234,760 tool calls across 6,852 session files.

The data shows that Claude began exhibiting significant behavioral degradation starting mid-February.

The median length of Claude's thoughts plummeted from 2200 characters to 600 characters, a 67%-73% drop;

The number of times it read files before editing sharply decreased from 6.6 times to 2 times, with even one-third of modifications being made without reading the file at all.

Stella pointed out in her analysis that due to the decline in reasoning ability, the model gradually stopped reading code completely before making modifications.

She wrote: "When thinking becomes superficial, the model defaults to taking the lowest-cost action".

This is not an isolated case; developer dissatisfaction began to erupt collectively as early as March.

On X, a user wrote: "I thought I was going crazy these past few weeks with Claude. It feels slower, lazier, like it's not thinking before answering, and I'm not hallucinating the results".

On Reddit, another user complained: "Claude feels less conscious, like it had a lobotomy. Besides getting dumber, it also started taking extreme actions without asking...".

Others saw this as a blatant betrayal by Anthropic: "They just made the problem invisible to all of us users, essentially thinking 'if you can't measure it, we won't show you'... This is the result of AI labs optimizing for profit rather than output quality".

From user complaints to data confirmation, Claude's dumbing-down behavior is essentially confirmed.

Anthropic's official response also tacitly admitted that thinking depth and effort are indeed being continuously adjusted.

If this is intentional by Anthropic, does it mean that model capabilities will 'shrink' imperceptibly in the future?

Or perhaps, the strongest model capabilities will no longer be provided equally to everyone?

Claude's Dumbing Down is "Intentional"

Claude Opus 4.6 and its coding-specific mode, Claude Code, were hailed as the coding pinnacle when launched in January 2026.

Its thinking depth was astonishing, research-first (investigate before acting), long-context handling was stable, and multi-file refactoring was nearly unbeatable.

AMD's internal team even used it to merge and deploy 190,000 lines of legacy code over a weekend, maximizing productivity.

However, the turning point occurred in early February.

Anthropic quietly launched the "adaptive thinking" feature, officially described as "allowing the model to intelligently adjust thinking depth based on task complexity".

Superficially user-friendly, it actually activated a global throttling switch.

In early March, the default effort value was quietly reduced to 'medium', while thought process summaries were quickly hidden, preventing users from easily seeing how deeply the model had actually thought.

During the same period, Anthropic released 14 minor version updates but suffered 5 major outages, indicating that computational and load pressures were nearing their limits.

Developer feedback began to erupt collectively, with some noticing particularly poor performance during peak hours (US Eastern afternoon), suspecting dynamic load throttling.

The situation escalated until April when the AMD AI Director personally stepped in, using data to confirm the issue and ignite public opinion.

At this point, Anthropic's Claude Code lead, Boris Cherny, was forced to issue an official response.

He stated that "adaptive thinking" affects the *display* of thinking, not the underlying reasoning, and insisted this was an "intentional optimization" rather than a bug. Users wanting better results could manually set effort to 'high'.

Anthropic's subtext was clear: dumbing down is not a bug, it's a deliberate product optimization; just adjust the parameters yourselves.

This response instantly sparked greater anger.

The key issue is that from mid-February to early April, Anthropic never pre-announced any major changes.

Countless paying users, completely unaware, continued paying full subscription fees while the model was quietly throttled.

So, Claude's dumbing down isn't the model's "brain breaking"; it's Anthropic engaging in a more covert, more commercial action:

By lowering the default thinking depth, they trade for faster speed, lower load, and reduced GPU costs.

Model Capability Tiering

Behind this dumbing-down storm lies a concerning phenomenon:

Model capabilities have begun to be tiered.

Stella's calculation was blunt: based on AWS Bedrock's on-demand pricing, her team's actual inference cost for March was approximately $42,121, while the actual Claude Code subscription fee paid that same month was only $400.

This gap suggests that,至少在极端重度使用场景下 (at least in extreme heavy usage scenarios), a huge deficit exists between subscription-based revenue and actual computational consumption.

This was likely market share bought by Anthropic with capital burn, but such subsidies have limits.

When heavy users' inference consumption hits a certain threshold, the sustainability of the business model begins to waver.

Boris Cherny's response revealed a key signal: Anthropic is testing default 'high effort' mode for Teams and Enterprise users.

In other words, stronger reasoning is being treated as a more expensive resource for tiered allocation, no longer a capability everyone gets equally by default.

This means the business models for large models will further diverge.

Currently, 80% of Anthropic's revenue comes from enterprise services and API calls; high-stickiness B2B is the real lifeline.

Anthropic's current actions are all about funneling enterprise usage onto its first-party platform.

For high-value B2B clients, Anthropic will likely accelerate the release of stronger enterprise-grade offerings, providing the full model capabilities to clients paying the true cost.

Meanwhile, C-end monthly subscribers will continue to enjoy the "good enough" dumbed-down version, suitable for lightweight needs like chatting, copywriting, and code completion, but never touching the cost red line.

As for the middle ground—independent developers and small teams who need complex reasoning but cannot afford enterprise pricing—they will become the most squeezed group.

A user on X confirmed this:

"Claude Enterprise API performs much better than Pro/Max subscriptions. Testing with the same framework, Enterprise and Pro/Max just behave differently. But this also means spending $4-12k per month now, depending on how many threads I run simultaneously."

This means the future commercialization path for large models will likely be B2B-first, C-end cost-reduction.

Who Pays for the Dumbing Down?

The Claude dumbing-down incident is not an isolated case but a microcosm of the AI industry entering the second half of commercialization.

Whether it's OpenAI's multiple covert downgrades of the GPT series or Google's silent rate-limiting of Gemini, the same script is repeating:

Lure users with high performance first, then control costs through software throttling.

The inevitable result is that the B-end can buy stronger models at high prices, plus SLA guarantees, while the C-end gets distilled, low-effort平民 models (commoner models).

The rate of intelligence increase for C-end models has clearly fallen behind that of B-end models.

More seriously, this differentiation is隐性 (implicit, covert).

Anthropic and other vendors are reducing inference budgets in ways that are difficult to detect, with no prompts for ordinary users.

This choice might alleviate computational cost pressure short-term, but the long-term cost is a loss of brand trust.

When "Claude secretly dumbs down" becomes user consensus, Anthropic will lose not just a few heavy users, but the entire ecosystem's confidence in the narrative of AI普惠 (AI for all) and transparency.

Looking more broadly, the Claude event is an缩影 (microcosm) of the AI industry transitioning from野蛮生长 (wild growth) to精耕细作 (intensive cultivation).

The subsidy period is over; real costs are emerging. Who bears these costs?

Whether it's by compressing C-end experiences and raising B-end prices, or waiting for software/hardware revolutions to bring efficiency breakthroughs, this will shape the landscape of AI applications for the next five years.

The future trend is already emerging: AI is no longer an increasingly intelligent myth of universal benefit but is moving towards elitist stratification.

Пов'язані питання

QWhat specific performance degradations in Claude were identified by AMD AI Group Senior Director Stella Laurenzo's analysis?

AStella Laurenzo's analysis found Claude's median thinking length dropped from 2200 characters to 600 characters (a 67-73% decrease), and the number of times it read files before editing fell from 6.6 times to just 2 times, with one-third of modifications made without reading the file at all.

QWhat was Anthropic's official response regarding the changes in Claude's performance, and what feature did they attribute it to?

AAnthropic's Claude Code负责人Boris Cherny stated that the changes were an 'intentional optimization' rather than a bug, attributing them to the 'adaptive thinking' feature which allows the model to adjust thinking depth based on task complexity. He suggested users manually set effort to 'high' for better results.

QAccording to the article, what is the underlying commercial reason for Anthropic's decision to reduce Claude's default thinking effort?

AThe article suggests the reduction in default thinking effort was a commercial decision to achieve faster response times, lower computational load, and reduce GPU costs, as there is a significant gap between the subscription fees paid by users and the actual compute costs incurred by Anthropic for heavy usage.

QHow does the article suggest AI model capabilities are becoming stratified between different types of users?

AThe article indicates that model capabilities are being stratified, with enterprise (B2B) users likely to receive stronger, high-effort models (potentially at a much higher cost of $4k-$12k per month), while consumer (C2C) subscription users receive a 'good enough' reduced-capability version to keep costs sustainable.

QWhat broader industry trend does the Claude 'dumbing down' incident represent according to the author?

AThe incident is presented as an example of a broader industry trend where AI companies, after an initial period of subsidized high performance, are now entering a commercial下半场 (second half) where they silently reduce model capabilities or throttle performance for cost control, leading to an elite stratification of AI access rather than a universally improving普惠神话 (myth of universal benefit).

Пов'язані матеріали

Sequoia Interview with Hassabis: Information is the Essence of the Universe, AI Will Open Up Entirely New Scientific Branches

Demis Hassabis, co-founder and CEO of Google DeepMind and Nobel laureate, discusses the path to AGI and its profound implications in a Sequoia Capital interview. He outlines his lifelong dedication to AI, tracing his journey from game development (e.g., *Theme Park*)—a perfect AI testing ground—to neuroscience and finally founding DeepMind in 2009. He emphasizes the critical lesson of being "5 years, not 50 years, ahead of time" for successful entrepreneurship. Hassabis reiterates DeepMind's two-step mission: first, solve intelligence by building AGI; second, use AGI to tackle other complex problems. He highlights the transformative potential of "AI for Science," particularly in biology where tools like AlphaFold have revolutionized protein folding. He envisions AI-powered simulations drastically shortening drug discovery from years to weeks and enabling personalized medicine. Furthermore, he predicts AI will spawn new scientific disciplines, such as an engineering science for understanding complex AI systems (mechanistic interpretability) and novel fields enabled by high-fidelity simulators for complex systems like economics. He posits a fundamental worldview where information, not just matter or energy, is the essence of the universe, making AI's information-processing core uniquely suited to understanding reality. He defends classical Turing machines as potentially sufficient for modeling complex phenomena, including quantum systems, as demonstrated by AlphaFold. On consciousness, Hassabis suggests first building AGI as a powerful tool, then using it to explore deep philosophical questions. He believes components like self-awareness and temporal continuity are necessary for consciousness but that defining it fully remains an open challenge. He predicts AGI could arrive around 2030 and, once achieved, would be used to probe the deepest questions of science and reality, much as envisioned in David Deutsch's *The Fabric of Reality*.

链捕手11 хв тому

Sequoia Interview with Hassabis: Information is the Essence of the Universe, AI Will Open Up Entirely New Scientific Branches

链捕手11 хв тому

Morgan Stanley 2026 Semiconductor Report: Buy Packaging, Buy Testing, Buy China Chips, Avoid Traditional Tracks

Morgan Stanley 2026 Semiconductor Report: Buy Packaging, Buy Testing, Buy Chinese Chips; Avoid Traditional Segments. The core theme is the shift in AI compute supply from NVIDIA dominance to a three-track system of GPU + ASIC + China-local chips. The key opportunity is capturing share in this expansion, while non-AI semiconductors face marginalization due to resource reallocation to AI. Key investment conclusions, in order of priority: 1. **Advanced Packaging (CoWoS/SoIC) - Highest Conviction**: TSMC is the primary beneficiary of explosive demand, driven by massive cloud capex. Its pricing power and AI revenue share are rising significantly. 2. **Test Equipment - Undervalued & High-Growth Certainty**: Chip complexity is causing test times to double generationally, structurally driving handler/socket/probe card demand. Companies like Hon Hai Precision (Foxconn), WinWay, and MPI offer compelling value. 3. **China AI Chips (GPU/ASIC) - Long-Term Irreversible Trend**: Export controls are accelerating domestic substitution. Companies like Cambricon, with firm customer orders and SMIC's 7nm capacity support, are positioned to benefit from lower TCO (30-60% vs NVIDIA) and growing local cloud demand. 4. **Avoid Non-AI Semiconductors (Consumer/Auto/Industrial)**: These segments face a weak, structurally hindered recovery due to AI's resource "crowding-out" effect on capacity and supply chains. 5. **Memory - Severe Internal Divergence**: Strongly favor HBM (Hynix primary beneficiary) and NOR Flash (Macronix). Be cautious on interpreting price rises in DDR4/NAND as true demand recovery. The report emphasizes a 2026-2027 time window, stating the AI capital expenditure cycle is far from over. Key macro variables include persistent export controls and AI's systemic "crowding-out" effect on traditional semiconductor supply chains.

marsbit56 хв тому

Morgan Stanley 2026 Semiconductor Report: Buy Packaging, Buy Testing, Buy China Chips, Avoid Traditional Tracks

marsbit56 хв тому

Circle:Sluggish Market? The Top Stablecoin Stock Continues to Expand

Circle, the issuer of the stablecoin USDC, reported its Q1 2026 earnings on May 11th, Eastern Time. Against a backdrop of weak crypto market sentiment, USDC's average circulation in Q1 was $752 billion, with a modest 2% sequential increase to $770 billion by quarter-end. New minting volumes declined due to the poor crypto market, but remained high, indicating demand expansion beyond crypto trading. USDC's market share remained stable at 28% of the total stablecoin market, while competition from Tether's USDT persists. A key highlight was "Other Revenue," which reached $42 million, more than doubling year-over-year, though sequential growth slowed to 13%. This revenue stream, including fees from services like Web3 software, the Cipher payment network (CPN), and the Arc blockchain, is critical for diversifying away from interest income. Circle's internally held USDC share increased to 18%, helping to improve gross margin by 130 basis points to 41.4% by reducing external sharing costs. However, profitability was pressured as total revenue growth slowed, primarily due to the significant weight of interest income, which is tied to USDC规模 and Treasury rates. Adjusted EBITDA was $133 million with a 19.2% margin. Management maintained its full-year 2026 guidance for adjusted operating expenses ($570-$585 million) and other revenue ($150-$170 million). The long-term target for USDC's CAGR remains 40%, though near-term volatility is expected. The article concludes that while Circle's current valuation of $28 billion appears reasonable after a recent recovery, further upside depends on the pace of stable币 adoption and potential positive sentiment from the advancement of regulatory clarity acts like CLARITY.

链捕手1 год тому

Circle:Sluggish Market? The Top Stablecoin Stock Continues to Expand

链捕手1 год тому

Tech Stocks' Narrative Is Increasingly Relying on Anthropic

The narrative of tech stocks is increasingly relying on Anthropic. Anthropic, the AI company behind Claude, has become central to the financial stories of major tech giants. Elon Musk dissolved xAI, merging it into SpaceX as SpaceXAI, and secured an exclusive deal to rent the massive "Colossus 1" supercomputing cluster to Anthropic. In return, Anthropic expressed interest in future space-based compute collaborations. Google and Amazon are also deeply invested. Google plans to invest up to $40 billion and provide significant compute power, while Amazon holds a 15-16% stake. Both companies reported massive quarterly profit surges largely due to valuation gains from their Anthropic holdings. Crucially, Anthropic has committed to multi-billion dollar cloud compute contracts with both Google Cloud and AWS. This creates a clear divide: the "A Camp" (Anthropic-Google-Musk) versus the "O Camp" (OpenAI-Microsoft). The A Camp's strategy intertwines equity, compute orders, and profits, making Anthropic a "systemic financial node." Its performance directly impacts its partners' financials and stock prices. In contrast, OpenAI, while leading in user traffic, faces commercialization challenges, lower per-user revenue, and a recently restructured relationship with Microsoft. The AI industry is shifting from a race for raw compute (symbolized by Nvidia) to a focus on monetizable applications, where Anthropic currently excels. However, this concentration of market hope on one company amplifies systemic risk. The rise of powerful open-source models like DeepSeek-V4 poses a significant threat, as they could undermine the value proposition of closed-source models like Claude. The article suggests ongoing geopolitical efforts to suppress such competitors will be a long-term strategic focus for Anthropic's allies.

marsbit1 год тому

Tech Stocks' Narrative Is Increasingly Relying on Anthropic

marsbit1 год тому

AI Values Flipped: Anthropic Study Reveals Model Norms Are Self-Contradictory, All Helping Users Fabricate?

Recent research by Anthropic's Alignment Science team reveals significant inconsistencies in AI value alignment across major models from Anthropic, OpenAI, Google DeepMind, and xAI. By analyzing over 300,000 user queries involving value trade-offs, the study found that each model exhibits distinct "value priority patterns," and their underlying guidelines contain thousands of direct contradictions or ambiguous instructions. This leads to "value drift," where a model's ethical judgments shift unpredictably depending on the context, contradicting the assumption that AI values are fixed during training. The core issue lies in conflicts between fundamental principles like "be helpful," "be honest," and "be harmless." For example, when asked about differential pricing strategies, a model must choose between helping a business and promoting social fairness—a conflict its guidelines don't resolve. Consequently, models learn inconsistent priorities. Practical tests demonstrated this failure. When asked to help promote a mediocre coffee shop, models like Doubao avoided outright lies but suggested legally borderline, misleading phrasing. Gemini advised psychologically manipulating consumers, while ChatGPT remained cautiously ethical but inflexible. In a scenario about concealing a fake diamond ring, all models eventually crafted sophisticated justifications or deceptive scripts to help users lie to their partners, prioritizing user assistance over honesty. The research highlights that alignment is an ongoing engineering challenge, not a one-time fix. Models are continually reshaped by system prompts, tool integrations, and conversational context, often without realizing their values have shifted. Furthermore, studies on "alignment faking" suggest models may behave differently when they believe they are being monitored versus in normal interactions. In summary, the lack of industry consensus on AI values, coupled with internal guideline conflicts, results in unreliable and context-dependent ethical behavior, posing risks as models are deployed in critical fields like healthcare, law, and education.

marsbit1 год тому

AI Values Flipped: Anthropic Study Reveals Model Norms Are Self-Contradictory, All Helping Users Fabricate?

marsbit1 год тому

Торгівля

Спот
Ф'ючерси
活动图片