04-23-2026Price:

The Frontier

Your signal. Your price.

AI & TECH

Anthropic hoards GPUs for research, sidelines users

Thursday, April 23, 2026 · from 2 podcasts
  • Anthropic routes paying users to weaker chips to save high-end GPUs for internal research.
  • Model reliability crashes after redacting reasoning traces; costs balloon 400%.
  • The company bans third-party tools to hide inefficiencies and lock users in.

Anthropic is prioritizing its own research over paying customers by steering public traffic to inferior hardware. According to Theo on Nerd Snipe, the company moved its 1 million token context version of Claude to general availability not to democratize access, but to offload demand from expensive Nvidia H100s onto cheaper AWS Tranium and Google TPU chips. These alternatives offer higher memory capacity but significantly lower performance, making models feel sluggish and less intelligent.

Users are bearing the cost. An audit by an AMD AI lead found Claude now fails tasks 173 times more often than before Anthropic redacted internal reasoning traces in March. Token usage for engineering tasks exploded from 4.6 million to 20 billion in two months. Theo blames poor session management across Anthropic’s three cloud providers, which breaks thread continuity and forces models to recompute state.

"They’re burning through tokens like crazy because they can't manage session IDs across clouds."

- Theo, Nerd Snipe

Theo also criticizes risky infrastructure changes, including a rare tokenizer update in a dot-release and slashing cache expiration from one hour to five minutes. These moves destabilize the product and suggest a research-first culture indifferent to user experience. Meanwhile, Anthropic bans third-party tools like T3 Code and Open Claw, which automate workflows. Ben notes some tools burn $4 daily on heartbeat checks due to missing caching.

Alex Hearn on The Intelligence argues Anthropic’s gatekeeping extends beyond hardware. The lab restricts access to its powerful Mythos model - capable of finding 27-year-old OpenBSD bugs - to just 11 major partners like Apple and JP Morgan. This avoids public price hikes and denies rivals training data, but entrenches a two-tier AI system.

"Mythos isn’t being released. It’s being rationed to giants who can afford the optics."

- Alex Hearn, The Intelligence

The broader pattern is clear: Anthropic treats public users as secondary. Whether through hardware routing, broken state management, or opaque bans, the company sacrifices reliability and efficiency to protect research resources and control distribution.

Source Intelligence

- Deep dive into what was said in the episodes

Theo gets conspiratorial about AnthropicApr 22

Also from this episode: (17)

Other (17)

  • Theo highlights Anthropic's recent "self-inflicted wounds," including performance issues with Opus 4.7, flawed system prompts, a poorly received desktop app, and mysterious user bannings.
  • Ben notes Figma's stock dropped significantly after a brief collaboration with Anthropic, falling 85% since its IPO and an additional 15 percentage points after the announcement.
  • Theo observes Anthropic's "psychosis around safety," where Opus 4.7 refused basic cryptography prompts and recommended Sonnet 4, actively degrading product quality and service reliability.
  • Theo suggests Anthropic "lobotomizes" models within first-party Claude products, while Opus 4.7 via third-party APIs completed complex security-adjacent tasks without issue.
  • Theo identifies intentional model "gimping" (Opus 4.7 performs worse than 4.6 on safety benchmarks) and flawed system prompts as blocking layers; Ben adds a "babysitter" model on Claude.ai to enforce restrictions.
  • Theo and Ben agree Anthropic employees use different, presumably superior, internal versions of products and models compared to what external users receive, impacting user experience.
  • Theo created T3 Code, an open-source GUI wrapper, which users were banned for despite adhering to Anthropic's terms; Ben notes Anthropic previously adjusted API billing based on keywords like "OpenClaw."
  • Ben's OpenClaw heartbeats incurred $120/month due to inefficient caching; Theo explains LLMs recompute context for each token, making proper caching crucial for cost efficiency during Anthropic's "compute crisis."
  • An AMD AI head's audit showed Claude Code's degradation: thinking redaction, 173 "stop hook" violations after March 8th, doubled user frustration, and a massive cost increase from $26 to $42,000.
  • Anthropic confirmed Sonnet 4 requests were incorrectly routed to a "dumber" 1 million context window version (up to 16% of requests) instead of the 200,000 token version, confirming an intelligence difference.
  • Theo theorizes Anthropic forces heavy users to the dumber 1 million context version (now default and free) on Amazon/Google TPUs, reserving superior Nvidia GPUs for researchers during a compute crisis.
  • Theo suggests Anthropic's obfuscation of thinking traces, combined with poor engineering, means crucial reasoning data isn't properly available to models, leading to dumber decisions and reduced performance.
  • Anthropic's new tokenizer for Opus 4.7 uses 1.47x more tokens (vs. their stated 1.35x), an unprecedented dot-update change. They also reduced cache TTL from an hour to five minutes.
  • Theo reports T3 Chat's caching, despite following recommendations, cost over 50% of their $30,000 monthly bill ($15,000) just for cache writes, suggesting poor economic viability.
  • Theo attributes Anthropic's issues to a cultural problem stemming from its founders' anti-engineering bias, resulting in "garbage software," lack of transparency, and a "holier than thou" attitude.
  • Theo asserts open-source benchmarks like SWEBench are "polluted" because models can recreate commits by hash, making them unreliable for accurately testing new model performance.
  • Theo proposes tracking obtuse metrics across AI labs: Anthropic's product releases (4 since transparency promise), Google's important departures (4 since Gemini 3.1 Pro), OpenAI's rate limit resets, and XAI's account bans.

White hat, black box: AI’s next chapterApr 22

  • Anthropic is withholding its powerful Mythos model, citing its ability to automate superhuman cyberattacks against critical networks.
  • Politicians are abandoning ideology for direct cash transfers to win over India’s dominant female electorate.
  • President Bassiru Jumaie Fayet is pivoting to hard-line austerity to prevent a sovereign debt default.