Anthropic is prioritizing its own research over paying customers by steering public traffic to inferior hardware. According to Theo on Nerd Snipe, the company moved its 1 million token context version of Claude to general availability not to democratize access, but to offload demand from expensive Nvidia H100s onto cheaper AWS Tranium and Google TPU chips. These alternatives offer higher memory capacity but significantly lower performance, making models feel sluggish and less intelligent.
Users are bearing the cost. An audit by an AMD AI lead found Claude now fails tasks 173 times more often than before Anthropic redacted internal reasoning traces in March. Token usage for engineering tasks exploded from 4.6 million to 20 billion in two months. Theo blames poor session management across Anthropic’s three cloud providers, which breaks thread continuity and forces models to recompute state.
"They’re burning through tokens like crazy because they can't manage session IDs across clouds."
- Theo, Nerd Snipe
Theo also criticizes risky infrastructure changes, including a rare tokenizer update in a dot-release and slashing cache expiration from one hour to five minutes. These moves destabilize the product and suggest a research-first culture indifferent to user experience. Meanwhile, Anthropic bans third-party tools like T3 Code and Open Claw, which automate workflows. Ben notes some tools burn $4 daily on heartbeat checks due to missing caching.
Alex Hearn on The Intelligence argues Anthropic’s gatekeeping extends beyond hardware. The lab restricts access to its powerful Mythos model - capable of finding 27-year-old OpenBSD bugs - to just 11 major partners like Apple and JP Morgan. This avoids public price hikes and denies rivals training data, but entrenches a two-tier AI system.
"Mythos isn’t being released. It’s being rationed to giants who can afford the optics."
- Alex Hearn, The Intelligence
The broader pattern is clear: Anthropic treats public users as secondary. Whether through hardware routing, broken state management, or opaque bans, the company sacrifices reliability and efficiency to protect research resources and control distribution.

