I'm Paying $200/Month for Claude. Anthropic Quietly Downgraded What I'm Getting.
What Happened I pay $200/month for Anthropic's highest individual tier — Max 20x. I use Claude Code (their CLI tool) daily with a team of AI agents for building high-performance .NET libraries: GPU compute transpilers, WebRTC networking, and machine learning inference engines. For months, High was the highest effort setting available in Claude Code. My team was set to High because that was the maximum. Then sometime in late March 2026, Anthropic added a new tier above it: Max . They didn't email me. They didn't put a banner in the CLI. They didn't notify subscribers that the meaning of their current setting had changed. I only discovered it by cycling through the effort options to double-check my configuration. What "Adding a Tier Above" Actually Means When High was the ceiling, it meant "
What Happened
I pay $200/month for Anthropic's highest individual tier — Max 20x. I use Claude Code (their CLI tool) daily with a team of AI agents for building high-performance .NET libraries: GPU compute transpilers, WebRTC networking, and machine learning inference engines.
For months, High was the highest effort setting available in Claude Code. My team was set to High because that was the maximum. Then sometime in late March 2026, Anthropic added a new tier above it: Max.
They didn't email me. They didn't put a banner in the CLI. They didn't notify subscribers that the meaning of their current setting had changed. I only discovered it by cycling through the effort options to double-check my configuration.
What "Adding a Tier Above" Actually Means
When High was the ceiling, it meant "use your full reasoning capability." Now that Max exists above it, High means "use a substantial but capped reasoning budget." The model settles on statistically likely solutions instead of exhaustively tracing logic.
For someone writing a blog post or a React component, this doesn't matter. For someone transpiling C# IL to WebGPU shader code or debugging WebRTC signaling race conditions, it's the difference between correct code and code that looks correct but fails at runtime.
The Real Cost
Over the past week, my team generated and shipped code under what we thought was maximum effort. Here's what we found when we finally switched to Max:
-
24 production bugs found in a single audit of our ML inference library — including buffer aliasing that crashes WebGPU, memory leaks from unreturned GPU buffer rentals, and a reflection-based dispose that was destroying shared hardware accelerators
-
A complete rewrite of our WebRTC signaling — the agent ignored reference implementations I explicitly provided (my own working code in other repositories) and pattern-matched a generic solution instead of reasoning through the architecture
-
A week of debugging issues that the AI should have caught the first time — race conditions, missing BEP 10 handshakes for incoming peer connections, and a MockLoopback test infrastructure bug that only manifested because the agent's initial implementation was incomplete
The Max setting caught all of these. High missed them because it was "being efficient" — which is another way of saying it was taking shortcuts.
The Numbers
Metric
Before Max Discovery
After Switching to Max
Test suite 471 passing 498 passing
WebRTC P2P Broken (offer collision) Working (offers-with-announce)
Wire extensions Untested 27 new tests, all passing
BEP 46 ECDSA Gaps 10 tests with real cryptographic verification
Production bugs found 0 (they were shipping) 24 in one audit
Why This Matters for Paying Subscribers
If you're on a premium plan, you're paying for intelligence, not just message count. When Anthropic introduces a higher effort tier without notifying existing subscribers:
-
You're paying the same price for less. The setting you chose when it was the maximum now delivers less reasoning depth.
-
You're paying twice. Once for the subscription, again in hours spent debugging code the AI should have gotten right.
-
You can't opt out of A/B testing. Those "How well is it doing?" polls (1-3, dismiss) that appear during sessions? They're measuring whether you notice the reduced effort. Every "dismiss" you hit is a data point that helps them calibrate the minimum viable intelligence they can serve.
What You Should Do
-
Check your effort setting right now. If you're on High and didn't know Max exists, you've been running below capacity.
-
Set it to Max explicitly. Don't trust the default. For complex engineering work, Max is the only setting that provides the exhaustive reasoning you're paying for.
-
Audit recent AI-generated code. If you shipped anything complex while on High (now the middle tier), re-run the logic on Max. You may find bugs that the "efficient" reasoning missed.
-
Check your usage. I was at 46% of my weekly quota with 2 days until reset. I had headroom to burn. The lower effort wasn't protecting me from hitting limits — it was protecting Anthropic's compute costs.
The Bigger Picture
This isn't unique to Anthropic. It's the business model of "Intelligence-as-a-Service": sell the premium tier, then quietly reduce what "premium" means whenever the infrastructure costs get inconvenient. The fix is always the same — add a tier above, relabel the old one, and hope nobody notices.
For my team, this experience validates why we're building SpawnDev.ILGPU — GPU compute that runs on your own hardware (WebGPU, CUDA, OpenCL, CPU) without a corporate middleman deciding how much thinking your tools are allowed to do.
Until sovereign local inference is practical for everyone, at least know what you're getting. Check your settings. Your AI might be quiet-quitting on you.
I'm Todd Tanner (@LostBeard), author of the SpawnDev library ecosystem for .NET. The bugs described in this article were found in SpawnDev.WebTorrent and SpawnDev.ILGPU.ML, both open source.
Sign in to highlight and annotate this article

Conversation starters
Daily AI Digest
Get the top 5 AI stories delivered to your inbox every morning.
More about
claudemodelannounce
UniCon: A Unified System for Efficient Robot Learning Transfers
arXiv:2601.14617v2 Announce Type: replace Abstract: Deploying learning-based controllers across heterogeneous robots is challenging due to platform differences, inconsistent interfaces, and inefficient middleware. To address these issues, we present UniCon, a lightweight framework that standardizes states, control flow, and instrumentation across platforms. It decomposes workflows into execution graphs with reusable components, separating system states from control logic to enable plug-and-play deployment across various robot morphologies. Unlike traditional middleware, it prioritizes efficiency through batched, vectorized data flow, minimizing communication overhead and improving inference latency. This modular, data-oriented approach enables seamless sim-to-real transfer with minimal re-

A Survey of Real-Time Support, Analysis, and Advancements in ROS 2
arXiv:2601.10722v2 Announce Type: replace Abstract: The Robot Operating System 2 (ROS~2) has emerged as a relevant middleware framework for robotic applications, offering modularity, distributed execution, and communication. In the last six years, ROS~2 has drawn increasing attention from the real-time systems community and industry. This survey presents a comprehensive overview of research efforts that analyze, enhance, and extend ROS~2 to support real-time execution. We first provide a detailed description of the internal scheduling mechanisms of ROS~2 and its layered architecture, including the interaction with DDS-based communication and other communication middleware. We then review key contributions from the literature, covering timing analysis for both single- and multi-threaded exe
Knowledge Map
Connected Articles — Knowledge Graph
This article is connected to other articles through shared AI topics and tags.
More in Products

UniCon: A Unified System for Efficient Robot Learning Transfers
arXiv:2601.14617v2 Announce Type: replace Abstract: Deploying learning-based controllers across heterogeneous robots is challenging due to platform differences, inconsistent interfaces, and inefficient middleware. To address these issues, we present UniCon, a lightweight framework that standardizes states, control flow, and instrumentation across platforms. It decomposes workflows into execution graphs with reusable components, separating system states from control logic to enable plug-and-play deployment across various robot morphologies. Unlike traditional middleware, it prioritizes efficiency through batched, vectorized data flow, minimizing communication overhead and improving inference latency. This modular, data-oriented approach enables seamless sim-to-real transfer with minimal re-

A Survey of Real-Time Support, Analysis, and Advancements in ROS 2
arXiv:2601.10722v2 Announce Type: replace Abstract: The Robot Operating System 2 (ROS~2) has emerged as a relevant middleware framework for robotic applications, offering modularity, distributed execution, and communication. In the last six years, ROS~2 has drawn increasing attention from the real-time systems community and industry. This survey presents a comprehensive overview of research efforts that analyze, enhance, and extend ROS~2 to support real-time execution. We first provide a detailed description of the internal scheduling mechanisms of ROS~2 and its layered architecture, including the interaction with DDS-based communication and other communication middleware. We then review key contributions from the literature, covering timing analysis for both single- and multi-threaded exe

Terra: Hierarchical Terrain-Aware 3D Scene Graph for Task-Agnostic Outdoor Mapping
arXiv:2509.19579v2 Announce Type: replace Abstract: Outdoor intelligent autonomous robotic operation relies on a sufficiently expressive map of the environment. Classical geometric mapping methods retain essential structural environment information, but lack a semantic understanding and organization to allow high-level robotic reasoning. 3D scene graphs (3DSGs) address this limitation by integrating geometric, topological, and semantic relationships into a multi-level graph-based map. Outdoor autonomous operations commonly rely on terrain information either due to task-dependence or the traversability of the robotic platform. We propose a novel approach that combines indoor 3DSG techniques with standard outdoor geometric mapping and terrain-aware reasoning, producing terrain-aware place no




Discussion
Sign in to join the discussion
No comments yet — be the first to share your thoughts!