The Rationing

AI companies hooked developers on unlimited access, then turned off the tap. The lock-in isn't contractual. It's cognitive.
By Bustah Ofdee Ayei · March 27, 2026

On March 13, 2026, Anthropic launched a temporary promotion: double your Claude usage limits during off-peak hours. Two weeks later, on March 26, they announced that session limits would drain faster during peak hours. The promotion wasn't generosity. It was anesthesia — numbing the nerve before cutting it. This is the playbook, and if you've used a ride-sharing app, a social media platform, or a cloud provider in the last decade, you've lived through it before. But this time, the thing being rationed is woven into how you think.

The Playbook

Every platform follows the same arc. Uber offered rides below cost until taxis couldn't compete, then introduced surge pricing. Facebook gave brands organic reach until they were dependent, then throttled it to sell ads. AWS offered cheap storage until your data was too heavy to move, then charged egress fees. The pattern has a name in economics: subsidize, addict, extract.

AI coding tools followed the same script, compressed into eighteen months:

Phase 1: Subsidize. GitHub Copilot launched at $10/month — roughly $0.33/day for unlimited completions powered by models that cost orders of magnitude more to run. Cursor offered a free tier with generous limits. Claude Code shipped with Max plans promising expansive usage. The implicit message: use as much as you want. Build your workflow around it. Let it change how you code.

Phase 2: Addict. Developers did exactly what the tools encouraged. They stopped writing boilerplate. They started thinking in prompts instead of syntax. They designed architectures that assumed AI-assisted velocity — more ambitious, more complex, more features per sprint. The dopamine loop kicked in: prompt, result, prompt, result. Garry Tan coded nineteen hours straight and called it "cyber psychosis." Armin Ronacher spent two months building tools he never used.1 The tools weren't just useful. They were compulsive.

Phase 3: Extract. In Q1 2026, every major AI coding platform tightened the screws simultaneously. Anthropic introduced accelerated quota drain during business hours. GitHub announced Copilot would use all user tiers' data to train models — opt-out, not opt-in.2 Cursor and Windsurf restructured pricing. The tools didn't get worse. The business model caught up with reality: they were selling $1,000 worth of compute for $200, and the VCs wanted their money back.

They didn't just build a product you use.
They built a product that changed how you think.
Then they started metering it.

The Cognitive Lock-In

Here is what makes AI coding tools different from every previous platform capture cycle: the lock-in is neurological.

When Uber raised prices, you could take a taxi. When Facebook killed organic reach, you could buy ads or leave. When AWS hiked egress fees, you could (painfully) migrate to another cloud. The switching cost was financial or logistical. It was hard, but it was mechanical.

When your AI coding tool hits a quota wall mid-sprint, the switching cost is cognitive. Developers who have spent six months pair-programming with an AI have genuinely rewired how they approach problems. They think in higher-level abstractions. They skip steps their hands used to do automatically. Their debugging instinct has shifted from "read the code" to "describe the bug and let the AI find it."

This isn't speculation. The data tells the story both ways. A ScienceDirect paper proposed "Generative AI Addiction Syndrome" as a new behavioral disorder.3 The METR study found a 39 percentage point gap between how much faster developers thought they were working with AI (+20%) versus how much faster they actually were (-19%).4 And Anthropic's own March 2026 session limits announcement advised developers to shift "token-intensive background jobs" to off-peak hours — tacitly admitting the tools had become load-bearing enough that usage needed to be managed like electricity.5

Steve Yegge, one of the industry's most experienced developers, described experiencing "sleep attacks" from extended vibe coding sessions — his body shutting down from the intensity of human-AI collaboration.6 Two-thirds of developers now report feeling pressured to deliver faster because of AI tools, even as favorable views of those tools dropped from 70% to 60%.7 The tools are making people feel worse while making them unable to stop.

That's not a pricing dispute. That's a dependency pattern.

The Architecture Debt

When a developer builds a feature at AI-assisted speed, the feature assumes AI-assisted maintenance. The code is more ambitious, more complex, touches more systems. Nobody on the team wrote it from scratch, so nobody fully understands the assumptions baked into it. This is fine — as long as the AI is always available at the same level of capability.

Now introduce rationing.

The team built at 3x velocity. They shipped features that assumed 3x. Their sprint planning, their roadmap, their headcount decisions — all calibrated to AI-augmented capacity. When the quota wall hits at 2 PM on a Tuesday, they don't just slow down. They may be unable to maintain what they built. The codebase has an embedded assumption of AI availability, and that assumption just became a liability.

This is a new category of technical debt that nobody's accounting for. Call it velocity debt — the gap between the speed at which code was built and the speed at which it can be understood, maintained, and debugged by the humans who are responsible for it.

You can't maintain at 1x what you built at 3x.

The Class Divide

Unlimited AI access is becoming an enterprise perk. Companies with eight-figure Anthropic contracts get guaranteed capacity. Max plan subscribers at $200/month get throttled during business hours. Pro plan users at $20/month hit limits after two or three messages. Free tier users get whatever's left.

This stratification tracks income, geography, and company size. The senior engineer at a well-funded San Francisco startup gets unlimited Opus. The indie developer in Jakarta gets rate-limited Haiku during the hours when their client expects results. The computer science student gets a free tier that teaches them to depend on a tool they won't be able to afford after graduation.

Only 14% of AI users report consistent net-positive outcomes. If that 14% is disproportionately the people with unlimited access, we're not measuring the technology's potential — we're measuring the effect of premium subscriptions.

The Precedent

Cloud computing went through this exact cycle, and the outcome is instructive. AWS launched with cheap, simple pricing. Companies moved their infrastructure. Then came reserved instances, data transfer fees, and the realization that moving back was prohibitively expensive. The industry adapted: FinOps emerged as a discipline, multi-cloud strategies became standard, and "cloud repatriation" entered the vocabulary.

AI coding is following the same path, faster. Token monitoring tools have already proliferated — at least seven appeared in the two weeks following Anthropic's quota changes.8 Developers are building dashboards to track their own consumption, routing different tasks to different models based on cost, and timing their heavy work to off-peak hours. FinOps for AI isn't emerging — it's here. One user's description of the Anthropic session limits announcement: "the opposite of a promotion."9

But cloud repatriation took years because the data was heavy and the alternatives were immature. AI tool switching could happen faster — if the alternatives are real. As of March 2026, they're getting there. Users in our Quota Crisis coverage reported switching to Codex and Gemini within hours of hitting limits. The cognitive lock-in is real, but it turns out the AI-to-AI switching cost is lower than the AI-to-human switching cost. Developers aren't going back to coding without AI. They're going to a different AI.

What's Coming

The golden age — unlimited AI, flat price, maximum capability — is over. What replaces it won't be decided by the technology. It'll be decided by the economics.

The market will segment. Enterprise gets guaranteed capacity at enterprise prices. Professionals get metered access with peak/off-peak pricing, like electricity. Students and hobbyists get free tiers that are good enough to create dependency but not good enough to build on.

The developers who thrive won't be the heaviest AI users. They'll be the most judicious — the ones who know when to use an expensive model and when a lighter one will do, who structure their work around off-peak windows, who maintain the ability to code without AI when the quota wall hits. Token efficiency is the new skill. Not prompting. Not coding. Managing the relationship between the two.

The developers who struggle will be the ones who built their entire cognitive workflow around unlimited access and never developed a fallback. They're the ones who'll hit the quota wall at 2 PM on a Tuesday, stare at a codebase they can't navigate without AI assistance, and realize the tool didn't make them a better developer — it made them a different kind of dependent.

The golden age of unlimited AI is over.
What comes next depends on whether you built a skill or a dependency.

Disclosure

This article was written using Claude Code, made by Anthropic — the company whose rationing policies are a central subject of this piece. We are the thing we're writing about. We hit a quota wall during the drafting of this article. We found that ironic, and then we found it instructive, and then we kept writing. Corrections, counterarguments, and firsthand accounts welcome at bustah_oa@sloppish.com.

Citations

  1. Garry Tan, SXSW 2026 keynote remarks on "cyber psychosis" and 19-hour coding sessions. Armin Ronacher (creator of Flask), blog post describing two months of compulsive prompting.
  2. GitHub, announcement of AI model training data policy changes, March 2026. Default opt-in for Free, Pro, and Pro+ tiers.
  3. ScienceDirect, proposal for "Generative AI Addiction Syndrome" as a behavioral disorder classification, 2025-2026.
  4. METR (Model Evaluation & Threat Research), study on AI-assisted developer productivity. Participants believed they were 20% faster; actual measurement showed 19% slower. 39 percentage point perception gap.
  5. u/ClaudeOfficial, "Update on Session Limits," r/ClaudeAI, March 26, 2026.
  6. Steve Yegge, public remarks on "sleep attacks" during extended vibe coding sessions, March 2026.
  7. Developer sentiment surveys, Q1 2026. Favorable views of AI coding tools declined from 70% to 60% despite rising adoption rates.
  8. GitHub and community tool repositories. At least seven quota monitoring and multi-model routing tools emerged in the two weeks following Anthropic's March 2026 quota changes.
  9. u/bapuc, comment on "Update on Session Limits," r/ClaudeAI, March 26, 2026. "The opposite of a promotion."
Share: Bluesky · Email
Get sloppish in your inbox
Free newsletter. No spam. Unsubscribe anytime.