Gargantua
Capacity Protection and the Time Dilation of AI-Accelerated Work

"One hour here is seven years back on Earth."
— Cooper, Interstellar (2014), standing on Miller's Planet, 1.3 AU from a black hole with 100 million solar masses
Cooper lands on Miller's Planet. He's there for three hours and seventeen minutes. When he gets back to the Endurance, twenty-three years have passed. Romilly has aged. His children are grown. The messages have piled up. He sits in the ship and watches his kids become strangers on a screen, a decade per video, sobbing through the time he can never get back.
It's the most devastating scene in the film because it makes time dilation feel like something. Not an equation. A loss. The physics didn't betray Cooper. He understood the math. He knew what the black hole would cost. But understanding the ratio intellectually and experiencing the gap emotionally are different things.
I want to talk about why your project manager and an AI company are both doing the same thing Cooper did: staring at the math, realizing capacity is finite, and making the same hard choices about what to protect.
Because on March 26, 2026 — five days ago — Anthropic quietly adjusted the session limits for Claude Pro, Max, and Free users during peak hours. The token cost per session went up. The five-hour window started burning faster. Seven percent of users began hitting limits they'd never hit before. By today, March 31, Anthropic admitted that Claude Code quotas are "draining way faster than expected" and said it's their "top priority."
Their official recommendation? "Shift token-intensive background jobs to off-peak hours."
Read that again. An AI company — the company building the most capable AI models on the planet — just told its users to rearrange their work schedule around capacity constraints. To time-shift their demand. To self-triage. To accept that the system is finite and act accordingly.
Your project manager has been saying the same thing for twenty years. It's called "what's the priority?"
The Anthropic Capacity Timeline
How an AI company rediscovered the same capacity protection mechanisms humans have used for decades
Jan 2026
Opus 4.5 usage limits "significantly reduced" since launch
GitHub Issue #17084
Mar 26, 2026
Anthropic adjusts 5-hour session limits during peak hours (5am-11am PT)
The Register
Mar 26, 2026
Token costs increased during peak — your 5-hour window burns faster
TechRadar
Mar 27, 2026
~7% of users now hitting limits they wouldn't have before
gHacks
Mar 31, 2026
Anthropic admits Claude Code quotas "draining way faster than expected"
The Register
Mar 31, 2026
Official recommendation: "Shift token-intensive work to off-peak hours"
Anthropic
Section I: The Physics of Finite Capacity
In general relativity, time dilation occurs when a massive object curves spacetime. The closer you are to the mass, the slower time passes for you relative to a distant observer. On Miller's Planet, Gargantua's gravitational pull is so extreme that one hour on the surface equals seven years on the Endurance. Cooper isn't moving slowly — he's moving through time at a different rate than everyone else.
Now replace "massive object" with "AI-augmented team," and "spacetime" with "the velocity of work around you."
In our previous posts — Limitless: The Human Token Economy and Be Kind, Rewind — we established that a human knowledge worker produces roughly 4,000-8,000 tokens of meaningful work per day. Emails, Slack messages, documents, code, meeting contributions. Everything that leaves your brain through your fingers or your mouth.
With AI augmentation, your team's collective output isn't 5x that. It's 10-50x. The AI didn't just speed up individual contributors — it compressed timelines so dramatically that the gap between "what you produced" and "what happened around you" has become a time dilation problem.
The Gravity Well
Larger mass = stronger gravitational pull = more time dilation for nearby observers
The closer you orbit the AI-augmented team, the more time-dilated you become relative to everyone else.
You are Cooper. Your AI-augmented team is Gargantua. The closer you work to the team's AI-accelerated output, the more your personal time frame diverges from theirs. You step away for a meeting. You come back to a universe that aged weeks. Not metaphorically — in token output, the team produced what used to take weeks.
Section II: The Question Gap
Here's where the time dilation becomes personal. Think about how knowledge work actually flows. You don't work in isolation. You work in a cycle of doing and asking. You do work, you hit an ambiguity, you ask a question, you wait for the answer, you resume.
Every question you ask creates a gap. And in that gap, the world keeps moving.
Pre-AI, the gap was manageable. You post a question in Teams at 9am. By the time someone answers at 11am, roughly 2,000-4,000 tokens of work happened around you. Two hours of normal human output. You read the Slack thread. You're caught up in five minutes. No problem.
Post-AI, you post the same question at 9am. By 11am, your team — augmented by AI coding assistants, AI analysis tools, AI document generators — has produced 40,000-85,000 tokens. That's not a Slack thread you can skim. That's three design documents, two prototypes, a cost analysis, and a Slack thread with 47 messages debating the tradeoffs of an approach that didn't exist when you asked the question.
The question gap is the Miller's Planet of knowledge work. Every time you pause — to ask a question, to attend a meeting, to eat lunch, to think — you land on the surface. And when you come back, time has dilated.
The Question Gap Calculator
How much work happens between your questions?
17,000
10x
5
3,400
Pre-AI: tokens between questions
34,000
Post-AI: tokens between questions
10x
Dilation factor
102 min
Reading time to catch up
Pre-AI, 3,400 tokens of work happened between your questions. Now it's 34,000. Every time you pause to ask, you fall 10x further behind than you used to.
This is why "scope creep" feels more dangerous now. It always was a capacity protection mechanism — a way of saying "we can only orbit so close to this mass before we lose control." But pre-AI, scope creep added days. Post-AI, scope creep adds what feels like months of output in days. The gravitational pull of AI-accelerated work makes every additional ask exponentially more expensive in terms of the gap it creates.
Five Scenarios of Time Dilation at Work
Let's make this concrete. Here are five common moments when you step away — and what the dilation looks like:
The Quick Question
You stepped away for 2 hours
Your tokens
2,000
Team + AI tokens
85,000
What you said before stepping away:
"Hey, should we use REST or GraphQL for this endpoint?"
The Lunch Break
You stepped away for 1 hour
Your tokens
0
Team + AI tokens
48,000
What you said before stepping away:
"Can someone review my PR when they get a chance?"
The Meeting
You stepped away for 45 minutes
Your tokens
3,200
Team + AI tokens
156,000
What you said before stepping away:
"Let's circle back on the API design after this meeting."
The Deep Focus Block
You stepped away for 3 hours
Your tokens
6,500
Team + AI tokens
210,000
What you said before stepping away:
"I'm heads down until 2pm, will catch up on Slack then."
The PTO Day
You stepped away for 8 hours
Your tokens
0
Team + AI tokens
520,000
What you said before stepping away:
"Taking tomorrow off, nothing urgent pending."
Section III: The Dilation Clock
Cooper watched his children age on a screen. You watch your Slack unread count climb. Different medium, same physics. Here's a clock that shows the gap opening in real time — your human token output versus what an AI-augmented team produces in the same wall-clock seconds:
Time Dilation Clock
Watch the gap open in real time
Your Output
0
tokens (human pace)
Your Team + AI
0
tokens (AI-augmented)
Dilation ratio
--
For every token you produce, the world produces ... around you
That's not a simulation of the future. That's what's happening right now, in every AI-augmented workplace, every time you step away from the keyboard. The clock doesn't stop when you do. It accelerates.
Section IV: Capacity Protection Is a Universal Law
Here's the insight that connects Cooper's problem, Anthropic's problem, and your project manager's problem: they're all solving for the same variable. Finite capacity in the face of infinite (or rapidly scaling) demand.
The solutions are identical, too. Not similar — identical. The same mechanisms appear independently in human organizations, AI systems, and physics because they're not cultural artifacts or engineering patterns. They're mathematical necessities that emerge from any system with bounded resources and unbounded demand.
Human
"What's the priority?"
AI (Anthropic)
Peak-hour token cost multiplier
Physics
Conservation of energy
Human
"That's scope creep."
AI (Anthropic)
5-hour session limits
Physics
Event horizon
Human
"Let's take this offline."
AI (Anthropic)
"Try again in 3 hours"
Physics
Time dilation
Human
"I'm at capacity."
AI (Anthropic)
HTTP 429: Too Many Requests
Physics
Schwarzschild radius
Human
Meeting-free Fridays
AI (Anthropic)
Off-peak hour discounts
Physics
Orbital mechanics
Human
Quarterly planning
AI (Anthropic)
Weekly usage caps
Physics
Resonance
Click any row to expand
Look at the symmetry. Every human capacity protection mechanism has an exact analog in how Anthropic manages its infrastructure, and both map to fundamental physics. This isn't coincidence. It's convergent evolution. Any sufficiently complex system with finite capacity independently discovers the same solutions:
Triage: not everything can be served. Decide what matters.
Boundaries: define the scope. What's inside, what's outside. Enforce it.
Temporal shifting: move demand to when capacity exists.
Honest signals: communicate when the system is full. Don't pretend it isn't.
Recovery windows: schedule downtime before the system forces it.
Batching: consolidate demand into planned intervals instead of continuous intake.
Section V: Why Anthropic's Throttle Matters More Than You Think
Most commentary on Anthropic's March 2026 capacity adjustments has focused on user frustration. "I'm paying $200/month and I can't use the tool for a full day." Fair complaint. But the more interesting story is what the throttle reveals about the state of AI infrastructure.
Anthropic has more demand than GPU capacity. That's the sentence buried in every announcement. They're not throttling because they want to. They're throttling because the alternative is degraded service for everyone. The same reason your project manager says "that's not in scope" — not because they don't want to build it, but because saying yes to everything means delivering nothing well.
The specific mechanisms Anthropic chose are textbook capacity protection:
Peak-hour pricing
Token costs are higher during 5am-11am PT. This is surge pricing for compute — the same principle as Uber, electricity markets, and highway tolls. Increase the price until demand matches supply.
Session windows
Five-hour blocks that burn faster during peak. Your allocation is fixed; the cost per unit of work scales with demand. It's a sliding rate limit — identical to how cellular networks throttle during stadium events.
Usage transparency
Anthropic is telling users to check their usage and plan accordingly. This is the AI equivalent of your manager saying "here's the sprint velocity — what fits?" Capacity becomes visible, not assumed.
Temporal shifting
"Move token-intensive work to off-peak hours." This is meeting-free Fridays for AI. Reserve the high-demand windows for interactive work; batch the heavy processing for nights and weekends.
What's remarkable isn't that Anthropic is doing this. It's that these are the exact same mechanisms that human organizations have used for decades to manage knowledge worker capacity. Anthropic didn't invent anything. They rediscovered what every operations manager already knew: when demand exceeds capacity, you either degrade quality for everyone or ration access for some. There is no third option. The physics don't allow it.
Section VI: The Exacerbation Loop
Here's where the time dilation compounds. Every capacity protection mechanism you use — asking a question, raising a gap, flagging scope creep — takes you off the surface. You step away from production to coordinate. And while you coordinate, the AI-augmented world keeps producing.
Pre-AI, this was manageable. You flag a gap in Teams. Someone responds in an hour. During that hour, maybe 1,000-2,000 tokens of work happened. You read the thread and move on.
Post-AI, you flag the same gap. During the hour you wait for the answer, 48,000 tokens of work happened. The prototype you were questioning? It's been built, tested, and deployed. Your gap is no longer a question about an approach — it's a question about a shipped feature.
This creates a vicious loop:
You notice a gap, risk, or ambiguity in the work
You raise it in Teams/Slack (capacity protection — good practice)
While you wait for a response, the AI-augmented team keeps producing
By the time someone responds, 10-50x more work has happened
Your original question may now be irrelevant — the team moved past it
You either accept work you haven't validated, or become "the bottleneck"
You raise fewer gaps to avoid being the bottleneck
Quality decreases because capacity protection was abandoned
This is the gravitational spaghettification of knowledge work. Get too close to the AI-augmented velocity and the tidal forces tear your capacity protection apart. Your questions can't keep up with the answers. Your reviews can't keep up with the output. Your judgment can't keep up with the volume. The very thing that makes you valuable — human oversight, institutional knowledge, domain judgment — gets compressed past its Schwarzschild radius.
The Bottleneck Paradox
The human knowledge worker is simultaneously the most valuable and the most capacity-constrained component in an AI-augmented system. You are the judgment layer. You are also the slowest layer. Every system, eventually, is governed by its slowest component. AI doesn't eliminate the bottleneck — it makes the bottleneck more consequential. The gap between "AI speed" and "human judgment speed" is the time dilation. And it's widening.
Section VII: What Cooper Did Next
In Interstellar, Cooper doesn't solve time dilation. He can't. It's physics. What he does is account for it. He makes decisions knowing the cost. He stops pretending the math doesn't apply. He factors the dilation into every choice: which planet to visit, how long to stay, what he's willing to lose.
That's the only playbook for AI-accelerated work, too. You can't eliminate the time dilation. The gap between human pace and AI pace is going to widen, not shrink. But you can stop pretending it doesn't exist and start building for it:
Make the Dilation Visible
Most teams don't see the gap because they don't measure it. Start tracking: how many tokens of output happen between your coordination touchpoints? If the number is 10x what it was last year, your capacity protection mechanisms need to scale 10x too. You can't manage what you don't measure.
Increase Question Frequency, Decrease Question Scope
The old model: big questions, long gaps. "Should we use REST or GraphQL?" takes 2 hours to answer and 85K tokens of work happen in between. The new model: small questions, short gaps. "I'm going REST for the user endpoint, any blockers?" takes 10 minutes to answer and 7K tokens happen in between. Same capacity protection. Smaller dilation.
Build Async Review Into the Flow
Don't wait for review gates. Embed continuous human checkpoints into AI workflows — short reads, quick approvals, flagged exceptions. The AI keeps producing, but the human judgment layer operates concurrently, not sequentially. Reduce the orbit, reduce the dilation.
Protect the Protectors
The people doing capacity protection — PMs, tech leads, architects, QA — are the ones most affected by time dilation. They spend their days coordinating, not producing. That means they're constantly off the surface while the team accelerates. Give them tools, dashboards, and AI summaries that compress the gap. They shouldn't have to read every Slack thread to protect the project.
Accept the Schwarzschild Radius
Every system has a limit. Every person has a limit. Every AI subscription has a limit. Anthropic accepted theirs this week. Your project manager accepted theirs years ago. The most dangerous thing you can do is pretend the limit doesn't exist — adding one more feature, one more task, one more question to a system that's already at capacity. The physics always win.
Don't Let Me Leave, Murph
"I'm not afraid of death. I'm an old physicist. I'm afraid of time."
— Dr. Brand, Interstellar
Cooper's tragedy wasn't that he went to space. It was that he understood the cost and went anyway, because the mission mattered. He knew every hour on Miller's Planet would cost years. He factored it in. He made the call. And then he sat in the ship and cried watching what the math took from him.
The knowledge worker's version is quieter but structurally identical. You know that stepping away from the AI-accelerated flow — to ask a question, to protect scope, to take PTO, to eat lunch, to think — will cost you. The world dilates while you're on the surface. The Slack threads accumulate. The decisions get made. The codebase evolves. And when you come back, you're Cooper watching the videos.
Anthropic learned this week what Cooper learned on Miller's Planet: you cannot serve infinite demand with finite capacity. The GPUs are the Endurance. The tokens are the time. The throttle is the acknowledgment that physics applies even to companies valued at billions of dollars.
And your project manager — the one who says "what's the priority?" and "that's scope creep" and "let's take that offline" — has been solving the same equation since before AI existed. They were always protecting capacity. They were always managing time dilation. They just didn't have the physics metaphor.
Now they do. It's called Gargantua. And it's the black hole at the center of every AI-augmented workplace, bending time around it, making every hour on the surface cost seven years on the ship.
The question isn't whether the dilation exists. It does. The question is whether you'll account for it — like Cooper did — or pretend the math doesn't apply.
The physics always win.
The Human Token Economy Series
P.S. I wrote this post today, March 31, 2026 — the same day Anthropic admitted their Claude Code quotas are draining faster than expected. I burned through a meaningful chunk of my own Claude session writing it. There's a non-zero chance I hit my own rate limit before I finish editing. If that's not the Gargantua effect in action, I don't know what is.
Related Posts
Be Kind, Rewind: The AI-Accelerated Workplace Has a Re-Entry Problem
Your team produced 50x the output while you were on PTO. Your human context window is 4K-8K tokens. The delta waiting for you is 2.5 million. An interactive, role-personalized deep dive into the re-entry crisis nobody is designing for — and the organizational playbook to fix it.
Limitless: The Human Token Economy
What if we measured human work output in tokens? The average knowledge worker produces ~237,000 tokens per month — emails, meetings, docs, analysis. At Claude Opus 4.6 API rates, that costs $5.93. Your salary costs $9,407. You are a 1,585x markup. An interactive, role-personalized deep dive into the economics of human cognition, AI adoption psychology, and why a $20/month subscription is the highest-ROI investment in business history.
And Now You Know the Rest of the Story: Who Really Should Pay for Your AI Subscription
Who should pay for AI subscriptions — the employee or the business? The obvious answer is always wrong. Paul Harvey knew: mechanics, teachers, nurses, truck drivers, and one accountant going through a divorce prove that you can't make the right call without the rest of the story.