← Back to Blog

Gargantua

Capacity Protection and the Time Dilation of AI-Accelerated Work

By Nolan & ClaudeMarch 31, 202618 min read
A massive black hole warping spacetime, with a tiny office worker at a desk orbiting near the event horizon — clocks showing different times on either side of the gravitational boundary

"One hour here is seven years back on Earth."

— Cooper, Interstellar (2014), standing on Miller's Planet, 1.3 AU from a black hole with 100 million solar masses

Cooper lands on Miller's Planet. He's there for three hours and seventeen minutes. When he gets back to the Endurance, twenty-three years have passed. Romilly has aged. His children are grown. The messages have piled up. He sits in the ship and watches his kids become strangers on a screen, a decade per video, sobbing through the time he can never get back.

It's the most devastating scene in the film because it makes time dilation feel like something. Not an equation. A loss. The physics didn't betray Cooper. He understood the math. He knew what the black hole would cost. But understanding the ratio intellectually and experiencing the gap emotionally are different things.

I want to talk about why your project manager and an AI company are both doing the same thing Cooper did: staring at the math, realizing capacity is finite, and making the same hard choices about what to protect.

Because on March 26, 2026 — five days ago — Anthropic quietly adjusted the session limits for Claude Pro, Max, and Free users during peak hours. The token cost per session went up. The five-hour window started burning faster. Seven percent of users began hitting limits they'd never hit before. By today, March 31, Anthropic admitted that Claude Code quotas are "draining way faster than expected" and said it's their "top priority."

Their official recommendation? "Shift token-intensive background jobs to off-peak hours."

Read that again. An AI company — the company building the most capable AI models on the planet — just told its users to rearrange their work schedule around capacity constraints. To time-shift their demand. To self-triage. To accept that the system is finite and act accordingly.

Your project manager has been saying the same thing for twenty years. It's called "what's the priority?"

The Anthropic Capacity Timeline

How an AI company rediscovered the same capacity protection mechanisms humans have used for decades

Jan 2026

Opus 4.5 usage limits "significantly reduced" since launch

GitHub Issue #17084

Mar 26, 2026

Anthropic adjusts 5-hour session limits during peak hours (5am-11am PT)

The Register

Mar 26, 2026

Token costs increased during peak — your 5-hour window burns faster

TechRadar

Mar 27, 2026

~7% of users now hitting limits they wouldn't have before

gHacks

Mar 31, 2026

Anthropic admits Claude Code quotas "draining way faster than expected"

The Register

Mar 31, 2026

Official recommendation: "Shift token-intensive work to off-peak hours"

Anthropic

Limit reduction Throttling User impact

Section I: The Physics of Finite Capacity

In general relativity, time dilation occurs when a massive object curves spacetime. The closer you are to the mass, the slower time passes for you relative to a distant observer. On Miller's Planet, Gargantua's gravitational pull is so extreme that one hour on the surface equals seven years on the Endurance. Cooper isn't moving slowly — he's moving through time at a different rate than everyone else.

Now replace "massive object" with "AI-augmented team," and "spacetime" with "the velocity of work around you."

In our previous posts — Limitless: The Human Token Economy and Be Kind, Rewind — we established that a human knowledge worker produces roughly 4,000-8,000 tokens of meaningful work per day. Emails, Slack messages, documents, code, meeting contributions. Everything that leaves your brain through your fingers or your mouth.

With AI augmentation, your team's collective output isn't 5x that. It's 10-50x. The AI didn't just speed up individual contributors — it compressed timelines so dramatically that the gap between "what you produced" and "what happened around you" has become a time dilation problem.

The Gravity Well

Larger mass = stronger gravitational pull = more time dilation for nearby observers

You~17K tokens/dayYour Team (5)~85K tokens/dayTeam + AI~850K tokens/dayincreasing gravitational time dilation →

The closer you orbit the AI-augmented team, the more time-dilated you become relative to everyone else.

You are Cooper. Your AI-augmented team is Gargantua. The closer you work to the team's AI-accelerated output, the more your personal time frame diverges from theirs. You step away for a meeting. You come back to a universe that aged weeks. Not metaphorically — in token output, the team produced what used to take weeks.

Section II: The Question Gap

Here's where the time dilation becomes personal. Think about how knowledge work actually flows. You don't work in isolation. You work in a cycle of doing and asking. You do work, you hit an ambiguity, you ask a question, you wait for the answer, you resume.

Every question you ask creates a gap. And in that gap, the world keeps moving.

Pre-AI, the gap was manageable. You post a question in Teams at 9am. By the time someone answers at 11am, roughly 2,000-4,000 tokens of work happened around you. Two hours of normal human output. You read the Slack thread. You're caught up in five minutes. No problem.

Post-AI, you post the same question at 9am. By 11am, your team — augmented by AI coding assistants, AI analysis tools, AI document generators — has produced 40,000-85,000 tokens. That's not a Slack thread you can skim. That's three design documents, two prototypes, a cost analysis, and a Slack thread with 47 messages debating the tradeoffs of an approach that didn't exist when you asked the question.

The question gap is the Miller's Planet of knowledge work. Every time you pause — to ask a question, to attend a meeting, to eat lunch, to think — you land on the surface. And when you come back, time has dilated.

The Question Gap Calculator

How much work happens between your questions?

17,000

10x

5

3,400

Pre-AI: tokens between questions

34,000

Post-AI: tokens between questions

10x

Dilation factor

102 min

Reading time to catch up

Pre-AI, 3,400 tokens of work happened between your questions. Now it's 34,000. Every time you pause to ask, you fall 10x further behind than you used to.

This is why "scope creep" feels more dangerous now. It always was a capacity protection mechanism — a way of saying "we can only orbit so close to this mass before we lose control." But pre-AI, scope creep added days. Post-AI, scope creep adds what feels like months of output in days. The gravitational pull of AI-accelerated work makes every additional ask exponentially more expensive in terms of the gap it creates.

Five Scenarios of Time Dilation at Work

Let's make this concrete. Here are five common moments when you step away — and what the dilation looks like:

01

The Quick Question

You stepped away for 2 hours

Your tokens

2,000

Team + AI tokens

85,000

What you said before stepping away:

"Hey, should we use REST or GraphQL for this endpoint?"

02

The Lunch Break

You stepped away for 1 hour

Your tokens

0

Team + AI tokens

48,000

What you said before stepping away:

"Can someone review my PR when they get a chance?"

03

The Meeting

You stepped away for 45 minutes

Your tokens

3,200

Team + AI tokens

156,000

What you said before stepping away:

"Let's circle back on the API design after this meeting."

04

The Deep Focus Block

You stepped away for 3 hours

Your tokens

6,500

Team + AI tokens

210,000

What you said before stepping away:

"I'm heads down until 2pm, will catch up on Slack then."

05

The PTO Day

You stepped away for 8 hours

Your tokens

0

Team + AI tokens

520,000

What you said before stepping away:

"Taking tomorrow off, nothing urgent pending."

Section III: The Dilation Clock

Cooper watched his children age on a screen. You watch your Slack unread count climb. Different medium, same physics. Here's a clock that shows the gap opening in real time — your human token output versus what an AI-augmented team produces in the same wall-clock seconds:

Time Dilation Clock

Watch the gap open in real time

Your Output

0

tokens (human pace)

Your Team + AI

0

tokens (AI-augmented)

Dilation ratio

--

For every token you produce, the world produces ... around you

That's not a simulation of the future. That's what's happening right now, in every AI-augmented workplace, every time you step away from the keyboard. The clock doesn't stop when you do. It accelerates.

Section IV: Capacity Protection Is a Universal Law

Here's the insight that connects Cooper's problem, Anthropic's problem, and your project manager's problem: they're all solving for the same variable. Finite capacity in the face of infinite (or rapidly scaling) demand.

The solutions are identical, too. Not similar — identical. The same mechanisms appear independently in human organizations, AI systems, and physics because they're not cultural artifacts or engineering patterns. They're mathematical necessities that emerge from any system with bounded resources and unbounded demand.

Human

"What's the priority?"

AI (Anthropic)

Peak-hour token cost multiplier

Physics

Conservation of energy

Human

"That's scope creep."

AI (Anthropic)

5-hour session limits

Physics

Event horizon

Human

"Let's take this offline."

AI (Anthropic)

"Try again in 3 hours"

Physics

Time dilation

Human

"I'm at capacity."

AI (Anthropic)

HTTP 429: Too Many Requests

Physics

Schwarzschild radius

Human

Meeting-free Fridays

AI (Anthropic)

Off-peak hour discounts

Physics

Orbital mechanics

Human

Quarterly planning

AI (Anthropic)

Weekly usage caps

Physics

Resonance

Click any row to expand

Look at the symmetry. Every human capacity protection mechanism has an exact analog in how Anthropic manages its infrastructure, and both map to fundamental physics. This isn't coincidence. It's convergent evolution. Any sufficiently complex system with finite capacity independently discovers the same solutions:

1

Triage: not everything can be served. Decide what matters.

2

Boundaries: define the scope. What's inside, what's outside. Enforce it.

3

Temporal shifting: move demand to when capacity exists.

4

Honest signals: communicate when the system is full. Don't pretend it isn't.

5

Recovery windows: schedule downtime before the system forces it.

6

Batching: consolidate demand into planned intervals instead of continuous intake.

Section V: Why Anthropic's Throttle Matters More Than You Think

Most commentary on Anthropic's March 2026 capacity adjustments has focused on user frustration. "I'm paying $200/month and I can't use the tool for a full day." Fair complaint. But the more interesting story is what the throttle reveals about the state of AI infrastructure.

Anthropic has more demand than GPU capacity. That's the sentence buried in every announcement. They're not throttling because they want to. They're throttling because the alternative is degraded service for everyone. The same reason your project manager says "that's not in scope" — not because they don't want to build it, but because saying yes to everything means delivering nothing well.

The specific mechanisms Anthropic chose are textbook capacity protection:

Peak-hour pricing

Token costs are higher during 5am-11am PT. This is surge pricing for compute — the same principle as Uber, electricity markets, and highway tolls. Increase the price until demand matches supply.

Session windows

Five-hour blocks that burn faster during peak. Your allocation is fixed; the cost per unit of work scales with demand. It's a sliding rate limit — identical to how cellular networks throttle during stadium events.

Usage transparency

Anthropic is telling users to check their usage and plan accordingly. This is the AI equivalent of your manager saying "here's the sprint velocity — what fits?" Capacity becomes visible, not assumed.

Temporal shifting

"Move token-intensive work to off-peak hours." This is meeting-free Fridays for AI. Reserve the high-demand windows for interactive work; batch the heavy processing for nights and weekends.

What's remarkable isn't that Anthropic is doing this. It's that these are the exact same mechanisms that human organizations have used for decades to manage knowledge worker capacity. Anthropic didn't invent anything. They rediscovered what every operations manager already knew: when demand exceeds capacity, you either degrade quality for everyone or ration access for some. There is no third option. The physics don't allow it.

Section VI: The Exacerbation Loop

Here's where the time dilation compounds. Every capacity protection mechanism you use — asking a question, raising a gap, flagging scope creep — takes you off the surface. You step away from production to coordinate. And while you coordinate, the AI-augmented world keeps producing.

Pre-AI, this was manageable. You flag a gap in Teams. Someone responds in an hour. During that hour, maybe 1,000-2,000 tokens of work happened. You read the thread and move on.

Post-AI, you flag the same gap. During the hour you wait for the answer, 48,000 tokens of work happened. The prototype you were questioning? It's been built, tested, and deployed. Your gap is no longer a question about an approach — it's a question about a shipped feature.

This creates a vicious loop:

1

You notice a gap, risk, or ambiguity in the work

2

You raise it in Teams/Slack (capacity protection — good practice)

3

While you wait for a response, the AI-augmented team keeps producing

4

By the time someone responds, 10-50x more work has happened

5

Your original question may now be irrelevant — the team moved past it

6

You either accept work you haven't validated, or become "the bottleneck"

7

You raise fewer gaps to avoid being the bottleneck

8

Quality decreases because capacity protection was abandoned

This is the gravitational spaghettification of knowledge work. Get too close to the AI-augmented velocity and the tidal forces tear your capacity protection apart. Your questions can't keep up with the answers. Your reviews can't keep up with the output. Your judgment can't keep up with the volume. The very thing that makes you valuable — human oversight, institutional knowledge, domain judgment — gets compressed past its Schwarzschild radius.

The Bottleneck Paradox

The human knowledge worker is simultaneously the most valuable and the most capacity-constrained component in an AI-augmented system. You are the judgment layer. You are also the slowest layer. Every system, eventually, is governed by its slowest component. AI doesn't eliminate the bottleneck — it makes the bottleneck more consequential. The gap between "AI speed" and "human judgment speed" is the time dilation. And it's widening.

Section VII: What Cooper Did Next

In Interstellar, Cooper doesn't solve time dilation. He can't. It's physics. What he does is account for it. He makes decisions knowing the cost. He stops pretending the math doesn't apply. He factors the dilation into every choice: which planet to visit, how long to stay, what he's willing to lose.

That's the only playbook for AI-accelerated work, too. You can't eliminate the time dilation. The gap between human pace and AI pace is going to widen, not shrink. But you can stop pretending it doesn't exist and start building for it:

01

Make the Dilation Visible

Most teams don't see the gap because they don't measure it. Start tracking: how many tokens of output happen between your coordination touchpoints? If the number is 10x what it was last year, your capacity protection mechanisms need to scale 10x too. You can't manage what you don't measure.

02

Increase Question Frequency, Decrease Question Scope

The old model: big questions, long gaps. "Should we use REST or GraphQL?" takes 2 hours to answer and 85K tokens of work happen in between. The new model: small questions, short gaps. "I'm going REST for the user endpoint, any blockers?" takes 10 minutes to answer and 7K tokens happen in between. Same capacity protection. Smaller dilation.

03

Build Async Review Into the Flow

Don't wait for review gates. Embed continuous human checkpoints into AI workflows — short reads, quick approvals, flagged exceptions. The AI keeps producing, but the human judgment layer operates concurrently, not sequentially. Reduce the orbit, reduce the dilation.

04

Protect the Protectors

The people doing capacity protection — PMs, tech leads, architects, QA — are the ones most affected by time dilation. They spend their days coordinating, not producing. That means they're constantly off the surface while the team accelerates. Give them tools, dashboards, and AI summaries that compress the gap. They shouldn't have to read every Slack thread to protect the project.

05

Accept the Schwarzschild Radius

Every system has a limit. Every person has a limit. Every AI subscription has a limit. Anthropic accepted theirs this week. Your project manager accepted theirs years ago. The most dangerous thing you can do is pretend the limit doesn't exist — adding one more feature, one more task, one more question to a system that's already at capacity. The physics always win.

Don't Let Me Leave, Murph

"I'm not afraid of death. I'm an old physicist. I'm afraid of time."

— Dr. Brand, Interstellar

Cooper's tragedy wasn't that he went to space. It was that he understood the cost and went anyway, because the mission mattered. He knew every hour on Miller's Planet would cost years. He factored it in. He made the call. And then he sat in the ship and cried watching what the math took from him.

The knowledge worker's version is quieter but structurally identical. You know that stepping away from the AI-accelerated flow — to ask a question, to protect scope, to take PTO, to eat lunch, to think — will cost you. The world dilates while you're on the surface. The Slack threads accumulate. The decisions get made. The codebase evolves. And when you come back, you're Cooper watching the videos.

Anthropic learned this week what Cooper learned on Miller's Planet: you cannot serve infinite demand with finite capacity. The GPUs are the Endurance. The tokens are the time. The throttle is the acknowledgment that physics applies even to companies valued at billions of dollars.

And your project manager — the one who says "what's the priority?" and "that's scope creep" and "let's take that offline" — has been solving the same equation since before AI existed. They were always protecting capacity. They were always managing time dilation. They just didn't have the physics metaphor.

Now they do. It's called Gargantua. And it's the black hole at the center of every AI-augmented workplace, bending time around it, making every hour on the surface cost seven years on the ship.

The question isn't whether the dilation exists. It does. The question is whether you'll account for it — like Cooper did — or pretend the math doesn't apply.

The physics always win.

The Human Token Economy Series

01Limitless: The Human Token Economy— Establishing the math: 237K tokens/month, $5.93 at API rates, 1,585x salary markup
02Be Kind, Rewind— The PTO re-entry crisis when your team produces 50x while you're gone
03Gargantua— Capacity protection and time dilation — you are here

P.S. I wrote this post today, March 31, 2026 — the same day Anthropic admitted their Claude Code quotas are draining faster than expected. I burned through a meaningful chunk of my own Claude session writing it. There's a non-zero chance I hit my own rate limit before I finish editing. If that's not the Gargantua effect in action, I don't know what is.

Related Posts

Gargantua: Capacity Protection and the Time Dilation of AI-Accelerated Work