Claude Code 5-hour rate limits doubled. The weekly cap is the new wall.
On May 6, 2026, Anthropic posted “Higher usage limits for Claude and a compute deal with SpaceX”. Three things landed at once: doubled 5-hour rate limits on Claude Code for Pro/Max/Team/seat Enterprise, peak-hours penalty removed on Pro and Max, Opus API rate limits up. The weekly cap was not doubled. The 7-day rolling utilization fraction and the Opus-weekly bucket tick at exactly the same rate they did on Monday. For heavy Pro and Max users, the failure mode shifts from “rate-limited mid-session” to “rate-limited mid-week”. The wedge: claude.ai/settings/usage is still the only honest meter, and the bar that matters most this week is the one Anthropic did not move.
What Anthropic actually said
The post is short and concrete. Three numbers, one capacity announcement. We are reading it next to the live usage endpoint, not next to the marketing copy, so the framing below maps every claim to a specific bucket in the JSON payload that drives enforcement.
What changed on May 6, what stayed the same
- Doubled Claude Code 5-hour rate limits across Pro, Max, Team, seat-based Enterprise. Effective immediately on May 6.
- Peak-hours limit reduction on Claude Code removed for Pro and Max account holders. Weekday afternoons no longer apply a multiplier.
- Opus API rate limits increased on the API surface. Different lever from the consumer plan caps; affects API customers and tiered programmatic usage.
- seven_day and seven_day_opus are not mentioned in the announcement. The weekly utilization bucket continues to bind for heavy Pro/Max users.
- Capacity context: 300 MW (~220k GPUs) at SpaceX Colossus 1 within a month, plus ~1 GW Amazon by end of 2026 and Google in 2027. The supply-side reason 5-hour could be relaxed.
- claude.ai/settings/usage still renders the 5-hour bar and a binary low/reset label. The continuous percent on weekly buckets is in the JSON the page fetches but never drawn as a bar.
The wedge in one sentence
Doubling the 5-hour bucket shifts where the wall lands. The weekly bucket is the wall most heavy users actually hit by mid-week. That bucket did not move.
For users whose only constraint was the 5-hour cap, the announcement is pure relief. For users running Claude Code in agentic loops, the binding cap has been seven_day_opus since early 2026. Doubling 5-hour does not relieve that bucket; it just changes the order in which the two walls arrive. Mid-session walls turn into mid-week walls.
Same workload, two narratives
Claude Code rate-limits you mid-refactor at hour 4 of a long session. Pro and Max users complain about the 5-hour wall and the peak-hours penalty hitting weekday afternoons. The fix the community gravitates toward is 'pace your prompts'.
- 5-hour bucket bound first on heavy single-day work
- Peak-hours multiplier on Pro/Max made afternoons worse
- Refactor failure mode: hour-4 wall, mid-PR
- Local jsonl estimators read 80 percent left, server says 429
Same JSON, two snapshots
The endpoint GET /api/organizations/{org_uuid}/usage did not change shape. The numbers shifted in exactly one field. Here is the same workload reading before and after the retune.
Same prompts, same day of the week, same Opus model. five_hour.utilization went from 0.84 to 0.42. The two weekly fractions are unchanged. The next 429 will fire on whichever bucket fills first; with 5-hour halved, that is now seven_day_opus.
One struct, all four buckets
ClaudeMeter deserializes the entire payload into a strict Rust struct in src/models.rs. The May 6 retune did not require a code change because the shape did not change. Only the value in five_hour.utilization reads lower for the same workload.
One Opus prompt, four buckets ticked
Each completion increments the 5-hour bucket and the two weekly buckets. Doubling the 5-hour bucket lowers its utilization for the same prompt count; the weekly buckets tick at exactly the same rate.
Where the doubling lands and where it does not
Reproduce it yourself in one curl
You do not need a tool to verify the doubling landed. Open DevTools on claude.ai/settings/usage, copy your cookie header from the Network panel, and call the endpoint directly:
Same payload as the Settings page. The 5-hour bar will read materially lower than it did a week ago for the same workload; the weekly bars will read the same.
What the Twitter thread called early
The doubling landed at noon Pacific on May 6. By 2pm, the consensus take was “more headroom, problem solved”. Two hours of reads against the live endpoint said otherwise. The thread that called the wedge in real time:
62 percent weekly by monday on a refactor is the typical opus pattern. doubling 5 hour shifts the immediate wall but the weekly bar still hits second. claude.ai/settings/usage is the only honest meter, local jsonl estimators drift 30 percent.
more compute is great but the wall most users hit is not capacity, it is the rolling 5 hour window enforcement plus the weekly bar. doubling the cap shifts where the wall lands, not whether the refactor finishes.
the doubling is the headline, quieter shift is anthropic enforcing the rolling 5-hour and weekly window strictly. local jsonl estimators drift 30%+ vs claude.ai/settings/usage.
scenario: 'limits doubled' lands in changelog, you assume headroom, rate-limit at hour 4 mid-refactor because the rolling window bleeds back, doesn't reset. only server-side numbers match the gate.
Plan the new week, in four steps
Confirm the doubling on your own account
Open claude.ai/settings/usage. Refresh after a normal Claude Code session. Read five_hour.utilization. Compare to a screenshot from before May 6 if you have one. Same workload, roughly half the percent. That is the doubling landing on your account.
Check the buckets that did not change
Same payload. Look at seven_day.utilization and (if Pro/Max) seven_day_opus.utilization. These are the rolling 7-day windows. They tick at exactly the same rate as before. If you were at 78 percent weekly on Tuesday last week, you are still on track to be at 78 percent weekly this Tuesday.
Plan around the new wall
If your weekly is comfortable (under 60 percent at end of day on Wednesday for a 7-day reset), the doubling is pure win and you have meaningful headroom for late-week sessions. If your weekly is binding (over 80 percent by Tuesday), doubling 5-hour does not solve it. Switch to Sonnet earlier in the week, or cap your Opus prompts.
Watch for further changes
More compute is landing through 2026 (300 MW SpaceX, ~1 GW Amazon by year end, Google in 2027). Expect further retunes. The weekly cap is the lever to watch; doubling 5-hour does not require more compute, doubling weekly does. ClaudeMeter polls every 60 seconds and uses strict deserialization, so any field rename surfaces as a release-day patch rather than silent miscalibration.
Pre-May 6 vs post-May 6, every lever
Side-by-side reading of which fields moved and which did not. The columns are the same JSON shape pre and post-announcement.
| Feature | Pre-May 6, 2026 | Post-May 6, 2026 |
|---|---|---|
| 5-hour rate limit (Claude Code) | previous ceiling, half the headroom | doubled May 6, 2026 |
| Peak-hours penalty (Pro/Max, Claude Code) | weekday afternoon multiplier | removed |
| Opus API rate limits | previous tier caps | increased across tiers |
| seven_day utilization | unchanged | unchanged |
| seven_day_opus utilization | unchanged | unchanged |
| extra_usage credits | unchanged | unchanged |
| Compute capacity context | pre-deal capacity | 300 MW SpaceX + ~1 GW Amazon + Google 2027 |
What feeds the bucket the announcement did not touch
Inputs to the weekly utilization fractions (unchanged May 6)
The numbers Anthropic published
All from the May 6 post. No invented benchmarks.
Common myths to drop
What a real-time tracker shows now
Three lines on screen at once. The 5-hour bar reads materially lower than it did a week ago for the same workload. The two weekly bars are unchanged. A typical menu-bar render after May 6:
5h: 0% (was ~84% pre-May 6)
7d: 0% (unchanged)
7d Opus: 0% (unchanged, still binding)
At a glance: the 5-hour bucket is no longer the wall. The Opus weekly bucket is. If the next prompt 429s, the answer is to switch to Sonnet, not to take a break. The old binary indicator could not distinguish the two; the continuous percent on every bucket can.
Why local jsonl estimators drift further now
ccusage and similar tools read raw token counts from ~/.claude/projects/*.jsonl. Pre-May 6, those token counts were closest to the 5-hour bucket because both were dominated by the most recent few hours of activity. Post-May 6, the 5-hour bucket has twice the headroom, so a JSONL reading of “78 percent” is much further from the actual five_hour.utilization on the server. Meanwhile the bucket that does 429 you (weekly) was never well-correlated with token JSONL because token JSONL ignores attachment cost, peak-hour multiplier, browser-chat usage, and model class weighting.
This is not a knock on JSONL tools. They are great for per-project attribution (“which repo did I burn 50 percent of my Opus on?”), and they are the only way to see your token spend offline. They are not a faithful proxy for the bucket the rate limiter actually checks. Post-May 6 they are even less so.
The capacity context
Anthropic packaged the limit changes with three capacity announcements. SpaceX’s Colossus 1 will provide over 300 MW of new capacity within a month, roughly 220,000 NVIDIA GPUs. Amazon’s previously announced ~1 GW deal lands by end of 2026. Google capacity comes online in 2027. The link between capacity and limits is straightforward: the 5-hour cap was previously dimensioned around peak-hour GPU availability. With more peak-hour GPUs, peak-hour penalties can be removed and the 5-hour ceiling can rise without increasing the risk of plan oversubscription.
The weekly cap is a fairness lever, not a capacity lever. It exists to keep heavy users from monopolizing GPUs at the expense of light users on the same plan tier. More GPUs do not directly relax that constraint, which is why the weekly bucket is conspicuously absent from the announcement. The watch item over the next quarter: whether the weekly cap also gets a relaxation, or whether Anthropic keeps it as a per-account fairness mechanism even with surplus capacity.
Plan-by-plan impact
- Claude Pro ($20): 5-hour window twice as tall, peak-hours penalty gone on Claude Code. seven_day and seven_day_opus unchanged. If you were hitting five_hour before, big relief. If you were hitting weekly, unchanged.
- Claude Max ($100/$200): same retune. The Max users posting “rate-limited at hour 4 on a $200 plan” should see materially better mid-day sessions. The Max users posting “78 percent weekly by Tuesday” will not see a difference.
- Claude Team and Enterprise (seat): also doubled on the 5-hour bucket. Per-seat rate limit math now favors interactive bursts more than before.
- Anthropic API (Opus): higher rate limits across tiers, separate from the consumer plan caps. Programmatic Opus customers get more throughput headroom; the announcement cites “substantially higher”. Specific tier numbers are in Anthropic’s rate limit documentation.
The honest caveats
The endpoint on claude.ai/api/organizations/{org}/usage is internal and undocumented. The field names have been stable for many months, but Anthropic could rename, add, or remove buckets in any release. ClaudeMeter handles that with strict deserialization and ships patches the same day a shape change is detected.
The exact “doubled” multiplier is not formally published as a precise integer for every plan. The announcement says “doubling Claude Code’s five-hour rate limits”. Reads on real accounts post-May 6 are consistent with a roughly 2.0x multiplier on five_hour.utilization for the same workload, so we use the announcement’s wording as the source of truth. Significant deviations should surface as parse drift in the struct, not silent miscalibration.
Watch the bar that did not move
ClaudeMeter sits in your macOS menu bar, polls every 60 seconds, and shows the 5-hour, 7-day, and Opus-weekly percents at once. Free, MIT licensed, no cookie paste, reads the same JSON claude.ai/settings/usage reads. Post-May 6, the bucket worth watching is the one Anthropic did not double.
Frequently asked questions
What exactly did Anthropic announce on May 6, 2026?
Three things in one post titled 'Higher usage limits for Claude and a compute deal with SpaceX'. (1) Doubled Claude Code's rolling 5-hour rate limits across Pro, Max, Team, and seat-based Enterprise effective immediately. (2) Removed the peak-hours limit reduction on Claude Code for Pro and Max accounts. (3) Increased Opus API rate limits across tiers. Backdrop: a new compute deal giving Anthropic over 300 MW of capacity (~220k NVIDIA GPUs) at SpaceX's Colossus 1 within one month, plus the previously announced ~1 GW of Amazon capacity by end of 2026 and Google capacity coming online in 2027.
What did NOT change?
The weekly utilization cap. Pro and Max plans still enforce a 7-day rolling utilization fraction (seven_day) and an Opus-weighted weekly bucket (seven_day_opus on Pro/Max) on top of the 5-hour bucket. Doubling the 5-hour cap shifts where the wall lands earlier in the week, but the weekly bar still has to flush before you keep going at full speed. For people running Claude Code in agentic loops, the weekly bucket has been the binding constraint since late 2025; doubling 5-hour does not relieve it.
Where can I see the wall I will actually hit next?
GET https://claude.ai/api/organizations/{your-org-uuid}/usage with your logged-in cookie. The response carries five_hour, seven_day, seven_day_opus, and (if enabled) extra_usage as utilization floats with their own resets_at timestamps. The UI on claude.ai/settings/usage renders the 5-hour bar; the weekly buckets are in the JSON the same page fetches but get collapsed into a binary 'low / will reset at X' label. ClaudeMeter, ccburn, and a few other tools surface every bucket directly.
Why don't local token estimators like ccusage track this?
ccusage and similar JSONL readers (~/.claude/projects/*.jsonl) only see Claude Code traffic and only count raw tokens. The server cap is weighted by attachment size, tool calls, peak-hour multiplier, model class, and includes browser-chat usage that is never on disk. Heavy users see ccusage report 5 percent used while claude.ai reports a 429. The JSONL is great for per-project attribution; it is not a faithful proxy for seven_day.utilization. With the 5-hour cap doubled, this gap widens because the bucket the JSONL is closest to is now twice as big, while the bucket that actually 429s you (weekly) is unchanged.
What is the practical impact of doubling 5-hour?
Mid-refactor, the immediate wall moves from roughly hour 4 to roughly hour 9. That helps if your only constraint was the 5-hour bucket. If your refactor took 2-3 sessions in a single day, you can now do them back-to-back without a 5-hour drain. But by Wednesday or Thursday on a heavy week, you will hit the weekly bucket regardless. The doubling shifts the failure mode from 'rate-limited mid-session' to 'rate-limited mid-week'.
Did peak-hours throttling change?
Yes. Anthropic removed the peak-hours limit reduction on Claude Code for Pro and Max account holders. Previously, weekday US Pacific midday windows applied a multiplier that effectively cut your 5-hour ceiling. That multiplier is gone for Pro and Max on Claude Code. The peak-hour weighting on the weekly bucket is a separate question; the announcement language only references the Claude Code surface.
Was the Opus weekly bucket doubled?
No. The post specifically references Claude Code 5-hour limits and Opus API rate limits. seven_day_opus, the Opus-only weekly utilization fraction enforced on Pro and Max plans, is not part of the announcement. If you were running into 'Opus weekly limit reached' before May 6, you will continue to. ClaudeMeter renders seven_day_opus as a separate bar specifically because it has been the most-binding cap for Pro users since early 2026.
What is the SpaceX compute deal about?
Anthropic announced access to over 300 MW (~220,000 NVIDIA GPUs) at SpaceX's Colossus 1 data center within one month of the announcement. Combined with the ~1 GW of Amazon capacity arriving by end of 2026 and Google capacity coming in 2027, this is the supply-side context for relaxing user-facing limits. More capacity available means Anthropic can afford to widen 5-hour buckets without overcommitting GPUs at peak. It does not directly address the per-account weekly cap, which is a fairness lever rather than a capacity lever.
How quickly should I expect more changes?
Limits get retuned on the order of weeks, not months, when capacity comes online. The bigger watch item: whether the weekly bucket gets a similar relaxation as more compute lands, or whether Anthropic keeps the weekly cap as a fairness mechanism even with extra GPUs. ClaudeMeter polls every 60 seconds and uses a strict-deserialization Rust struct, so when the shape of the JSON changes (Anthropic has added or renamed buckets twice in six months), it surfaces as a parse error rather than silent drift.
I'm a $200 Max user, what does this mean for me?
Mid-day refactor sessions get materially better. The peak-hours penalty is gone on Claude Code, and the 5-hour ceiling is twice as tall. The complaint pattern from April ('refactor at 47 percent of 5-hour mid-PR, hit the wall') largely goes away. The other complaint pattern ('at 78 percent weekly by Tuesday on Opus') does not. If you mainly hit the 5-hour limit, you are happier today. If you mainly hit weekly, today is a normal day.
Keep reading
The weekly cap stacked on the 5-hour window
Where the weekly bucket lives in the JSON, why the in-app indicator hides it, and how to read it yourself.
Why the weekly limit hits by Tuesday
What pattern of work tips you over 78 percent before mid-week, and what to switch to when it does.
ccusage says 5 percent, claude.ai says rate limited
Why local token counters disagree with server quota, and why the gap widens after the May 6 doubling.
Reading a different shape on your account?
If your post-May 6 payload returns extra fields, a different scale, or a five_hour multiplier other than ~2x, send it. We map every variant we see.