Claude Code users hitting usage limits 'way faster than expected'
samizdis
293 points
178 comments
March 31, 2026
Related Discussions
Found 5 related stories in 55.9ms across 3,471 title embeddings via pgvector HNSW
- Claude Code users hitting usage limits 'way faster than expected' steveharing1 · 18 pts · April 02, 2026 · 90% similar
- Claude usage limits hitting faster than expected Austin_Conlon · 11 pts · March 31, 2026 · 82% similar
- Anthropic discourages Claud demand during peak productivity hours dude250711 · 15 pts · March 26, 2026 · 64% similar
- Claude Code adjusting down 5hr limits laacz · 27 pts · March 26, 2026 · 63% similar
- Anthropic: "During peak hours you'll move through session limits faster" CharlesW · 12 pts · March 26, 2026 · 63% similar
Discussion Highlights (20 comments)
elephanlemon
Yesterday (pro plan) I ran one small conversation in which Claude did one set of three web searches, a very small conversation with no web search, and I added a single prompt to an existing long conversation. I was shocked to see after the last prompt that I had somehow hit my limit until 5:00pm. This account is not connected to an IDE or Code, super confusing.
p2hari
I cancelled my pro plan last month. I was using Claude as my daily driver. In fact had the API plan also and topped it with $20 more. So it was around $40 each month. Starting from December last year it has been like this. When sessions could last a couple of hours with some deep boilerplate and db queries etc. to architecture discussion and tool selection. Slowly the last two months it just gets over. One prompt and few discussions as to why this and not that and it is done.
giancarlostoro
I'm guessing their newer models are taking way more compute than they can afford to give away. The biggest challenge of AI will eventually be, how to bring down how much compute a powerful model takes. I hope Claude puts more emphasis into making Haiku and Sonnet better, when I use them via JetBrains AI it feels like only Opus is good enough, for whatever odd reason.
Asmod4n
When asking it to write a http library which can decode/parse/encode all three versions of it the usage limit of the day gets hit with one sentence. In the pro plan. Even when you hand it a library which does hpack/huffmann.
lukewarm707
please tell me if i'm crazy. i just refuse to use openai/google/anthropic subscriptions, i only use open source models with ZDR tokens. - i like privacy in my work, and i share when i wish. somehow we accepted that our prompts and work may be read and moderated by employees. would you accept people moderating what you write in excel, google docs, apple pages? - i want a consistent tool, not something that is quantised one day, slow one day, a different harness one day, stops randomly. - unless i am missing something, the closed source models are too slow for me to watch what they are doing. i feel comfortable with monitoring something, usually at about 200-300tps on GLM 5. above that it might even be too fast!
stavros
Anthropic went about this in a really dishonest way. They had increased demand, fine, but their response was to ban third-party clients (clients they were fine with before), and to semi-quietly reduce limits while keeping the price the same. Unilaterally changing the deal to give customers less for the same price should not be legal, but companies have slowly boiled the frog in such a way that now we just go "welp, it's corporations, what can you do", and forget that we actually used to have some semblance of justice in the olden days.
robviren
I find Claude code to be a token hog. No matter how confidently the papers say context rot is not an issue I find curating context to be highly important to output quality. Manually managing this in the Claude Webui has helped with my use cases more than freely tossing Claude code at it. Likely I am using both "wrong" but the way I use it is easier for me to reason about and minimize context rot.
jdefr89
Over reliance on LLMs is going to become such a disaster in a way no one would have thought possible. Not sure exactly what, who, when, or where.. Just that having your entire product or repo dependent on a single entity is going to lead to some bad times…
shafyy
What is the best way to get start with open weight models? And are they a good alternative to Claude Code?
nprateem
I literally ran out of tokens on the antigravity top plan after 4 new questions the other day (opus). Total scam. Not impressed.
kneel
I asked it to complete ONE task: You've hit your limit · resets 2am (America/Los_Angeles) I waited until the next day to ask it to do it again, and then: You've hit your limit · resets 1pm (America/Los_Angeles) At which point I just gave up
ZeroCool2u
I'm finishing my annual paid Pro Gemini plan, so I'm on the free plan for Claude and I asked one (1) single question, which admittedly was about a research plan, using the Sonnet 4.6 Extended thinking model and instantly hit my limit until 2 PM (it was around 8 or 9 AM). Just a shockingly constrained service tier right now.
firebot
The first hit is free.
dinakernel
This turned out to be a bug. https://x.com/om_patel5/status/2038754906715066444?s=20 One reddit user reverse engineered the binary and found that it was a cache invalidation issue. They are doing some hidden string replacement if the claude code conversation talks about billing or tokens. Looks like that invalidates the cache at that point. If that string appears anywhere in the conversation history, I think the starting text is replaced, your entire cache rebuilds from scratch. So, nothing devious, just a bug.
1970-01-01
This has been verified as a bug. Naturally, people should see some refunds or discounts, but I expect there won't be anything for you unless you make a stink. https://old.reddit.com/r/ClaudeCode/comments/1s7zg7h/investi...
aliljet
There's a weird 'token anxiety' you get on these platforms. And you basically don't know how much of this 'limit' you may consume at any time. And you actually don't even know what the 'limit' is or how it's calculated. So far, people have just assumed Anthropic will do the kind thing and give you more than you could ever use...
pxtail
Recently after noticing how quickly limits are consumed and reading others complaints about same issue on reddit I was wondering how much about this is real error or bug hidden somewhere and how much it's about testing what threshold of constraining limits will be tolerated without cancelling accounts. Eventually, in case of "shit hits the fan" situation it can be always dismissed by waving hands and apologizing (or not) about some abstract "bug". The lack of transparency and accountability behind all of this is incredible in my perception.
ryan42
claude automatically enabled "extra usage" on my pro account for me (I had it disabled) and the total got to $49 extra before I noticed. I sent an email asking wtf but I don't expect much.
delphic-frog
The token usage differs day to day - that's the most frustrating part. You can't effectively plan a development session if you aren't sure how far you'll likely get into a feature.
spongebobstoes
try codex, it's really good and doesn't have the same limits issues