Top
Best
New

Posted by samizdis 7 hours ago

Claude Code users hitting usage limits 'way faster than expected'(www.theregister.com)
204 points | 137 comments
pxtail 5 hours ago|
Recently after noticing how quickly limits are consumed and reading others complaints about same issue on reddit I was wondering how much about this is real error or bug hidden somewhere and how much it's about testing what threshold of constraining limits will be tolerated without cancelling accounts. Eventually, in case of "shit hits the fan" situation it can be always dismissed by waving hands and apologizing (or not) about some abstract "bug".

The lack of transparency and accountability behind all of this is incredible in my perception.

vintagedave 4 hours ago||
I've run into this, and I highly doubt I am one of the more extraordinary users. I have delays between working with it, don't have many running at once, am running on smaller codebases, etc. Yet just a few minutes ago I hit a quota. In the past I did far more work with it without running into the quota.

I emailed their support a few days ago with details, concerns, a link to the twitter thread from one of their employees, and a concrete support request, which had an AI agent ('Fin') tell me:

> While our Support team is unable to manually reset or work around usage limits, you can learn about best practices here. If you’ve hit a message limit, you’ll need to wait until the reset time, or you can consider purchasing an upgraded plan (if applicable).

I replied saying that was not an appropriate answer.

You're absolutely right re the lack of transparency and accountability. On one hand, Anthropic generates good will by appearing to have a more ethical stance then OpenAI, and a better product. On the other hand, they kill it fast through extremely poor treatment of their customers.

If they have a bug, they need to resolve it: and in the meantime refund quotas. 'Unable to' - that's shocking. This is simple and reasonable. It's basic customer service. I don't know if they realise the damage their attitude is doing.

Kim_Bruning 3 hours ago||
Fin is the most useless thing ever. There's no obvious way to get reports in front of a human in a timely manner, and there's no clue to believe fin interactions are retained.

This does mean ultimately no loyalty. I can't stay loyal to a brand that doesn't actually respond to inquiries, bug reports or down reports at all.

I do understand that Anthropic is operating at a tremendous scale and can't have enough humans in the loop. This sounds like a good use for ai classification and triage, really!

traceroute66 21 minutes ago||
> I can't stay loyal to a brand that doesn't actually respond to inquiries, bug reports or down reports at all.

Amen to this.

Being in business means having to respond to customer enquiries at some point.

Given the amount of billions being pumped into Anthropic's pockets and given the millions their senior-leadership no doubt pay themselves, I'm sure they could spare a bit of cash to get off their backsides and sort out the Customer Service.

I simply do not buy the "poor Antropic, they are operating at scale, they are too busy winning to deal with customer service" argument that comes up time and time again.

The fact is there are many large businesses, many large governments that are able to deal with customers "at scale".

Scale means you respond a bit slower, maybe a few days or at most a couple of weeks AT MOST. But complete silence for months or years is inexcusable.

All of my experiences with "Fin" matches that of my friends and colleagues .... namely that "Fin" is a synonym for "black hole". I've got "tickets" opened with "Fin" months ago that have not had a modicum of reply.

joshuak 48 minutes ago|||
It is also interesting to observe that your most valuable accounts in this kind of pricing model are the ones that are least used and therefore are not confronted by the limits. Heavy users canceling their accounts in frustration is a win for Anthropic not a punishment, at least a short term.
JambalayaJimbo 4 hours ago|||
Once you get used to using claude as an abstraction layer you start getting pretty reckless with it.

My organization has the concept of "premium models" where our limits reset every month. I hit my limit pretty quickly last month because I was burning tokens doing things that would have been a simple bash loop in the past - all because I was used to interfacing with Claude at the chat layer for all my automation needs and not thinking any more about it.

devmor 4 hours ago||
This is a real danger that I think a lot of people will run into as prices go up more and more in the future.

Completely outside of the productivity debate, offloading cognitive tasks to LLMs leaves you less practiced in them and less ready to do them when the LLM isn't available. When you have to delegate only certain tasks to the LLM for financial reasons, you may find yourself very frustrated.

foxyv 2 hours ago|||
I suspect that Claude had a bug that undercounted tokens and they fixed it.
mmmlinux 1 hour ago||
I wonder if that was why they were offering the bonus off hours limits. Ease people in to the transition.
joshuafuller 5 hours ago|||
This feels a lot like the same playbook we’re seeing with dynamic pricing in retail, just applied to compute instead of products. You never really know what you’re getting, and the rules shift under you.

What makes it worse is the lack of transparency. If there were clear, hard limits, people could plan around it. Instead it’s this moving target that makes it impossible to trust for real work.

At some point it stops feeling like a bug and starts feeling like a pricing experiment on users.

bayarearefugee 5 hours ago|||
The clear trend over the past decade or so has been using analytics and data gathering to extract maximum rents from every customer in every industry and AI is going to massively accelerate this.

The only way out is government regulation which means we are screwed in the US (our government is too far gone to represent average citizen interests in any meaningful way) but Europeans maybe have a chance if they get it together and demand change.

tartoran 5 hours ago|||
What a horrid glimpse in the future. I hope we won't get there and we all collectively fight back with our wallets.
ryandrake 4 minutes ago|||
It's going to get much worse. We're soon going to have enough data and compute (and are losing enough online privacy) to allow every company to apply personalized pricing down to the individual. My local restaurant is going to know that I am willing to buy a burger for at most $4.57 and my neighbor is only willing to pay $2.91 for it, and they will have the ability to charge us individually. Every business is going to soak each of us us to the maximum extent that the data says they can.
Tade0 5 hours ago|||
I'm worried that the present is actually living off a line of credit that will be spent/closed soon.
tjoff 1 hour ago|||
Working as intended? They openly state that how quickly your limit is reached depends on many factors (that you don't know) as well as current load on their systems.

Could just be that usage has gone up.

thisisit 4 hours ago|||
They keep running experiments like free $50 in extra use credits or 2x usage outside certain windows where inference is very slow. You can’t help but think this is all a slowly boiling the frog experiment. Experimenting how much they can charge.
nicce 5 hours ago||
Are they going to pay back if subscription was payed but token limit was less than advertised? Is there some tiny text somewhere preventing just suing or pulling money back with credit cards?
jadar 5 hours ago||
Part of the issue is that they don't actually advertise what the token limit is. Just some vague, "this is 5x more than free, and 5x more than pro". They seem to be free to change the basis however they please, because most of us are more than happy to use what they give us at the discounted subscription pricing.
midnightdiesel 47 minutes ago||
It seems like Anthropic is constantly changing the rules and pulling out rugs, and always entirely by surprise. I’m not sure if they’re incompetent or just careless, but I stopped paying them because of this a while ago, and my days are much more interesting and enjoyable using my own brain instead.
carefree-bob 46 minutes ago|
As long as they keep losing money and are reliant on investments to pay their operating expenses, they are going to be thrashing about in search of a sustainable business model and I don't blame them.
garrickvanburen 19 minutes ago||
Considering: - Anthropic decides how much a token is worth. - Users have no visibility or ability to control in how many tokens a given response will burn.

This is the only expected answer. https://forstarters.substack.com/p/for-starters-59-on-credit...

dinakernel 5 hours ago||
This turned out to be a bug. https://x.com/om_patel5/status/2038754906715066444?s=20

One reddit user reverse engineered the binary and found that it was a cache invalidation issue.

They are doing some hidden string replacement if the claude code conversation talks about billing or tokens. Looks like that invalidates the cache at that point.

If that string appears anywhere in the conversation history, I think the starting text is replaced, your entire cache rebuilds from scratch.

So, nothing devious, just a bug.

ibejoeb 4 hours ago||
> BUG 2: every time you use --resume, your entire conversation cache rebuilds from scratch. one resume on a large conversation costs $0.15 that should cost near zero.

I use it with an api key, so I can use /cost. When I did a resume, it showed the cost from what I thought was first go. I don't think it's clear what the difference is between api key and subscription, but am I believe that simply resuming cost me $5? The UI really make it look like that was the original $5.

replwoacause 5 hours ago|||
Nothing devious, but is Anthropic crediting users? In a sense, this is _like_ stealing from your customer, if they paid for something they never got.
arvid-lind 4 hours ago||
Not seeing any quota returned on my Pro account. My weekly usage went up to 20% in about one hour yesterday before I panicked and stopped the task. It was outside of the prime hours too which are supposed to run up your quota at a slower rate.
novaleaf 4 hours ago|||
your linked bug is a cherry pick of the worst case scenario for the first request after a resume.

While it should be fixed, this isn't the same usage issue everyone is complaining about.

kif 5 hours ago|||
Anecdotally when Claude was error 500'ing a few days ago, its retries would never succeed, but cancelling and retrying manually worked most of the time.
mook 4 hours ago|||
That is a summary and a picture of https://old.reddit.com/r/ClaudeAI/comments/1s7mkn3/psa_claud... it looks like?
TazeTSchnitzel 5 hours ago|||
That bug would only affect a conversation where that magic string is mentioned, which shouldn't be common.
dinakernel 4 hours ago||
I guess so - but for people working on billing section of a project or even if they include things like - add billing capability etc in Claude MD - it might be an issue, I think
pier25 5 hours ago||
https://xcancel.com/om_patel5/status/2038754906715066444
mszczodrak 4 minutes ago||
I've been hitting the API limit errors over Claude CLI, yet the total usage was 0% on the claude.ai website. Changing the model fixed the problem.
torginus 11 minutes ago||
I dunno, but CC might give away tokens for cheaper, but when I used Opus as standalone in Cursor, I definitely get way more mileage out of a token.

Considering how much progress I made vs how much I paid, I couldn't make a scientific assessement, but it felt pretty close.

p2hari 6 hours ago||
I cancelled my pro plan last month. I was using Claude as my daily driver. In fact had the API plan also and topped it with $20 more. So it was around $40 each month. Starting from December last year it has been like this. When sessions could last a couple of hours with some deep boilerplate and db queries etc. to architecture discussion and tool selection. Slowly the last two months it just gets over. One prompt and few discussions as to why this and not that and it is done.
iwontberude 4 minutes ago||
I have had the exact same experience (like super uncanny with prices etc). And now feel like I can only use my Claude subscription for the most basic issues. I’m getting range anxiety.
ramon156 5 hours ago||
After they force OpenCode to remove their Claude integration, and the insane token hogging, I also cancelled my subscription.
aliljet 5 hours ago||
There's a weird 'token anxiety' you get on these platforms. And you basically don't know how much of this 'limit' you may consume at any time. And you actually don't even know what the 'limit' is or how it's calculated. So far, people have just assumed Anthropic will do the kind thing and give you more than you could ever use...
sumtechguy 5 hours ago||
This reminds me of the early days of cell phones. Limits everywhere and you paid for it by the kilobyte. Think at one point I was paying 45c per text message. I hope this gets better and we do not need gigawatt datacenters to do this stuff.
jauntywundrkind 3 hours ago||
Yeah, I've been juggling some patches to opencode to help me see where my codex usage limits are at. As of a month ago, that information was not visible on the ChatGPT web UI.

You just work till suddenly the AI dumps you out, and sit there wondering how many hours or days you have to wait. It's incredible that this experience is at all ok, is accepted

robviren 5 hours ago||
I find Claude code to be a token hog. No matter how confidently the papers say context rot is not an issue I find curating context to be highly important to output quality. Manually managing this in the Claude Webui has helped with my use cases more than freely tossing Claude code at it. Likely I am using both "wrong" but the way I use it is easier for me to reason about and minimize context rot.
elephanlemon 6 hours ago|
Yesterday (pro plan) I ran one small conversation in which Claude did one set of three web searches, a very small conversation with no web search, and I added a single prompt to an existing long conversation. I was shocked to see after the last prompt that I had somehow hit my limit until 5:00pm. This account is not connected to an IDE or Code, super confusing.
master_crab 6 hours ago|
Tool calls (particularly fetching for context) eats the context window heavily. I explicitly send MCP calls to sub agents because they are so “wordy”.
bensyverson 6 hours ago||
Everyone who has not hit this bug thinks it’s user error… It’s not. It happened to me a few days ago, and the speed at which I tore through my 5 hour usage cap was easily 10x faster than normal.

Also: sub agents do not get you free usage. They just protect your main context window.

dmd 4 hours ago|||
I'm on Max. This morning, just to test, before doing anything else whatsoever, I was at 0%, and I typed 'test one two three' into CC.

That put me at 12%.

I have no MCPs except the built in claude-in-chrome.

This is clearly a bug.

piva00 6 hours ago||||
Don't they consume less of the token quota in case the subagents are running cheaper models like Sonnet and Haiku compared to Opus?
bensyverson 5 hours ago||
Correct—I just wouldn't want folks to mistakenly think that the context fill % corresponds 1:1 with session token use.
master_crab 5 hours ago|||
Yes, sorry. I meant it more as a descriptor of how many tokens it consumes. You are still stuck burning money.
More comments...