Claude has become practically unusable for Pro users in the past few days. The Opus 4.7 blew through an entire 5 hour limit in one question and didn’t even finish answering it. Zero value delivered.
Opus 4.6 is giving 2, maybe 3 questions before blowing through the Pro 5 hour limit as well. We are forced to use Sonnet which makes the same mistakes over and over and then to start trying with other companies. To make matters worse, it reuses old code as we try to survive between credit expiry so it re-introduced issues into the code with the limited credits, that we had already fixed on our own and with other models.
Anthropic in just a few days has gotten me to try GLM 5.1, the new Kimi, and back to OpenAI. OpenAI also seems to introduce new bugs without being carefully micromanaged. The advantage Claude has is that the models are more careful and can refactor code instead of leading to bloat as they go. But the throttling happening now is breaking things and making the entire subscription unusable. I really hope they fix it soon.
My personal LLM coding stack is now OpenCode, Claude Sonnet for ideation on spec with OpenWhispr for voice-to-text, GLM-5.1 for the orchestrating loop, GLM-4.7 for coding, and DeepSeek R1 for review and validation. It works much, much better than the Claude Code setup I have at work for substantially less money to boot.
At this rate I fully anticipate being able to run a comparable stack on a 128GB Mac Studio using quants of newer-generation distilled OSS models in a year or two. Being able to ramble to a computer for an hour about features and technical philosophy then have it build a nearly-working app for $50 is an exciting feeling. There's still a long tail of productionization and fixing what the model didn't adhere to but it's still incredible.
Im locked in for a year of claude pro, I encountered the same issues as you a couple weeks ago, Id get like one solid plan done and really really hope it was a 1 shot because that was legit all i was gonna get out of it for those 5 hours, and it would be ~10% of weekly usage to really make me feel scared to hit send.
I got the 20$ gpt tier, and now i just use claude to craft MD plan docs instead, and then i hand them off to gpt 5.4 and it has been working great. can do about 4x as much work or so based on my feelings(not accurate). if i have just small simple stuff to do i might still fire those off with sonnet and that seems plenty viable, but as soon as its an opus tier task i swap to this workflow.
Little annoying as now im kinda trying to manage a .claude/ and an .opencode/ folder but i kinda just have the .opencode/ stuff reference the .claude/ stuff so its a little less bleh.
I've been keeping within my usage because ive been in a funk a bit, but when i was slightly more worried id sorta just juggle whether claude or gpt would handle writing some initial tests as it did seem to kinda be imbalanced otherwise. seems like gpt just spam resets weekly usage throughout the week anyway so its prolly nbd.
> Claude has become practically unusable for Pro users in the past few days. The Opus 4.7 blew through an entire 5 hour limit in one question and didn’t even finish answering it
Glad I’m not the only one!
I’ve been limited so often this week I’ve setup half a dozen token compression tools in my workflow and had to do a crash course in token optimization.
Of course, it seems to only slightly delay the inevitable and doesn’t really solve the problem.
I wouldn't be surprised if folks start complaining to California government agencies like the Department of Consumer Affairs, and they take it seriously.
There is a lot of political capital to be earned by appearing to be "tough" on AI companies.
I thought it was just me. Even 4.6 is hitting limits so much sooner which I would have expected of 4.7
I hope people finally see why big tech is spending hundreds of billions in data center build out.
I have to guess that they're compute limited somewhere or the new models are incredibly overusing tokens, so I guess you need to wait for new data centers to come online?
I'm starting to think I've been A/B tested, because this was my experience for almost a year with Claude ever since I tried it for coding. Meanwhile, my coworkers seemed to be able to use it for long periods of time without getting rate limited.
One interesting variable is that I'm located in Vietnam while my coworkers are located in Norway and Europe.
To work around this issue I used Claude for coding with a Copilot subscription which was much cheaper and had virtually no rate limiting.
Copilot gives you some set amount of credits each month, but you can also pay as you go if you run out of credit which is much better than the 5 hour window crap claude code would give me.
The only opus model available now on copilot for some reason is 4.7 and it costs 7.5x tokens, while everything else is 1x, 0.33x or free.
But I switched to using GPT 5.4 medium for a month or so which I find very reasonable.