logoalt Hacker News

I cancelled Claude: Token issues, declining quality, and poor support

903 pointsby y42yesterday at 3:59 PM534 commentsview on HN

Comments

binyuyesterday at 5:45 PM

I feel like Anthropic is forcing their new model (Opus 4.7) to do much less guess work when making architectural choices, instead it prefers to defer back decisions to the user. This is likely done to mine sessions for Reinforcement-Learning signals which is then used to make their future models even smarter.

show 1 reply
exabrialyesterday at 7:00 PM

It's bad, really bad.

The filesystem tool cannot edit xml files with <name></name> elements in it

elevaetyesterday at 6:14 PM

I've been very happy using Codex in the VScode extension. Very high quality coding and generous token limits. I've been running Claude in the CLI over the last couple of months to compare and overall I prefer Codex, but would be happy with either.

hybrid_studyyesterday at 5:44 PM

Sometimes it feels like Anthropic uses token processing as a throttling tool, to their advantage.

_pdp_yesterday at 8:15 PM

Signup for all major providers (pro plan) and round-robin between all of them. This is the only way to protect against not having access to all of these heavily subsidised subscriptions. See what happened to Copilot.

yaloginyesterday at 5:30 PM

If someone wants to move off Claude what are the alternatives? More importantly can another system pick up from where Claude left off or is there some internal knowledge Claude keeps in their configuration that I need to extract before canceling?

show 1 reply
datavirtuetoday at 11:43 AM

I have enterprise plans for all AI services except Google. GitHub Copilot in VS Code is the best I have used so far. I hear a lot of complaints from people who are holding it wrong. In a single day I can have a beautiful greenfield app deployed. One dev. One day. Something that would have taken weeks with two teams bumping into each other. It's fully documented. Beautiful code. I read the reasoning prompts as it flows by to get an idea of what is going on. I work in phases and review the code and working product quickly after that. Minimal issues.

I'm an executive, the devs complaining are getting retrained or put on the chopping block.

My rockstars are now random contractor devs from Vietnam. The aloof FTE grey beards saying "I don't know, it doesn't work very good on X." Are getting a talking to or being sidelined/canned. So far most of my grey beards are adapting pretty well.

I'm not waiting on people to write code any more. No way in hell.

zendarryesterday at 4:33 PM

Seems like some of the token issues may be corrected now

https://www.anthropic.com/engineering/april-23-postmortem

show 2 replies
sreekanth850yesterday at 7:22 PM

Biggest issue i see is, models are not getting efficient. This is no where going to get commoditised. There will be a limit at which you can burn money at subsidised cost.

brachkowyesterday at 8:32 PM

As many others I had negative (not good as before) feeling about Claude Code lately

What I don't understand is these loud "voting with money" comments. What they are canceling is very subsidized plan to buy something that delivers a lot of value.

There are only two providers that can provide this level of models at very subsidized price - anthropic and openai. Both of them are bad in terms of reliability.

So I wonder what these people do after they "cancel" both of them? Do they see producing less result at same hourly rate as everyone else on the market as viable option?

mattasyesterday at 6:27 PM

I've see a post like this every week for the last 2 years. Are these models actually getting worse? Or do folks start noticing the cracks as they use them more and more?

DeathArrowyesterday at 4:37 PM

I use Claude Code with GLM, Kimi and MiniMax models. :)

I was worried about Anthropic models quality varying and about Anthropic jacking up prices.

I don't think Claude Code is the best agent orchestrator and harness in existence but it's most widely supported by plugins and skills.

show 1 reply
giancarlostoroyesterday at 4:38 PM

I'm torn because I use it in my spare time, so I've missed some of these issues, I don't use it 9 to 5, but I've built some amazing things, when 1 Million tokens dropped, that was peak Claude Code for me, it was also when I suspect their issues started. I've built up some things I've been drafting in my head for ages but never had time for, and I can review the code and refine it until it looks good.

I'm debating trying out Codex, from some people I hear its "uncapped" from others I hear they reached limits in short spans of time.

There's also the really obnoxious "trust me bro" documentation update from OpenClaw where they claim Anthropic is allowing OpenClaw usage again, but no official statement?

Dear Anthropic:

I would love to build a custom harness that just uses my Claude Code subscription, I promise I wont leave it running 24/7, 365, can you please tell me how I can do this? I don't want to see some obscure tweet, make official blog posts or documentation pages to reflect policies.

Can I get whitelisted for "sane use" of my Claude Code subscription? I would love this. I am not dropping $2400 in credits for something I do for fun in my free time.

show 3 replies
hedgehogyesterday at 4:42 PM

I used Opus via Copilot until December and then largely switched over to Claude Code. I'm not sure what the difference is but I haven't seen any of these issues in daily use.

nickdothuttonyesterday at 5:05 PM

Switched to local models after quality dropped off a cliff and token consumption seemed to double. Having some success with Qwen+Crush and have been more productive.

show 1 reply
chadlerivyesterday at 9:40 PM

Off topic: I do feel like this model switching content feels very circa 2010 "I'm quitting Facebook"

r0fltoday at 2:21 AM

I hope codex doesn’t decline the same way

I’m blown away by how good it is lately

sfmikeyesterday at 5:18 PM

i ran prompts used up a ton of usage, and got no return just showed error.

Asked support hey i got nothing back i tried prompting several times used a ton of usage and it gave no response. I'd just like usage back. What I payed for I never got.

Just bot response we don't do refunds no exceptions. Even in the case they don't serve you what your plan should give you.

caycepyesterday at 5:14 PM

If all Claude does is automate mundane code, why not just make a "meta library" of said common mundane code snippets?

show 2 replies
aleqsyesterday at 4:43 PM

The usage metering is just so incredibly inconsistent, sometimes 4 parallel Opus sessions for 3 hours straight on max effort only uses up 70% of a session, other times 20 mins / 3 prompts in one session completely maxes it out. (Max x20 plan) Is this just a bug on anthropic side or is the usage metering just completely opaque and arbitrary?

show 1 reply
AJRFyesterday at 9:57 PM

We are in the 'we need to IPO so screw our customers' phase of the cycle

SwellJoeyesterday at 5:37 PM

I don't get it. I use Claude Code every day, what I would consider pretty heavy usage...at least as heavy as I can use it while actually paying attention to what it's producing and guiding it effectively into producing good software. I literally never run into usage limits on the $100 plan, even when the bugs related to caching, etc. were happening that led to inflated token usage.

WTF are y'all doing that chews tokens so fast? I mean, sure, I could spin up Gas Town and Beads and produce infinite busy work for the agents, but that won't make useful software, because the models don't want anything. They don't know what to build without pretty constant guidance. Left to their own devices, they do busy work. The folks who "set and forget" on AI development are producing a whole lot of code to do nothing that needed doing. And, a lot of those folks are proud of their useless million lines of code.

I'm not trying to burn as many tokens as a possible, I'm trying to build good software. If you're paying attention to what you're building, there's so many points where a human is in the loop that it's unusual to run up against token limits.

Anyway, I assume that at some point they have to make enough money to pay the bills. Everything has been subsidized by investors for quite some time, and while the cost per token is going down with efficiency gains in the models/harnesses and with newer compute hardware tuned for these workloads, I think we're all still enjoying subsidized compute at the moment. I don't think Anthropic is making much profit on their plans, especially with folks who somehow run right at the edge of their token limit 24/7. And, I would guess OpenAI is running an even lossier balance sheet (they've raised more money and their prices are lower).

I dunno. I hear a lot of complaining about Claude, but it's been pretty much fine for me throughout 4.5, 4.6 and 4.7. It got Good Enough at 4.5, and it's never been less than Good Enough since. And, when I've tried alternatives, they usually proved to be not quite Good Enough for some reason, sometimes non-technical reasons (I won't use OpenAI, anymore, because I don't trust OpenAI, and Gemini is just not as good at coding as Claude).

show 3 replies
dannypostmayesterday at 10:34 PM

When I saw the German screenshot it all made sense to me.

captainregexyesterday at 7:37 PM

anyone remember the whole “delete uber” thing from 2017ish? good times

bad_haircut72yesterday at 4:42 PM

Waiting 60s every time I send a msg really kills the ux of claude

spaceman_2020yesterday at 9:28 PM

4.7 is the breaking point for me

It's almost unusable

postepowanieadmyesterday at 5:57 PM

Yeah, session limits are kinda show stoppers.

smashahtoday at 6:24 AM

Did the same with Google Ai Ultra. They rug pulled the subscribers. They changed the deal, we cancel. Simple.

zh_codeyesterday at 5:17 PM

I just cancelled my Max20 plan yesterday.

varispeedyesterday at 4:36 PM

It also seems to me they route prompts to cheaper dumber models that present themselves as e.g. Opus 4.7. Perhaps that's what is "adaptive reasoning" aka we'll route your request to something like Qwen saying it's Opus. Sometimes I get a good model, so I found I'll ask a difficult question first and if answer is dumb, I terminate the session and start again and only then go with the real prompt. But there is no guarantee model will be downgraded mid session. I wish they just charged real price and stopped these shenanigans. It wastes so much time.

show 1 reply
r00t-yesterday at 6:04 PM

Same, it's a mess.

danjlyesterday at 5:03 PM

This sounds just like all my neighbors complaining about their internet provider.

tamimiotoday at 1:16 AM

Very similar experience, although I didn’t use claude for anything in production, but I did try some tests with some few topics and questions on things that I know, and while initially it works very well, but as soon as you dive deeper you get all sort of extra none sense that was never asked to add/do nor it’s useful, just workarounds after workarounds after duct tape solutions, several times I would say “no, why are you introducing xyz, that will cause this and that” to get similar answer of “thanks for pushing back, you are right bla bla”.

We probably hit peak generative AI last year, now they probably use AI to improve the AI so it’s kinda garbage in garbage out, or maybe anthropic is deprioritizing users while favoring enterprise or even government where it provides better quality for higher contracts.

gizmodo59yesterday at 5:36 PM

Codex is becoming such a good product. I have the 100$ pro lite. I have Claude still but 20$. I rarely use it. Let’s see if they give generous limits and more importantly a model that’s better than 5.5. The mythos fear mongering did not give me a good impression that they care about the average developer.

johanneskanybaltoday at 12:14 AM

It's not magic but for me definitly claude is the way to go. Not expecting magic it's just another level of non-slop than the rest I've tried.

queuebertyesterday at 5:22 PM

Maybe this is an unpopular opinion, but I think choosing which companies to support during this period of pre-alignment is one way to vote which direction this all goes. I'm happy to accept a slightly worse coding agent if it means I don't get exterminated someday.

drivebyhootingyesterday at 5:03 PM

Imagine vibe coding your core consumer application and associated backend…

Oh wait, I don’t have to imagine. That’s what Anthropic does. A nice preview for what is in store for those who chose to turn off their brains and turn on their AI agents.

kissgyorgyyesterday at 8:47 PM

I cancelled in the minute my subscription stopped working in Pi. Not going back to the slopfest what Claude Code is.

gverrillayesterday at 8:33 PM

My main problem with claude code right now is observability. I've been experimenting a lot with vibe coding, but nowadays I can't even tell what it's doing. It's still delivering me value, but the trust on the company is going down and I've already started looking for alternatives.

josefritzishereyesterday at 8:10 PM

AI has a lot of future potential but at every level... it's still not very good. And certainly not good enough to validate the expense, let alone what the actual cost would be were it profitable.

wslhyesterday at 7:29 PM

Anthropic is astroscaling. We're essentially buying into a loop where speed and iteration take precedence over stability and support. If you view them as an experimental lab undergoing rapid atmospheric friction rather than a company, the "unreliability" is just the cost of being at the frontier. This is not an endorsement for Anthropic, just imagining their craziness on how you "can" grow in a fraction of time.

shevy-javayesterday at 6:42 PM

Those AI using software developers begin to show signs of addiction:

From "yay, claude is awesome" to "damn, it sucks". This is like with withdrawal symptoms now.

My approach is much easier: I'll stay the oldschool way, avoid AI and come up with other solutions. I am definitely slower, but I reason that the quality FOR other humans will be better.

ForOldHackyesterday at 6:39 PM

I have token issues three times a day, and I just upgraded to pro... and now this... now I cancel. my work flow was co-pilot to Gemini to Claude Code... and the bottle neck was always CC. Always. I am done. It should be pretty easy to replace CC.

AI used to be, the punched card replicator... its all replaceable.

moralestapiayesterday at 6:38 PM

The midwit curve of LLMs has OpenAI on both ends.

docheinestagesyesterday at 6:16 PM

Me too.

estimator7292yesterday at 6:06 PM

I just noticed today that it doesn't warn about approaching limits and just blows straight into billing extra tokens.

I'm pretty sure it used to warn when you got close to your 5hr limit, but no, it happily billed extra usage. Granted only about $10 today, but over the span of like 45 minutes. Not super pleased.

GrumpyGoblinyesterday at 4:32 PM

Cool

gexlatoday at 2:26 AM

We can't do it. We standardized. They got us.

semiinfinitelyyesterday at 7:38 PM

absolute garbage support was the reason why I canceled. who would have thought that an AI company has only bots as support agents

🔗 View 14 more comments