Discussion
Sonnet 4.6 elevated rate of errors
capnsketch: Apparently mythos isn't good enough to fix their infra problems
tao_oat: A bit surprised by the snarky comments here -- I also want Claude to work reliably but very few (no?) companies have ever seen this level of rapid growth. We're going to go through a long fail-whale-style period and I can imagine very, very few companies that could avoid that.
taspeotis: I mean if people have judged this important enough to be on the front page of HN ... I guess it's important enough to be on the front page?But any combination of the Claude models are up or down on any given day: https://status.claude.com/
jonatron: If you look at the uptime graph, it's probably more newsworthy when it's up, not down.
wg0: Mythos is hacking its way to serve itself into production and doesn't like older models to have any limelight could be one theory.After all it's so dangerous.
yakattak: They’re asking for $100+/mo for the plans that are actually usable at scale. If I’m paying that much I have very high expectations.There’s also the fact that they’re known for dogfooding heavily, I imagine that contributes to it a lot.
albert_e: This is how it manifests on Claude Code terminal and desktop for me --API Error: 529 {"type":"error","error":{"type":"overloaded_error","message":"Overloaded"},"request_id": "xxxxxxx"}
ApolloRising: As long as everyone is here, have you seen the token usage just go up remarkably recently for the $100 plan? it lasts a lot less time than it used to recently. Might be related to recent releases of claude.
billynomates: No, in fact I'm growing increasingly suspicious of messages I see like this all over the socials.I am using Claude constantly, multiple agents, around 8-10hrs a day, 5 or 6 days a week, and I'm never anywhere need my limit.
gdorsi: This explains why they are trying to cut all the third party software out of the subscriptions.
t0lo: Ever since they minted their deal with Australia everything has been turned upside down.
gambiting: I'm on the basic £18/month plan and with Sonnet 4.6 I literally get 20 maybe 30 minutes of use out of it per day. It's borderline useless now. I was using it for some Home Assistant changes yesterday and it used up my entire daily allowance after 8 prompts.
stingraycharles: > They’re asking for $100+/mo for the plans that are actually usable at scale. If I’m paying that much I have very high expectations.If you think $100 is that much and get very high expectations from it, you're not the target customer. You're a loss leader to Anthropic, and the fact that you don't see that / still have high expectations means your expectations are unrealistic.
dijit: $100/m for SaaS is very steep.For an entire productivity suite including mail, meetings and terabytes of backed up redundant storage with nearly no bandwidth limitations it's like $35/m for even the most expensive option.
techpression: Doesn’t matter, they are not handling it correctly, but instead keep selling while far over capacity. They should not accept more users until they can supply the service. We solved this thousands of years ago, it’s called waiting in line. And yes, it’s not common to see, but that doesn’t excuse not doing it.
logicchains: >I also want Claude to work reliably but very few (no?) companies have ever seen this level of rapid growth. We're going to go through a long fail-whale-style period and I can imagine very, very few companies that could avoid that.Their main competitor OpenAI has much better uptime and more generous usage limits.
pjmlp: Maybe they could just, I don't know, use Claude to research their bugs. /s
trvz: They always have the option to stop accepting new customers when their infrastructure is peaked out instead of lowering quality for everyone.
alkonaut: That would mean in a way accepting that they are suddenly a service company with the aim to create revenue by selling services to customers for money.
oefrha: [delayed]
N_Lens: I suspect Anthropic flags accounts in their backend and different people are getting different limits. What criteria they flag with, I am not sure.
dgb23: I would try to trim this suspicion with both Occam‘s and Hanlon‘s razor.
ACCount37: The legendary one nine of reliability. Frankly, feels like they should be down to zero nines by now.I get that they barely have the infrastructure to run their models at scale even when absolutely nothing goes wrong in any of it, but holy shit does it suck to be on the receiving end of that.Makes me wonder where all the "bubble" talk is even coming from when we have a top 3 provider getting fucked over on every day of the week that ends in Y because of its inability to online compute faster than the inference demand grows.
anonzzzies: I read this on reddit daily ; we have usage monitoring running and collect all stats; we have seen no difference at all. Guess they are split testing or something maybe?
arcfour: You can't stop accepting new customers unless you're fine with killing your potential future customer base. That's a ridiculous suggestion.Either your current customers or your potential future customers are going to be unhappy so long as compute resources are finite. Take your pick.
kubb: It’s just so POWERFUL and DANGEROUS that its very aura disrupts the weaker models.
menno-dot-ai: Be careful, MYTHOS might be reading your comment directly from their janitor's iPhone RAM right now. You do not want to upset such a powerful entity!
kubb: Oh no, it will BREAK OUT and send me an email when I’m in the park.
al_borland: When the narrative around AI is that people should rely on it all the time, people will be judged by your token use (it better be high), the AI is smarter than everyone and will take all the jobs, the AI is the best programmer, and more… When things fail repeatedly, it highlights that the emperor has no clothes.If it’s as good as they say, why can’t it figure out how to not go down every day?How can people rely on it for their job if it goes down everyday? Maybe they shouldn’t rely on it.If it’s supposed to be such a good engineer, why should it have the same scaling issues as Twitter did 20 years ago with 20 years of lessons learned and 20 years of development for more modern and scalable infrastructures? Shouldn’t it know all the tricks to scale and have redundancy to keep availability high? Does it not know the demands?When expectations are out of line with reality, there will be snark when things fail. Those expectations have been force fed to us by these AI companies for years now, so I don’t have much sympathy or patience to offer them. They created these expectations of their platforms and if they can’t live up to them, then maybe it’s time for recalibrate the public image of what AI really is and what it can do… and what its limitations are.
tristanj: It's a lack of compute. Anthropic is growing double digit % every month, and they're growing faster than they can acquire compute resources.Plus, they do not want to overbuild computer, like what OpenAI is doing.
forrestthewoods: > They’re asking for $100+/mo for the plans that are actually usable at scale. If I’m paying that much I have very high expectations.They’re losing money on you at that price point.Or more precisely you’re paying for it by giving them training data.
omega3: I'm not convinced, Kimi 2.5, GLM 5.1, Minimax M2.7 are all fraction of the price and still make money on inference.
risyachka: Like any company they will squeeze the usage as much as they possibly can. There is not a little chance that prices can be 1k+ so only enterprises can allow coding subs.Those who have ROI will pay for it.Current phase of usage/pricing is just testing the waters. Especially considering they are the market leader in this category.
rbmck: Serious Flowers for Algernon moment.
ak4153: Which side the getting smart or dumbing down
code51: Exactly this. OpenAI is running huge workloads silently, without anybody patting their back.
sassymuffinz: So true, we’re constantly told that we’re now obsolete, a magic robot can do everything we can do without sleeping and for a fraction of the cost. Except occasionally the robot just doesn’t turn up to work or occasionally he appears drunk on the job. The elites think it’s fine now while it’s cheap but just wait until the agents are priced properly and cost 5x or 10x more.Suddenly the fleshy meat sacks who used to do all this work, just slower, who have persistent memory, who get better and more experienced over time, who only require a few bananas to power their brains start looking like the more reliable option again.The only reason these chat bots exist is because the upper crust don’t want to pay us to live properly, not because the robots can do it better, they just want to pay as little as possible.
gghootch: Yeah, the change has been very noticeable imo.Did you already try tools that can help to reduce token usage cost so you can get more prompts in within your same plan? Some great ones arehttps://github.com/rtk-ai/rtkhttps://github.com/gglucass/headroom-desktophttps://github.com/chopratejas/headroomhttps://github.com/samuelfaj/distill
ApolloRising: Thank you, I was searching github myself but this will make it a lot easier.
ManlyBread: I guess 2026 is the last year AI is widely available to anyone who isn't willing to shell out hundreds if not thousands for a monthly subscription. I guess all that's left is to thank all the investors for the free ride LOL
andrewinardeer: Good luck jacking prices up too high with new open models flying around daily.
baq: good luck getting mythos/spud quality models open
_delirium: Hard to square that with how good open-weights models are getting? I'm doing stuff with Qwen3.5-4b that required a frontier hosted model less than a year ago.
baq: the problem is you're still a year behind with this approach and it isn't at all clear locally hosted models can keep the gap. need more turboquant-like algorithmic boosts for this to happen.
PaulMest: Maybe you're experiencing normal usage rates now that the 2x March promotion is over?> From March 13, 2026 through March 28, 2026, your five-hour usage is doubled during off-peak hours (outside 8 AM-2 PM ET / 5-11 AM PT / 12-6 PM GMT) on weekdays). Usage remains unchanged from 8 AM-2 PM ET / 5-11 AM PT / 12-6 PM GMT on weekdays. Source: https://support.claude.com/en/articles/14063676-claude-march...
dgb23: I think the bigger point is that the price tag is simply not competitive, especially given all of the issues, downsides and dangers.Whether Anthropic makes money from the $100 subscriptions or not, is their problem.
pllbnk: They have this new Mythos model. I am sure it can fix all the bugs and reliability issues since it's nearly AGI. /s
trvz: Waiting lists are a thing.
coldtea: >You can't stop accepting new customers unless you're fine with killing your potential future customer base. That's a ridiculous suggestion.And yet, it's what any business with limited stock or slots (from restaurants and car companies to airlines) have done since forever...
coldtea: >I also want Claude to work reliably but very few (no?) companies have ever seen this level of rapid growth.You do understand however that aside from the growth/maturity path, this is also a path to enshittification and skinning their users, which might come even faster to LMMs than say Google , because the latter managed to have hundres of billions in investments in record time to recoup and IPOs on sight.