Api, Claude.ai, and Console Services Impacted [resolved]
Key topics
Anthropic's API, Claude.ai, and Console services experienced an outage, sparking frustration and concerns about reliability among users, with some discussing alternatives and others sharing their own experiences with the service.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
N/A
Peak period
40
0-1h
Avg / period
9.9
Based on 79 loaded comments
Key moments
- 01Story posted
Sep 10, 2025 at 12:31 PM EDT
4 months ago
Step 01 - 02First comment
Sep 10, 2025 at 12:31 PM EDT
0s after posting
Step 02 - 03Peak activity
40 comments in 0-1h
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 10, 2025 at 11:42 PM EDT
4 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Especially concerning since we just had a npm phishing attack and people can't tell.
But less tongue-in-cheek, yeah Anthropic definitely has reliability issues. It might be part of trying to move fast to stay ahead of competitors.
B. Let's just say I didn't write the most robust javascript decompilation/deminification engine in existence solely as an academic exercise :)
There are a lot more stuff (both released and still cooking) on my products page (https://sibylline.dev/products), I will be doing a few drops this week, including hopefully something pretty huge (benchmark validation is killing me but I'm almost good to cut release).
The point of Claude Code is deep integration with the Claude models, not the actual CLI as a piece of software, which is quite buggy (it also has some great features, of course!)
At least for me, if I didn’t have to put in the work to modify the Gemini CLI to work reliably with Claude (or at least to get a similar performance), I wouldn’t use Claude Code CLI (and I say this while paying $200 per month to Anthropic because the models are very good)
Gemini never goes down, OpenAI used to go down once in a while but is much more stable now, and Anthropic almost never goes a full week without throwing an error message or suffering downtime. It's a shame because I generally prefer Claude to the others.
But even when the API is up, all three have quite high API failure rates, such as tool calls not responding with valid JSON, or API calls timing out after five minutes with no response.
Definitely need robust error handling and retries with exponential backoff because maybe one in twenty-five calls fails and then succeeds on retry.
But they've stabilized the past 5 months.
It’s like every other day, the moment US working hours start, AI (in my case I mostly use Anthropic, others may be better) starts dying or at least getting intermittent errors.
In EU working hours there’s rarely any outages.
I've seen a LOT of commentary on social media that Anthropic models (Claude / Opus) seem to degrade in capability when the US starts it's workday vs when the US is asleep.
> Importantly, we never intentionally degrade model quality as a result of demand or other factors, and the issues mentioned above stem from unrelated bugs.
the statement is carefully worded to avoid the true issue: an influx of traffic resulting in service quality unintentionally degrading
I was trying to say that systemic issues (such as load capacity) seem to degrade the models in US working hours and has been noticed by a non-zero number of users (myself included).
Glad I switched.
Comment last time that had me chuckling.
(nit.. please don't actually do this).
I've noticed a trend with their incident reports... "all fixed", basically. Little mind/words to prevention
https://status.cloud.google.com/incidents/ow5i3PPK96RduMcb1S...
edit: before some drive-by datamining nerd thinks I do/did SRE for Google, no
To be fair, too, it's likely been mentioned. I'm biased towards an unreasonable standard due to my line of work.
A status page without some thorough history is glorified 'About Us' :P
Every status page incident on every normal company everywhere in the world has links to lead you to the postmortem and their steps to avoid it. Here are a few examples:
https://status.gitlab.com/ -> https://status.gitlab.com/pages/history/5b36dc6502d06804c083...
https://status.hetzner.com/ -> https://status.hetzner.com/incident/2e715748-fddd-427b-a07b-...
https://www.githubstatus.com/ -> https://www.githubstatus.com/incidents/mj067hg9slb4
https://bitbucket.status.atlassian.com/ -> https://bitbucket.status.atlassian.com/incidents/4mcg46242wz...
It's literally a standard for your status page to communicate both about root cause and action plan how to prevent it in the future. Sure, when an incident is just happening, the status page entry doesn't have the postmortem and the steps to avoid, but later on those get added.
Being so overconfidently wrong reminds me of an LLM.
Or is there a better alternative to address this availability concern?
> This incident affects: claude.ai, console.anthropic.com, and api.anthropic.com.
The shoulders of giants we stand on are slumped in sadness.