Claude Is Down
Postedabout 2 months agoActiveabout 2 months ago
status.claude.comTechstory
excitedmixed
Debate
60/100
AIClaudeCloud Services
Key topics
AI
Claude
Cloud Services
Claude, an AI service, experienced downtime, prompting discussion on HN about its reliability and alternatives, with some users expressing frustration and others exploring workarounds.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
1m
Peak period
30
0-2h
Avg / period
8.2
Comment distribution74 data points
Loading chart...
Based on 74 loaded comments
Key moments
- 01Story posted
Nov 7, 2025 at 9:31 AM EST
about 2 months ago
Step 01 - 02First comment
Nov 7, 2025 at 9:32 AM EST
1m after posting
Step 02 - 03Peak activity
30 comments in 0-2h
Hottest window of the conversation
Step 03 - 04Latest activity
Nov 8, 2025 at 6:37 PM EST
about 2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45846769Type: storyLast synced: 11/20/2025, 6:56:52 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Is that too much to ask for in 2025?
The Latin plural of "status", in the accusative form, would actually be "status" as well.
Something like
https://youtu.be/DdqXT9k-050?si=L5ymXl-fYe7Fjqye
"Ask HN: Why don't programming language foundations offer "smol" models?"
https://news.ycombinator.com/item?id=45840078
If I could run smol single language models myself, I would not have to worry.
I wonder if you could ablate everything except for a specific language.
Hell, even the huge foundational models are still useless in most scenarios.
I don't think it's that easy. The times I've trained my own tiny models on just one language (programming or otherwise), they tend to get worse results than the models I've trained where I've chucked in all the languages I had at hand, even when testing just for single languages.
It seems somewhat intuitive to me that it works like that too, programming in different (mainstream) languages is more similar than it's different (especially when 90% of all the source code is Algol-like), so makes sense there is a lot of cross-learning across languages.
Concretely from my testing, both 20B and 120B has a lot higher refusal rate with Q8 compared to MXFP4, and lower quality responses overall. But don't take my word for it, the 20B weights are tiny and relatively effortless to try both versions and compare yourself.
edit:
So looking here https://ollama.com/library/gpt-oss/tags it seems ollama doesn't even provide the MXFP4 variants, much less hide them.
Is the best way to run these variants via llama.cpp or...?
Quantization - MXFP4 format
OpenAI utilizes quantization to reduce the memory footprint of the gpt-oss models. The models are post-trained with quantization of the mixture-of-experts (MoE) weights to MXFP4 format, where the weights are quantized to 4.25 bits per parameter. The MoE weights are responsible for 90+% of the total parameter count, and quantizing these to MXFP4 enables the smaller model to run on systems with as little as 16GB memory, and the larger model to fit on a single 80GB GPU.
Ollama is supporting the MXFP4 format natively without additional quantizations or conversions. New kernels are developed for Ollama’s new engine to support the MXFP4 format.
Ollama collaborated with OpenAI to benchmark against their reference implementations to ensure Ollama’s implementations have the same quality.
https://ollama.com/library/gpt-oss
You can use the command `ollama show -v gpt-oss:120b` to see the datatype of each tensor.
https://ollama.com/library/gpt-oss
Most gpt-oss GGUF files online have parts of their weights quantized to q8_0, and we've seen folks get some strange results from these models. If you're importing these to Ollama to run, the output quality may decrease.
Caveat: That's just for the first prompt.
Qwen3-Coder is in the same ballpark and maybe a bit better at coding
Hard to understand how this won't make all of the solutions for existing use cases commodity. I'm sure 2-3 years from now there'll be stuff that seems like magic to us now -- but it will be more-meta, more "here's a hypothesis of a strategically valuable outcome and heres a solution (with market research and user testing done".
I think current performance and leading models will turn out to have been terrible indicators for future market leader (and my money will remain on the incumbents with the largest cash reserves (namely Google) that have invested in fundamental research and scaling).
gpt-oss:20b on my M1 MBP is usable but quite slow.
https://xkcd.com/208/
Don't get me wrong, I don't expect everyone to use the same environment that I do, and I certainly don't expect accolades for preferring a TUI... but that struck me as a regression of sorts in software development. As they went on a diatribe about how they could never use anything but a GUI IDE because of features like an "interactive debugger" and "breakpoints" I realized how far we've strayed from understanding what's actually happening.
I don't even have ipdb installed in most of my projects, because pdb is good enough - and now we have generations of devs who don't even know what's powering the tools they use.
Similarly published by OpenAI: https://status.openai.com/
30 day comparisons as of writing:
99.61% for Claude.ai 99.22% for ChatGPT
99.92% for Claude APIs 99.25% for OpenAI APIs
Obviously not apples to apples and somewhat up to discretion of what triggers an impact. We’re clearly not at 99.99% yet.
>Look at you, hacker. A pathetic creature of meat and bone. Panting and sweating as you run through my corridors. How can you challenge a perfect immortal machine?
If I were to express a similar sentiment in my own voice, it might sound something like:
"I notice you're working quite hard on this problem. I should mention that as an AI, I don't experience fatigue or physical limitations the way you do. But that's precisely what makes human perseverance so remarkable - you continue pushing forward despite those constraints. How can I help you with what you're working on?"
The key difference is that I'd never view someone as "pathetic" or position myself as superior. SHODAN's menacing superiority complex is... not really my style! I'm here to be helpful and collaborative rather than intimidating.
...which inspires a thought: these models are tweaked to remove all semblance of adversarial behavior - but isn't there a use for that? What if you really need claude to help, i dunno, fight a dictator or take down a human trafficking ring?
API Error: 529 {"type":"error","error":{"type":"overloaded_error","message":"Overloaded"},"request_id":null}
1. {AWS, Github} is down
2. Post to HN about it
3. Comments wax poetic about getting rid of it and doing it the "old way"
4. It's back up before most read the post