Why Isn't Everyone Using Cerebras?
Postedabout 2 months agoActiveabout 2 months ago
Techstory
calmmixed
Debate
20/100
CerebrasLarge Language ModelsAI Infrastructure
Key topics
Cerebras
Large Language Models
AI Infrastructure
I work at a mid-sized startup dealing with latency issues in customer-facing flows that use LLMs. Using OSS-120B seems preferable to 5-mini or Anthropic models in many cases when we need speed, intelligence, and cost control. Is there some catch here beyond needing to acquire higher rate limits?
The author is considering using Cerebras for their LLM needs but is unsure about the potential drawbacks, with the discussion revolving around the reliability and support of Cerebras' models.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
1h
Peak period
1
1-2h
Avg / period
1
Key moments
- 01Story posted
Nov 14, 2025 at 6:10 PM EST
about 2 months ago
Step 01 - 02First comment
Nov 14, 2025 at 7:39 PM EST
1h after posting
Step 02 - 03Peak activity
1 comments in 1-2h
Hottest window of the conversation
Step 03 - 04Latest activity
Nov 14, 2025 at 7:39 PM EST
about 2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45933257Type: storyLast synced: 11/17/2025, 6:05:26 AM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
I still don't know how long they'll support our chosen model.
On Oct 22 I got an email saying that
```
- qwen-3-coder-480b will be available until Nov 5, 2025
- qwen-3-235b-a22b-thinking-2507 will be available until Nov 14, 2025
```
That's not a lot of notice!
I don't want to spend all my time benchmarking new models for features I already built. I don't want my users' experience to be disturbed every few months.