A Year of Fast Apply – Our Path to 10k Tokens Per Second
Posted2 months agoActive2 months ago
relace.aiTechstory
calmpositive
Debate
20/100
Performance OptimizationMachine LearningAI Infrastructure
Key topics
Performance Optimization
Machine Learning
AI Infrastructure
The post discusses Relace's achievement of reaching 10k tokens per second with their 'Fast Apply' technology, with commenters exploring the technical details and potential applications.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
3h
Peak period
2
6-7h
Avg / period
1.2
Key moments
- 01Story posted
Oct 29, 2025 at 1:04 PM EDT
2 months ago
Step 01 - 02First comment
Oct 29, 2025 at 3:44 PM EDT
3h after posting
Step 02 - 03Peak activity
2 comments in 6-7h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 30, 2025 at 12:33 AM EDT
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45749763Type: storyLast synced: 11/20/2025, 1:26:54 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
vibecoding internal eval tools is the single best use case of ai accelerating ai i know of! nice to see
(sorry if this gets asked a lot) - any philsophical/methodology differences to MorphLLM that you'd call out since you seem to be a direct alternative?
It's hard to know for sure because their methods aren't public, but my guess is the dataset they constructed pushes the Fast Apply model to more aggressively fix mistakes introduced by the frontier model in the edit snippet.
This aligns with the fact that their flagship model (morph-v3-large) is 4x slower than ours -- the smoothings/hallucinations are not in the initial code or the edit snippet so they break speculative continuations more frequently. Their 2x faster model (morph-v3-fast) is likely quantized more aggressively (maybe fp4? and run on B200s?) because it exhibits very strange behaviors like hallucinating invalid characters at random points that make the code non-compilable.
From an accuracy POV, auto-smoothing is helpful for fixing obvious mistakes in the edit snippet like missed imports from well known packages. However, it does increase the frequency of code breaking hallucinations like invalid local imports among other functional changes that you might not want a small apply model to perform.