Processing Strings 109x Faster Than Nvidia on H100
Posted4 months agoActive4 months ago
ashvardanian.comTechstory
calmmixed
Debate
40/100
GPU OptimizationString ProcessingCuda
Key topics
GPU Optimization
String Processing
Cuda
The post claims to achieve 109x faster string processing than Nvidia on H100, sparking discussion on the accuracy and scope of this claim, as well as the techniques used to achieve it.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
3d
Peak period
2
78-84h
Avg / period
1.5
Key moments
- 01Story posted
Sep 20, 2025 at 11:30 PM EDT
4 months ago
Step 01 - 02First comment
Sep 24, 2025 at 3:15 AM EDT
3d after posting
Step 02 - 03Peak activity
2 comments in 78-84h
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 24, 2025 at 10:13 AM EDT
4 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Discussion (3 comments)
Showing 3 comments
trilogic
4 months ago
1 replyImpressive work anti diagonal DP on CUDA, clean MCUPS framing, and the multi language shipping is legit. The “109× faster than NVIDIA on H100” line is accurate for your chosen case (cuDF/nvtext, long strings), but it’s not a blanket “faster than NVIDIA,” and readers will assume that tighten the scope. Bio results are a good baseline, not SOTA; Hopper’s DPX and WFA style tiling/bucketing would likely move you a tier up. Hashing and 52 bit MinHash are clever, but you need full SMHasher reports and retrieval quality metrics, not just entropy/collisions. Publish exact versions, params, and end to end timings (I/O + marshaling), plus short string vs long string batches. If you add those and rename the headline to reflect the setup, the claims will be hard to poke holes in.
klysm
4 months ago
AI slop meter is off the charts
ozgrakkurt
4 months ago
Duplicate of https://news.ycombinator.com/item?id=45304807
View full discussion on Hacker News
ID: 45319791Type: storyLast synced: 11/20/2025, 5:33:13 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.