From Lossy to Lossless Reasoning
Posted2 months agoActive2 months ago
manidoraisamy.comTechstory
calmpositive
Debate
0/100
LogicReasoningArtificial Intelligence
Key topics
Logic
Reasoning
Artificial Intelligence
The article discusses the concept of lossless reasoning, exploring its potential applications and implications in fields like artificial intelligence, with the HN community showing interest but limited engagement.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
20m
Peak period
1
0-1h
Avg / period
1
Key moments
- 01Story posted
Oct 31, 2025 at 2:01 PM EDT
2 months ago
Step 01 - 02First comment
Oct 31, 2025 at 2:20 PM EDT
20m after posting
Step 02 - 03Peak activity
1 comments in 0-1h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 31, 2025 at 2:20 PM EDT
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45774840Type: storyLast synced: 11/17/2025, 8:11:06 AM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
OP here. In my previous post [1], I argued that code generation is the kingpin behind reasoning models. The bottleneck is LLMs generate code lossily due to tokenization fragmentation and treating code like natural language instead of structured graphs. In this post I propose:
1. Parsing user prompts into input graphs (using controlled English like ACE)
2. Parsing code into output graphs (AST trees)
3. Using graph transformers to map input graphs → output graphs
Core hypothesis: eliminating tokenization fragmentation ("ladlen" → ["lad", "len"]) and preserving tree structure could improve FrontierMath accuracy from 26% to 35-40%. No benchmarks yet. Just theory and a plan to test the improvement.
I've built compilers, not transformers, so would love technical feedback on:
- Is tokenization & linear structure really the bottleneck in code generation, or am I missing bigger issues?
- Is 35-40% improvement plausible, or overly optimistic?
- For those working on graph transformers: what approaches look promising?
Thanks in advance!
[1] Previous post - https://manidoraisamy.com/reasoning-not-ai.html
HN thread - https://news.ycombinator.com/item?id=45683113