Recursive Language Models (rlms)
Posted3 months agoActive3 months ago
alexzhang13.github.ioTechstory
skepticalmixed
Debate
80/100
Artificial IntelligenceLanguage ModelsRecursive Models
Key topics
Artificial Intelligence
Language Models
Recursive Models
The Recursive Language Models (RLMs) paper proposes a new approach to language modeling by wrapping an existing LM with a dynamic environment, sparking debate among commenters about its novelty and effectiveness.
Snapshot generated from the HN discussion
Discussion Activity
Active discussionFirst comment
2h
Peak period
16
0-6h
Avg / period
5.8
Comment distribution35 data points
Loading chart...
Based on 35 loaded comments
Key moments
- 01Story posted
Oct 15, 2025 at 1:43 PM EDT
3 months ago
Step 01 - 02First comment
Oct 15, 2025 at 3:57 PM EDT
2h after posting
Step 02 - 03Peak activity
16 comments in 0-6h
Hottest window of the conversation
Step 03 - 04Latest activity
Oct 19, 2025 at 2:34 PM EDT
3 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45596059Type: storyLast synced: 11/20/2025, 4:35:27 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
See e.g. https://textbooks.cs.ksu.edu/cc210/16-recursion/08-recursion...
As another example, a finite-state-machine language can have loops, but it can’t recurse unless there is external memory it has access to in a way that it can serve as a stack. Regular expressions also fall into that pattern; they can loop, but they can’t recurse. For that you need a pushdown automaton: https://en.wikipedia.org/wiki/Pushdown_automaton.
The authors use as an environment a Python REPL that itself can call other instances of the LM. The prompt is programmatically manipulated as a Python variable on the REPL.
The motivation is for the LM to use Python commands, including commands that call other LM instances, to figure out how best to modify the context at inference time.
The results from early testing look impressive at a first glance: An RLM wrapping GPT-5-mini outperforms GPT-5 by a wide margin on long-context tasks, at significant lower cost.
I've added this to my reading list.
EDIT: makes me think of many computation systems in various substrates, and how they work. Focus vs distraction/creativity. ADHD workers in hierarchies of capitalism, purpose of breadth vs depth of exploration at various levels of the stack, who's at the "top" and why, etc etc
It was even before the rise of LLMs
The authors may want to consider a more specific name
It’s not relying on the LM context much. You can generally code away for an hour before you run out of context and have to run a compression step or just start fresh.
please correct me if I'm wrong..this is just subagent architecture?
> Lastly, in our experiments we only consider a recursive depth of 1 — i.e. the root LM can only call LMs, not other RLMs.
> but we felt that for most modern “long context” benchmarks, a recursive depth of 1 was sufficient to handle most problems.
I don't think a size 2 call stack algorithm should be regarded as 'recursive'.
It feels a little disingenuous to call it a Recursive Language Model when the recursive depth of the study was only 1.
1.Recursion is used to break down the large context and dispatch to different LLM calls to get the useful context.
2.This may lead to longer test-time execution on large contexts (even with parallelism in deep recursion), and the monetary cost may increase rapidly.
I think it’s a different idea from using RAG or manually maintaining a context window
correct me if I'm wrong
[1] https://viper.cs.columbia.edu/static/viper_paper.pdf
It simply hopes two drunks are more coherent than one drunk.