Gen-0 / Embodied Foundation Models That Scale with Physical Interaction
Posted2 months agoActiveabout 2 months ago
generalistai.comTechstory
excitedpositive
Debate
20/100
AIRoboticsMachine Learning
Key topics
AI
Robotics
Machine Learning
The GEN-0 project demonstrates embodied foundation models that scale with physical interaction, sparking discussion on the implications and potential applications of autonomous robotics.
Snapshot generated from the HN discussion
Discussion Activity
Moderate engagementFirst comment
10d
Peak period
6
Day 10
Avg / period
6
Key moments
- 01Story posted
Nov 5, 2025 at 5:25 AM EST
2 months ago
Step 01 - 02First comment
Nov 14, 2025 at 9:48 PM EST
10d after posting
Step 02 - 03Peak activity
6 comments in Day 10
Hottest window of the conversation
Step 03 - 04Latest activity
Nov 15, 2025 at 5:07 AM EST
about 2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45821402Type: storyLast synced: 11/20/2025, 12:32:34 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Imo the latter will be very useful for semantic planning and reasoning, but only after manipulation is solved.
A ballpark cost estimate -
- $10 to $20 hourly wages for the data collectors
- $100,000 to $200,000 per day for 10,000 hours of data
- ~1,500 to 2,500 data collectors doing 4 to 6 hours daily
- $750K to $1.25M on hardware costs at $500 per gripper
Fully loaded cost between $4M to $8M for 270,000 hours of data.
Not bad considering the alternatives.
For example, teleoperation is way less efficient - it's 5x-6x slower than human demos, and 2x-3x more expensive per hour of operator time. But could become feasible after low-level and mid-level manipulation and task planning is solved.
Thinking about it, I'm reminded of various "additive training" tricks. Teach an AI to do A, and then to do B, and it might just generalize that to doing A+B with no extra training. Works often enough on things like LLMs.
In this case, we use non-robot data to teach an AI how to do diverse tasks, and robot-specific data (real or sim) to teach an AI how to operate a robot body. Which might generalize well enough to "doing diverse tasks through a robot body".
3 more comments available on Hacker News