Endless AI-Generated Wikipedia
Posted4 months agoActive3 months ago
seangoedecke.comTechstory
calmmixed
Debate
40/100
AI-Generated ContentWikipediaLlms
Key topics
AI-Generated Content
Wikipedia
Llms
The author created an 'Endless Wikipedia' where AI generates new pages based on existing ones, sparking discussions on its potential, limitations, and potential issues like hallucinated content and cost.
Snapshot generated from the HN discussion
Discussion Activity
Active discussionFirst comment
2d
Peak period
12
42-48h
Avg / period
4.8
Comment distribution24 data points
Loading chart...
Based on 24 loaded comments
Key moments
- 01Story posted
Sep 25, 2025 at 5:13 AM EDT
4 months ago
Step 01 - 02First comment
Sep 26, 2025 at 10:22 PM EDT
2d after posting
Step 02 - 03Peak activity
12 comments in 42-48h
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 28, 2025 at 8:54 PM EDT
3 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45370760Type: storyLast synced: 11/20/2025, 2:30:18 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Guess it wasn't so endless after all.
Author is assuming malice, but honestly bots clicking links is just what happens to every public site on the internet. Not to mention going down the link clicking rabbit hole is common among wikipedia readers.
All that said, i don't really see the point. Wikipedia's human controls is what makes it exciting.
As a CS student ~20 years ago I wrote a small website to manage my todo list and hosted it on my desktop in the department. One day I found my items disappearing before my eyes. At first I assumed someone was intentionally messing with my app but logs indicated it was just a scraping bot someone was running.
It was a low stakes lesson on why GET should not mutate meaningful state. I knew when I built it anyone could click the links and I wasn’t bothered with auth since it was one accessible from within the department network. But I didn’t plan for the bots.
You know what doesn't care about Javascript and tries to click every link on your page? A search engine's web crawler.
https://thedailywtf.com/articles/The_Spider_of_Doom
I think the idea is sound, the potential is to have a much larger AI-wikipedia than the human one. Can it cover all known entities, events, concepts and places? All scientific publications? It could get 1000x larger than Wikipedia and be a good pre-training source of text.
Covering a topic I would not make the AI agent try to find the "Truth" but just to analyze the distribution of information out there. What are the opinions, who has them? I would also test a host of models in closed book mode and put an analysis of how AI covers the topic on its own, it is useful information to have.
This method has the potential to create much higher quality text than usual internet scrape, in large quantities. It would be comparative analysis text connecting across many sources, which would be better for the model than training on separate pieces of text. Information needs to circulate to be understood better.
"Tools such as Oracle Designer, Microsoft Visio, and open‑source platforms generate ER diagrams to aid developers in visualizing schema structures and ensuring Sean Goedecke."
I love the idea of "ensuring Sean Goedecke", and that developers are actively working to do so, lol! Something something John Connor something something
[1] https://www.endlesswiki.com/wiki/Entity%E2%80%91relationship...
>edit: I’ve disabled new page generation for now because someone ran a script overnight to endlessly click links and cost me $70.
Edit, well shit looks like there is a Minimalism page, but it didn't make any names clickable. Sean, looks like you need to tweak the code a bit?
https://www.endlesswiki.com/wiki/minimalism