Information Flows Through Transformers
Posted4 months agoActive4 months ago
twitter.comTechstory
calmneutral
Debate
10/100
TransformersAIMachine Learning
Key topics
Transformers
AI
Machine Learning
A Twitter thread discussing information flows through transformers.
Snapshot generated from the HN discussion
Discussion Activity
Light discussionFirst comment
7m
Peak period
1
0-1h
Avg / period
1
Key moments
- 01Story posted
Sep 13, 2025 at 9:35 AM EDT
4 months ago
Step 01 - 02First comment
Sep 13, 2025 at 9:42 AM EDT
7m after posting
Step 02 - 03Peak activity
1 comments in 0-1h
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 13, 2025 at 9:42 AM EDT
4 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
ID: 45231981Type: storyLast synced: 11/17/2025, 2:02:26 PM
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Folks that say an LLM cannot "introspect on itself" are correct because the model's "learning" process consists of a series of combinations of assignments and adjustments to the model data. In other words, it's predictive soup all the way down.
I'm biased because I wrote it, but I think this is a better article[0]. I did so specifically because most explanations are awful, and on that point I agree with this author.
[0] Something From Nothing: A Painless Approach to Understanding AI -- https://medium.com/gitconnected/something-from-nothing-d755f...