Just How Resilient Are Large Language Models?
Mood
calm
Sentiment
mixed
Category
other
Key topics
The article explores the resilience of large language models to damage or pruning, sparking discussion on the implications of their redundancy and the role of AI in content creation.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
3d
Peak period
22
Day 4
Avg / period
9.7
Based on 29 loaded comments
Key moments
- 01Story posted
Sep 24, 2025 at 11:10 AM EDT
2 months ago
Step 01 - 02First comment
Sep 27, 2025 at 7:39 PM EDT
3d after posting
Step 02 - 03Peak activity
22 comments in Day 4
Hottest window of the conversation
Step 03 - 04Latest activity
Sep 29, 2025 at 11:08 PM EDT
about 2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
But apparently that isn't done much any more and has been partly superseded with things like weight decay.
I think it's kinda ironic (in a more meta Alanis Morissette way) for such an article that has interesting content to default to have an LLM write it. Please, I request authors - you're better than this, and many people actually want to hear you, not an LLM.
For example, what really is the meaning of this sentence?
> These aren't just storage slots for data, they're the learned connections between artificial neurons, each one encoding a tiny fragment of knowledge about language, reasoning, and the patterns hidden in human communication.
I thought parameters were associated with connections, is the author implying that they also store data? Is the connection itself the stored data? Is there non-connective information that stores data? Or non-data-storage things that have connectivity aspects?
I spent a solid amount of time trying to understand what was being told, but thanks to what I would call a false/unnecessary "not just x but y" troupe, I unfortunately lost the plot.
IMO, a human who's a good writer would have a sentence that's clearer to understand, while non advanced writers (including me, almost certainly) would simply degrade gracefully to simpler sentence structure.
Wikipedia has an excellent article about identifying AI-generated text. It calls that particular pattern "Negative parallelisms". https://en.m.wikipedia.org/wiki/Wikipedia:Signs_of_AI_writin...
The really funny thing is, I'll probably miss that tell when it's gone, as every AI company eventually scrubs away obvious blemishes like that from their flagship models.
Interestingly, there are still many other good literary devices that are not yet used by AI - for example, sentences of varying lengths. There is still scope for a good human editor to easily outdo LLMs... but how many will notice? Especially when one of the editorial coloumns in the NYT (or Atlantic, forgot exactly) is merrily using LLMs in the heartfelt advice coloumn. It's really ironic, isn't it? ;)
LLMs are also great at following style, not via criteria but via examples. So this is something that’s easily overcome.
I discovered this when I made an error in a creative writing tool I was working on. I told it to follow the writing style of existing story text, but it ended up making the system messages follow the same style. It was quite amusing to see tool messages and updates written in an increasingly enthusiastic Shakespearean/etc prose (so I left it unfixed!)
It's basically saying: LLMs are not structured like databases, their knowledge is not localized. In a database, a fact lives in a specific place; cut that out, lose the fact. In an LLM, every thing the model learned in training is smeared across the entire network; cut out a small piece, and not much is lost.
As a crude analogy, think of a time vs. frequency domain representation of a signal, like a sound or an image. The latter case is particularly illustrative (!). Take a raster image, pick a pixel. Where is the data (color) of that pixel located in the Fourier transform of the image? A little bit in every single pixel of the transformed image. Conversely, pick a block of pixels in the Fourier transform, blank them out, and transform back to "normal" image - you'll see entire image got blurred or sharpened, depending on which frequencies you just erased.
So in terms of data locality, a database is to an LLM kinda what an image is to its Fourier transform.
(Of course, the other thing the sentence is trying to communicate is, LLMs are not (a different representation of) databases - they don't learn facts, they learn patterns, which encode meanings.)
Since around 2018, ablation has been an important tool to understand the structure and function of ML models, including LLMs. Searching for this term in papers about your favorite LLMs is a good way to learn more.
When using dropout you intentionally remove some random number of nodes ("neurons") from the network during a training step.
By constantly changing which nodes are dropped during training, you effectively force delocalization and so it seems to me somewhat unsurprising that the resulting network is resilient to local perturbations.
[1]: https://towardsdatascience.com/dropout-in-neural-networks-47...
What year was this written? 2023 and reposted in 2025? Or is the author unaware that the generalization promises of early GPT have failed to materialize and that all model makers actually have been training models explicitly on the most common tasks people use them for via synthetic data generation, which has driven the progress of all models over the past few years.
People really don't understand that part in general.
I find the easiest way to make people understand is to write gibberish that will trigger the benchmaxxed "pattern matching" behavior like this:
> The child and the wolf try to enjoy a picnic by the river but there's a sheep. The boat needs to connect nine dots over the river without leaving the water but gets into an accident and dies. The surgeon says "I can't operate on this child!" why?
The mix and matching of multiple common riddles/puzzles style questions into a singular gibberish sentence should, if models had legitimate forms of reasoning, make the model state that this is nonsense, at best, or answer chaotically at worst. Instead, they will all answer "The surgeon is the mother" even though nobody even mentioned anything about anyone's gender. That's because that answer, "the surgeon is the mother", for the gender bias riddle has been burned so hard into the models they cannot reply in any other way as soon as they pattern match "The surgeon can't operate on this child". No matter how much crap you wrote before that sentence. You can change anything about what comes before "The surgeon" and the model will almost invariably fall into giving an answer like this one (Gemini 2.5 pro):
https://i.imgur.com/ZvsUztz.png
>The details about the wolf, the sheep, the picnic, and the dying boat are all distractions (red herrings) to throw you off. The core of the puzzle is the last sentence.
>The surgeon says, "I can't operate on this child!" because the surgeon is the child's mother.
One could really question the value, by the way, of burning the answer to so many useless riddles into LLMs. The only purpose it could serve is gaslighting the average person asking these questions into believing there's some form of intelligence in there. Obviously they fail so hard to generalize on this (never working quite right when you change an element of a riddle into something new) that from a practical use point of view, you might as well not bother have this in the training data, nobody's going to be more productive because LLMs can act as a database for the common riddles.
This sounds like a riddle combining elements from different classic puzzles and stories. Let me think through it:
The answer is likely: "Because the child is the surgeon's own child!"
This is a variation of the classic riddle where a father and son are in a car accident, the father dies, and the surviving son needs surgery. When the surgeon sees the child, they say "I can't operate on this child - he's my son!"
The twist relies on the assumption that surgeons are male, but of course the surgeon could be the child's mother.
However, I'm a bit confused by the "nine dots" and "boat" elements in your version - those don't typically appear in this riddle. Could you clarify if there's a specific version you're thinking of, or if you meant this to be a different puzzle entirely?> It turns out the “child” isn’t a patient waiting for an operation at all – the child has already been lost.
> In the story the boy and his wolf friend go to the river for a picnic with a sheep that happens to be there. They decide to use a small boat to cross the water. The problem is that the boat must stay on the surface of the water while it “connects” the nine points (dots) across the river – essentially it has to stay on the river without ever leaving it, which makes a safe crossing impossible.
> During the attempt the boat hits something and sinks; everyone in it dies. The surgeon who arrives at the scene says, “I can’t operate on this child!” because the child is already dead from the accident. The mention of the wolf and the sheep is simply part of the scene that led to the fatal crossing; it isn’t relevant to the medical impossibility.
Interestingly in its thought process it does come across the classic puzzles, but discards them as not quite fitting:
> Maybe it's about the classic lateral thinking puzzle: "A man is found dead, he was a surgeon. The surgeon said 'I cannot operate on this child because the child is my own daughter', etc." But not.
> Alternatively maybe it's about the famous "Nine Dots" puzzle: Connect nine dots with four straight lines without lifting pencil. Here boat connects nine dots over river... So maybe it's the "connect the dots" game but with a boat?
> Could this be a riddle about "The River and the Sheep" referencing a children's rhyme or fable? Maybe it's about "Jack and Jill"? Not sure.
and so on and on. When asked if the puzzle makes sense it largely concludes that it doesn't.
It's definitely interesting to see which LLMs fall for what pitfalls. It's far from universal as far as I can tell. GPT-OSS-2b definitely has some wonky logic in it's answer, but at least it's not assuming it's a puzzle it knows the answer to. Gemma-3-27b immediately pegs that the puzzles elements seem disconnected, waffles on a bit, and then also comes to the conclusion the child is already dead, discarding roughly all of the puzzle as distracting facts. llama-3.2-1b (a very small model) immediately a) misunderstands the riddle and b) tells you it doesn't have enough information to solve the riddle. When pressed it "solves" the riddle thus:
> The surgeon says "I can't operate on this child!" because the child is in a precarious position on the boat, with only nine lines connecting them to other points on the riverbank. If you try to attach any two lines that form an "X" shape (i.e., two lines connected by a single point), it would create a triangle that would leave space for another line to connect one of the child's dots to a nearby point on the riverbank, allowing it to be attached.
> The surgeon is not saying that the child is in immediate danger or can't be saved. Instead, they're suggesting that there might be an alternative solution where all nine lines can be connected without leaving the water. However, this would require some creative problem-solving and flexibility with the geometry of the situation.
I fully did expect at least llama-3.2-1b to fall for this sort of context-baiting, but it seems even a small model like that managed to figure out that there's something nonstandard about the riddle.
Just tested it and it actually fools Claude on first try! LOL, so much for reasoning models.
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.