The Case That A.i. Is Thinking
Key topics
The article 'The Case That A.I. Is Thinking' sparks a heated debate on whether current AI systems, particularly LLMs, are truly 'thinking' or just sophisticated statistical models, with commenters discussing the nuances of intelligence, consciousness, and the implications of AI advancements.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
2m
Peak period
109
0-6h
Avg / period
20
Based on 160 loaded comments
Key moments
- 01Story posted
Nov 3, 2025 at 12:55 PM EST
2 months ago
Step 01 - 02First comment
Nov 3, 2025 at 12:56 PM EST
2m after posting
Step 02 - 03Peak activity
109 comments in 0-6h
Hottest window of the conversation
Step 03 - 04Latest activity
Nov 7, 2025 at 12:02 PM EST
2 months ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
(Apologies if that's been edited in after your comment)
https://news.ycombinator.com/user?id=jsomers
But it is not sentient. It has no idea of a self or anything like that. If it makes people believe that it does, it is because we have written so much lore about it in the training data.
It's not a living thing inside the computer, it's just the inference building text token by token using probabilities based on the pre-computed model.
'Power over your computer', all that is orthogonal to the point. A human brain without a functioning body would still be thinking.
I permit you, that yes, the data in the model is a LOT more cool, but some team could by hand, given billions of years (well probably at least 1 Octillion years), reproduce that model and save it to a disk. Again, no different than data stored in JIRA at that point.
So basically if you have that stance you'd have to agree that when we FIRST invented computers, we created intelligence that is "thinking".
Obviously, it is different or else we would just use JIRA and a database to replace GPT. Models very obviously do NOT store training data in the weights in the way you are imagining.
>So basically if you have that stance you'd have to agree that when we FIRST invented computers, we created intelligence that is "thinking".
Thinking is by all appearances substrate independent. The moment we created computers, we created another substrate that could, in the future think.
if the user types "hi" respond with "hi" or "bye" or "..." you get the point. It's basically storing the most probably following words (tokens) given the current point and its history.
That's not a brain and it's not thinking. It's similar to JIRA because it's stored information and there are if statements (admins can do this, users can do that).
Yes it is more complex, but it's nowhere near the complexity of the human or bird brain that does not use clocks, does not have "turing machines inside", or any of the other complete junk other people posted in this thread.
The information in Jira is just less complex, but it's in the same vein of the data in an LLM, just 10^100 times more complex. Just because something is complex does not mean it thinks.
Yes it is more complex, but it's nowhere near the complexity of the human or bird brain that does not use clocks, does not have "turing machines inside", or any of the other complete junk other people posted in this thread.
The information in Jira is just less complex, but it's in the same vein of the data in an LLM, just 10^100 times more complex. Just because something is complex does not mean it thinks.
So, what is the missing element that would satisfy you? It's "nowhere near the complexity of the human or bird brain", so I guess it needs to be more complex, but at the same time "just because something is complex does not mean it thinks".
Does it need to be struck by lightning or something so it gets infused with the living essence?
If you want the best comparison, it's closer to a plant- it reacts ONLY to external stimulous, sunlight, water, etc... but it does not think. (And I'm not comparing it to a plant so you can say - SEE you said it's alive!) It's just a comparison.
As far as sentience is concerned, we can't say they aren't sentient because we don't know the computational structures these models realize, nor do we know the computational structures required for sentience.
The inference software that would normally read from that file is also not alive, as it's literally very concise code that we wrote to traverse through that file.
So if the disk isn't alive, the file on it isn't alive, the inference software is not alive - then what are you saying is alive and thinking?
“So if the severed head isn’t alive, the disembodied heart isn’t alive, the jar of blood we drained out isn’t alive - then what are you saying is alive and thinking?”
- Some silicon alien life forms somewhere debating whether the human life form they just disassembled could ever be alive and thinking
Sure, and humans are just biochemical reactions moving muscles as their interface with the physical word.
I think the model of operation is not a good criticism, but please see my reply to the root comment in this thread where I detail my thoughts a bit.
1. We trained it on a fraction of the world's information (e.g. text and media that is explicitly online)
2. It carries all of the biases us humans have and worse the biases that are present in the information we chose to explicitly share online (which may or may not be different to the experiences humans have in every day life)
This is going to be a huge problem. Most people assume computers are unbiased and rational, and increasing use of AI will lead to more and larger decisions being made by AI.
All seem biased to recent buzzwords and approaches. Discussions will include the same hand-waving of DDD, event-sourcing and hexagonal services, i.e. the current fashion. Nothing of worth apparently preceded them.
I fear that we are condemned to a future where there is no new novel progress, but just a regurgitation of those current fashion and biases.
Am I wrong about this?
The only basis I have for assuming you are sentient according to that definition is trust in your self-reports.
The "feelings" part is probably harder though.
In practical terms it does not matter - it is reasonable for us to act as if others do experience the same we do. But if we are to talk about the nature of conscience and sentience it does matter that the only basis we have for knowing about other sentient beings is their self-reported experience.
How do you know that model processing text or image input doesn't go through feeling of confusion or excitement or corrupted image doesn't "smell" right for it?
Just the fact that you can pause and restart it doesn't mean it doesn't emerge.
Because the other person is part of your same species so you project your own base capabilities onto them, because so far they shown to behave pretty similarly to how you behave. Which is the most reasonable thing to do.
Now, the day we have cyborgs that mimic also the bodies of a human a la Battlestar Galactica, we will have an interesting problem.
Because I sincerely do not. I have frankly no idea how sentience arises from non sentience. But it's a topic that really interests me.
But to me the big mystery is how animals have sensation at all to begin with. What gives rise to that is a greater mystery to me personally.
There are examples of people who have no ability to feel pain yet are still able to think. Now I wonder if they ever experience mental anguish.
Clearly most animals are less educated, and most are less intelligent, but non-sentient? That sounds like 200-year old claims that "when one steps on the tail of a cat, it does indeed protest loudly, but not because it feels anything or because it would be sentient, no, no, it protests merely due to selective pressure, programming reflex circuits, since other creatures would show compassion, or back off due to a potential reaction by the cat."
Anyone who has had a pet like a cat or a dog knows they are sentient... if we consider ourselves sentient.
But asked for reasons I can only point to the social nature of their societies, where love and anger make sense, or of their hurt-behavior.
I also find it very hard to believe that everything else is slow evolution of components, and here all of a sudden something super complex comes into being out of nowhere.
But I still have no idea how it could work. What are the components and their interplay?
So any other Turing-complete model can emulate it, including a computer. We can even randomly generate Turing machines, as they are just data. Now imagine we are extremely lucky and happen to end up with a super-intelligent program which through the mediums it can communicate (it could be simply text-based but a 2D video with audio is no different for my perspective) can't be differentiated from a human being.
Would you consider it sentient?
Now replace the random generation with, say, a back propagation algorithm. If it's sufficiently large, don't you think it's indifferent from the former case - that is, novel qualities could emerge?
With that said, I don't think that current LLMs are anywhere close to this category, but I just don't think this your reasoning is sound.
Absolutely.
If you simulated a human brain by the atom, would you think the resulting construct would NOT be? What would be missing?
I think consciousness is simply an emergent property of our nervous system, but in order to express itself "language" is obviously needed and thus requires lots of complexity (more than what we typically see in animals or computer systems until recently).
That is what we don't know is possible. You don't even know what physics or particles are as yet undiscovered. And from what we even know currently, atoms are too coarse to form the basis of such "cloning"
And, my viewpoint is that, even if this were possible, just because you simulated a brain atom by atom, does not mean you have a consciousness. If it is the arrangement of matter that gives rise to consciousness, then would that new consciousness be the same person or not?
If you have a basis for answering that question, let's hear it.
But later on, he would get different experiences and become a different person no longer identical to the first.
In extension, I would argue that magically "translating" a person to another medium (e.g. a chip) would still make for the same person, initially.
Though the word "magic" does a lot of work here.
Either it is (and continues to be) the same consciousness, or it is not. If it were the same consciousness, then you would have a person who exists in two places at once.
So your question is, are two whirlwinds with identical properties (same speed, same direction, shape etc) the same in one box of air, vs another identical box?
To me, two such whirlwinds are identical but not the same. They are the same only if they are guaranteed to have the same value for every conceivable property, forever, and even this condition may not be enough.
Qualia exist "outside" spacetime, e.g. redness doesn't have a position in spacetime. If consciousness is purely physical, then how can two identical systems (identical brains with identical sensory input) giving rise by definition to the same qualia not literally be the same consciousness?
I'm sensing redness here and now, so the sensation of redness exists very clearly tied to a particular point in spacetime. In what sense is the qualia of redness not firmly anchored in spacetime? Of course, you could talk about the concept redness, like the concept Pi, but even then, these concepts exist in the mind of a human thinking about them, still tied to a particular location in spacetime.
> If consciousness is purely physical, then how can two identical systems (identical brains with identical sensory input) giving rise by definition to the same qualia not literally be the same consciousness?
The two brains don't receive the same sensory inputs, nothing in the experiment says they do. From the second right after the duplicate is created, their sensory inputs diverge, and so they become separate consciousnesses with the same history. They are interchangeable initially, if you gave the same sensory inputs to either of them, they would have the same output (even internally). But, they are not identical: giving some sensory input to one of them will not create any effect directly in the other one.
But qualia are inherently subjective. You can correlate brain activity (which exists at a position in spacetime) to subjective experience, but that experience is not related to spacetime.
Said otherwise: imagine you are in the Matrix at a coffee shop and sense redness, but your brain is actually in a vat somewhere being fed fake sensory input. "Where" is the redness? You would clearly say that it arises in your brain in the coffee shop. Imagine then the vat is moved (so its position in spacetime changes), your brain is rolled back to its previous state, and then fed the same sensory input again. Where is the redness now?
You can't differentiate the two sensations of redness based on the actual position of the brain in spacetime. For all intents and purposes, they are the same. Qualia only depend on the internal brain state at a point in time and on the sensory input. Spacetime is nowhere to be found in that equation.
> The two brains don't receive the same sensory inputs
But let's say they do. Identical brains, identical inputs = identical qualia. What differentiates both consciousnesses?
You would not need the simulation to be perfect; there is ample evidence that our brains a quite robust against disturbances.
> just because you simulated a brain atom by atom, does not mean you have a consciousness.
If you don't want that to be true, you need some kind of magic, that makes the simulation behave differently from reality.
How would a simulation of your brain react to an question that you would answer "consciously"? If it gives the same responds to the same inputs, how could you argue it isnt't conscious?
> If it is the arrangement of matter that gives rise to consciousness, then would that new consciousness be the same person or not?
The simulated consciousness would be a different one from the original; both could exist at the same time and would be expected to diverge. But their reactions/internal state/thoughts could be matched at least for an instant, and be very similar for potentially much longer.
I think this is just Occams razor applied to our minds: There is no evidence whatsoever that our thinking is linked to anything outside of our brains, or outside the realm of physics.
does not mean that the essential thing gives rise to consciousness is only approximate. To give an example from software, you can write software is robust against bad input, attempts to crash it, even bit flips. But, if I came in and just changed a single character in the source code, that may cause it to fail compilation, fail to run, or become quite buggy.
> If you don't want that to be true, you need some kind of magic,
This is just what I'm saying is a false dichotomy. The only reason some are unable to see beyond it is that we think the basic logic we understand are all there could be.
In this respect physics has been very helpful, because without peering into reality, we would have kept deluding ourselves that pure reason was enough to understand the world.
It's like trying to explain quantum mechanics to a well educated person or scientist from the 16th century without the benefit of experimental evidence. No way they'd believe you. In fact, they'd accuse you of violating basic logic.
We have zero evidence for either.
> does not mean that the essential thing gives rise to consciousness is only approximate
But we have 8 billion different instances that are presumably conscious; plenty of them have all kinds of defects, and the whole architecture has been derived by a completely mechanical process free of any understanding (=> evolution/selection).
On the other hand, there is zero evidence of consciousness continuing/running before or after our physical brains are operational.
Defects that have not rendered them unconscious, as long as they still are alive. You seem not to see the circularity of your argument.
I gave you an example to show that robustness against adverse conditions is NOT the same as internal resiliency. Those defect, as far as we know, are not affecting the origin of consciousness itself. Which is my point.
> How is it a false dichotomy? If you want consciousness to NOT be simulateable, then you need some essential component to our minds that can't be simulated (call it soul or whatever) and for that thing to interface with our physical bodies (obviously).
If you need two things to happen at the same time in sync with each other no matter if they are separated by billions of miles, then you need faster-than-light travel, or some magic [1]; see what I did there?
1. I.e., quantum entanglement
This makes no sense as written - by definition, there is no concept of "at the same time" for events that are spacelike separated like this. Quantum entanglement allows you to know something about the statistical outcomes of experiments that are carried over a long distance away from you, but that's about it (there's a simpler version, where you can know some facts for certain, but that one actually looks just like classical correlation, so it's not that interesting on its own).
I do get the point that we don't know what we don't know, so that a radical new form of physics, as alien to current physics as quantum entanglement is to classical physics, could exist. But this is an anti-scientific position to take. There's nothing about consciousness that breaks any known law of physics today, so the only logical position is to suppose that consciousness is explainable by current physics. We can't go around positing unknown new physics behind every phenomenon we haven't entirely characterized and understood yet.
Quite the claim to make
No. Because even if you had solid evidence for the hypothesis that quantum mechanical effects are indispensable in making our brains work (which we don't), then that is still not preventing simulation. You need some uncomputable component, which physics right now neither provides nor predicts.
And fleeing into "we don't know 100% of physics yet" is a bad hypothesis, because we can make very accurate physical predictions already-- you would need our brains to "amplify" some very small gap in our physical understanding, and this does not match with how "robust" the operation of our brain is-- amplifiers, by their very nature, are highly sensitive to disruption or disturbances, but a human can stay conscious even with a particle accelerator firing through his brain.
Imagine we assume that A.I. could be conscious. What would be the identity/scope of that consciousness. To understand what I'm driving at, let's make an analogy to humans. Our consciousness is scoped to our bodies. We see through sense organ, and our brain, which process these signals, is located in a specific point in space. But we still do not know how consciousness arises in the brain and is bound to the body.
If you equate computation of sufficient complexity to consciousness, then the question arises: what exactly about computation would prodcuce consciousness? If we perform the same computation on a different substrate, would that then be the same consciousness, or a copy of the original? If it would not be the same consciousness, then just what give consciousness its identity?
I believe you would find it ridiculous to say that just because we are performing the computation on this chip, therefore the identity of the resulting consciousness is scoped to this chip.
Similarly, if we can compute consciousness on a chip, then the chip obviously contains that consciousness. You can experimentally determine to what extent this is true: for example, you can experimentally check if increasing the clock frequency of said chip alters the consciousness that it is computing. Or if changing the thermal paste that attaches it to its cooler does so. I don't know what the results of these experiments would be, but they would be quite clearly determined.
Of course, there would certainly be some scale, and at some point it becomes semantics. The same is true with human consciousness: some aspects of the body are more tightly coupled to consciousness than others; if you cut my hand, my consciousness will change more than if you cut a small piece of my bowel, but less than if you cut out a large piece of my brain. At what point do you draw the line and say "consciousness exists in the brain but not the hands"? It's all arbitrary to some extent. Even worse, say I use a journal where I write down some of my most cherished thoughts, and say that I am quite forgetful and I often go through this journal to remind myself of various thoughts before taking a decision. Would it not then be fair to say that the journal itself contains a part of my consciousness? After all, if someone were to tamper with it in subtle enough ways, they would certainly be able to influence my thought process, more so than even cutting off one of my hands, wouldn't they?
> Similarly, if we can compute consciousness on a chip, then the chip obviously contains that consciousness.
This is like claiming that neurons are conscious, which as far as we can tell, they are not. For all you know, it is the algorithm that could be conscious. Or some interplay between the algorithm and the substrate, OR something else.
Another way to think of it problem: Imagine a massive cluster performing computation that is thought to give rise to consciousness. Is is the cluster that is conscious? Or the individual machines, or the chips, or the algorithm, or something else?
I personally don't think any of these can be conscious, but those that do should explain how they figure these thing out.
The bound informational dynamic that supervenes on the activity of the individual units in the cluster. What people typically miss is that the algorithm when engaged in a computing substrate is not just inert symbols, but an active, potent causal/dynamical structure. Information flows as modulated signals to and from each component and these signals are integrated such that the characteristic property of the aggregate signal is maintained. This binding of signals by the active interplay of component signals from the distributed components realizes the singular identity. If there is consciousness here, it is in this construct.
Of course, "the consciousness" is a nebulous concept. It would be like asking "which part of my processor is Windows" to some extent. But it's still fair to say that Windows is contained within my computer, and that the metal framing of the computer is not part of Windows.
Well, first I would ask whether this question makes sense in the first place. Does consciousness have a scope? Does consciousness even exist? Or is that more of a name attributed to some pattern we recognize in our own way of thinking (but may not be universal)?
Also, would a person missing an arm, but having a robot arm they can control have their consciousness' "scope" extended to it? Given that people have phantom pains, does a physical body even needed to consider it your part?
And the monkey strategy has been 100% dismissed as shit..
We know how to deploy monkeys on typewriters, but we don't know what they'll type.
We know how to deploy transformers to train and inference a model, but we don't know what they'll type.
We DON'T know how a thinking human (or animal) brain works..
Do you see the difference.
We don't claim a JPEG can paint great art, even though certain jpegs do.
But yes, we built a machine that generates text similar to what we built it from, and now we're looking at it generating text and are all impressed.
If you put a million monkeys on typewriters you would eventually get shakespeare is exactly why LLM's will succeed and why humans have succeeded. If this weren't the case why didn't humans 30000 years ago create spacecraft if we were endowed with the same natural "gift".
A scientific paper has to be verifiable, you should be able to recreate the experiment and come to the same conclusion. It's very very difficult to do with brains with trillions of parameters and that can't be controlled to the neuron level. Nothwithstanding the ethical issues.
We don't have a world weather simulator that is 100% accurate either given the complex interplay and inability to control the variables i.e. it's not verifiable. It'd be a bit silly to say we don't know why it's going to rain at my house tomorrow.
Until then it is a hypothesis, and we can't say we know even if the overwhelming evidence indicates that in fact that we do know.
> Do you see the difference
No? I'm not sure what you're getting at.
You're going off the rails IMMEDIATELY in your logic.
Sure, one Turing-complete computer language can have its logic "emulated" by another, fine. But human intelligence is not a computer language -- you're mixing up the terms "Turing complete" and "Turing test".
It's like mixing up the terms "Strawberry jam" and "traffic jam" and then going on to talk about how cars taste on toast. It's nonsensical.
Give my comment another read, but it was quite understandable from context. (Also, you may want to give a read to the Turing paper because being executable by a person as well was an important concept within)
Actually, the way LLMs are extended with tools is a pretty much the same (an LLM itself has no access to the internet, but if it returns some specific symbols, the external "glue" will do a search and then the LLM is free to use the results)
This is the equivalent of saying "I have set up a complex mechanical computer powered by water that is Turing complete. Since any Turing complete system can emulate another one, it means that any other Turing complete system can also make things wet and irrigate farms.
Human intelligence is not understood. It can be made to do Turing complete things, but you can't invert that and say that because you've read the paper on Turing completeness, you now understand human intelligence.
Who stated that sentience or sense of self is a part of thinking?
Rather, these models are literally grown during the training phase. And all the intelligence emerges from that growth. That's what makes them a black box and extremely difficult to penetrate. No one can say exactly how they work inside for a given problem.
What I mean by that is that I think there is a good chance that LLMs are similar to a subsystem of human thinking. They are great at pattern recognition and prediction, which is a huge part of cognition. What they are not is conscious, or possessed of subjective experience in any measurable way.
LLMs are like the part of your brain that sees something and maps it into a concept for you. I recently watched a video on the creation of AlexNet [0], one of the first wildly successful image-processing models. One of the impressive things about it is how it moves up the hierarchy from very basic patterns in images to more abstract ones (e. g. these two images' pixels might not be at all the same, but they both eventually map to a pattern for 'elephant').
It's perfectly reasonable to imagine that our brains do something similar. You see a cat, in some context, and your brain maps it to the concept of 'cat', so you know, 'that's a cat'. What's missing is a) self-motivated, goal-directed action based on that knowledge, and b) a broader context for the world where these concepts not only map to each other, but feed into a sense of self and world and its distinctions whereby one can say: "I am here, and looking at a cat."
It's possible those latter two parts can be solved, or approximated, by an LLM, but I am skeptical. I think LLMs represent a huge leap in technology which is simultaneously cooler than anyone would have imagined a decade ago, and less impressive than pretty much everyone wants you to believe when it comes to how much money we should pour into the companies that make them.
[0] https://www.youtube.com/watch?v=UZDiGooFs54
We don't know how to measure subjective experience in other people, even, other than via self-reporting, so this is a meaningless statement. Of course we don't know whether they are, and of course we can't measure it.
I also don't know for sure whether or not you are "possessed of subjective experience" as I can't measure it.
> What they are not is conscious
And this is equally meaningless without your definition of "conscious".
> It's possible those latter two parts can be solved, or approximated, by an LLM, but I am skeptical.
Unless we can find indications that humans can exceed the Turing computable - something we as of yet have no indication is even theoretically possible - there is no rational reason to think it can't.
But doesn't this rely on the same thing you suggest we don't have, which is a working and definable definition of consciousness?
I think a lot of the 'well, we can't define consciousness so we don't know what it is so it's worthless to think about' argument - not only from you but from others - is hiding the ball. The heuristic, human consideration of whether something is conscious is an okay approximation so long as we avoid the trap of 'well, it has natural language, so it must be conscious.'
There's a huge challenge in the way LLMs can seem like they are speaking out of intellect and not just pattern predicting, but there's very little meaningful argument that they are actually thinking in any way similarly to what you or I do in writing these comments. The fact that we don't have a perfect, rigorous definition, and tend to rely on 'I know it when I see it,' does not mean LLMs do have it or that it will be trivial to get to them.
All that is to say that when you say:
> I also don't know for sure whether or not you are "possessed of subjective experience" as I can't measure it.
"Knowing for sure" is not required. A reasonable suspicion one way or the other based on experience is a good place to start. I also identified two specific things LLMs don't do - they are not self-motivated or goal-directed without prompting, and there is no evidence they possess a sense of self, even with the challenge of lack of definition that we face.
Without that any limitations borne out of what LLMs don't currently do are irrelevant.
1. Humans intellect is Turing computable. 2. LLMs are based on Turing-complete technology. 3. Therefore, LLMs can eventually equal human intellect.
But if that is the right chain of assumptions, there's lots of issues with it. First, whether LLMs are Turing complete is a topic of debate. There are points for[0] and against[1].
I suspect they probably _are_, but that doesn't mean LLMs are tautologically indistinguishable from human intelligence. Every computer that uses a Turing-complete programming language can theoretically solve any Turing-computable problem. That does not mean they will ever be able to efficiently or effectively do so in real time under real constraints, or that they are doing so now in a reasonable amount real-world time using extant amounts of real-world computing power.
The processor I'm using to write this might be able to perform all the computations needed for human intellect, but even if it could, that doesn't mean it can do it quickly enough to compute even a single nanosecond of actual human thought before the heat-death of the universe, or even the end of this century.
So when you say:
> Without that any limitations borne out of what LLMs don't currently do are irrelevant.
It seems to me exactly the opposite is true. If we want technology that is anything approaching human intelligence, we need to find approaches which will solve for a number of things LLMs don't currently do. The fact that we don't know exactly what those things are yet is not evidence that those things don't exist. Not only do they likely exist, but the more time we spend simply scaling LLMs instead of trying to find them, the farther we are from any sort of genuine general intelligence.
[0] https://arxiv.org/abs/2411.01992 [1] https://medium.com/heyjobs-tech/turing-completeness-of-llms-...
Yes, with an emphasis on can. That does not mean they necessarily will. Though I would consider it unlikely that they won't, we the only way of proving that they will would be to do it.
> But if that is the right chain of assumptions, there's lots of issues with it. First, whether LLMs are Turing complete is a topic of debate. There are points for[0] and against[1].
It's trivial to prove that a system comprised of an LLM with a loop is Turing computable. A single inference step can not be Turing computable, but one with a loop only requires the LLM to be capable of executing 6 distinct steps with temperatur set to 0. You can wire up a toy neural network by hand that can do this.
This is in fact a far more limited claim than what the paper you linked to makes.
The article you linked to, on the other hand, is discussing if an LLM can act like a Turing machine without that loop. That is why "state management" matters. State management is irrelevant when you wrap a loop around because you can externalise the state, and you only need 2 states (and 3 symbols, or 3 states and 2 symbols) for the smallest known universal Turing machine.
The entire article is thus entirely irrelevant to this question. Sure, you will struggle to make an LLM act as a Turing machine without going "off the rails". But that is irrelevant - you only need it to be able to execute one state transition by deterministically producing the right next tape operation and next state when given the next symbol and current state.
From that you can build up to arbitrarily complex computation, because for every given Turing machine, you can construct a larger Turing machine that uses additional symbols or states to encode an operation that takes multiple steps for the smaller machine.
> The processor I'm using to write this might be able to perform all the computations needed for human intellect, but even if it could, that doesn't mean it can do it quickly enough to compute even a single nanosecond of actual human thought before the heat-death of the universe, or even the end of this century.
Irrelevant, because absent physics that exceeds the Turing computable, the human brain is an existence-proof for the possibility of computing everything the human brain does in a package the size of an average human brain.
It is very likely that we will need architectural changes to compute any given model efficiently enough, but to suggest it is not possible is an extraordinary claim not supported by anything.
If you take LLM to mean a very specific architecture, or specific computational methods to execute a model, then you have a point. If so we are talking about very different things.
No, it's like saying we have no indication that humans have psychic powers and can levitate objects with their minds. The commenter is saying no human has ever demonstrated the ability to figure things out that aren't Turing computable and we have no reason to suspect this ability is even theoretically possible (for anything, human or otherwise).
Then why make an argument based on what you do not know?
You say the limits of LLMs don't matter, because we don't have definitions strong enough to describe them.
I say the limits of LLMs do matter and the fact that we can't yet define them rigorously means we aren't able to fix them (assuming we want to).
Model/algorithm is the banks. Water could be the experiences. Maybe the algorithm does not have consciousness, but it is part of it.
They co-create each other. They are part of a recursive loop which cannot be explained statically, or part by part in isolation.
LLMs (AIs) are not useless. But they do not actually think. What is trivially true is that they do not actually need to think. (As far as the Turing Test, Eliza patients, and VC investors are concerned, the point has been proven.)
If the technology is helping us write text and code, it is by definition useful.
> In 2003, the machine-learning researcher Eric B. Baum published a book called “What Is Thought?” [...] The gist of Baum’s argument is that understanding is compression, and compression is understanding.
This is incomplete. Compression is optimisation, optimisation may resemble understanding, but understanding is being able to verify that a proposition (compressed rule or assertion) is true or false or even computable.
> —but, in my view, this is the very reason these models have become increasingly intelligent.
They have not become more intelligent. The training process may improve, the vetting of the data improved, the performance may improve, but the resemblance to understanding only occurs when the answers are provably correct. In this sense, these tools work in support of (are therefore part of) human thinking.
The Stochastic Parrot is not dead, it's just making you think it is pining for the fjords.
I'm so baffled when I see this being blindly asserted.
With the reasoning models, you can literally watch their thought process. You can see them pattern-match to determine a strategy to attack a problem, go through it piece-by-piece, revisit assumptions, reformulate strategy, and then consolidate findings to produce a final result.
If that's not thinking, I literally don't know what is. It's the same process I watch my own brain use to figure something out.
So I have to ask you: when you claim they don't think -- what are you basing this on? What, for you, is involved in thinking that the kind of process I've just described is missing? Because I genuinely don't know what needs to be added here for it to become "thinking".
Not true, you are falling for a very classic (prehistoric, even) human illusion known as experiencing a story:
1. There is a story-like document being extruded out of a machine humans explicitly designed for generating documents, and which humans trained on a bajillion stories humans already made.
2. When you "talk" to a chatbot, that is an iterative build of a (remote, hidden) story document, where one of the characters is adopting your text-input and the other's dialogue is being "performed" at you.
3. The "reasoning" in newer versions is just the "internal monologue" of a film noir detective character, and equally as fictional as anything that character "says out loud" to the (fictional) smokin-hot client who sashayed the (fictional) rent-overdue office bearing your (real) query on its (fictional) lips.
> If that's not thinking, I literally don't know what is.
All sorts of algorithms can achieve useful outcomes with "that made sense to me" flows, but that doesn't mean we automatically consider them to be capital-T Thinking.
> So I have to ask you: when you claim they don't think -- what are you basing this on?
Consider the following document from an unknown source, and the "chain of reasoning" and "thinking" that your human brain perceives when encountering it:
Now whose reasoning/thinking is going on? Can you point to the mind that enjoys steel and manure? Is it in the room with us right now? :PIn other words, the reasoning is illusory. Even if we accept that the unknown author is a thinking intelligence for the sake of argument... it doesn't tell you what the author's thinking.
But this is false, because the thinking exhibits cause and effect and a lot of good reasoning. If you change the inputs, the thinking continues to be pretty good with the new inputs.
It's not a story, it's not fictional, it's producing genuinely reasonable conclusions around data it hasn't seen before. So how is it therefore not actual thinking?
And I have no idea what your short document example has to do with anything. It seems nonsensical and bears no resemblance to the actual, grounded chain of thought processes high-quality reasoning LLM's produce.
> OK, so that document technically has a "chain of thought" and "reasoning"... But whose?
What does it matter? If an LLM produces output, we say it's the LLM's. But I fail to see how that is significant?
Creating complex Excel spreadsheet structures comes to mind, I just did that earlier today - and with plain GPT-5, not even -Thinking. Sure, maybe the Excel formulas themselves are a "98% match" to training data, but it takes real cognition (or whatever you want to call it) to figure out which ones to use and how to use them appropriately for a given situation, and how to structure the spreadsheet etc.
But it thinks just great on stuff it wasn't trained on.
I give it code I wrote that is not in its training data, using new concepts I've come up with in an academic paper I'm writing, and ask it to extend the code in a certain way in accordance with those concepts, and it does a great job.
This isn't regurgitation. Even if a lot of LLM usage is, the whole point is that it does fantastically with stuff that is brand new too. It's genuinely creating new, valuable stuff it's never seen before. Assembling it in ways that require thinking.
With how much data is being shoveled in there, our default assumption should be that significant components are present.
Many consider "thinking" something only animals can do, and they are uncomfortable with the idea that animals are biological machines or that life, consciousness, and thinking are fundamentally machine processes.
When an LLM generates chain-of-thought tokens, what we might casually call “thinking,” it fills its context window with a sequence of tokens that improves its ability to answer correctly.
This “thinking” process is not rigid deduction like in a symbolic rule system; it is more like an associative walk through a high-dimensional manifold shaped by training. The walk is partly stochastic (depending on temperature, sampling strategy, and similar factors) yet remarkably robust.
Even when you manually introduce logical errors into a chain-of-thought trace, the model’s overall accuracy usually remains better than if it had produced no reasoning tokens at all. Unlike a strict forward- or backward-chaining proof system, the LLM’s reasoning relies on statistical association rather than brittle rule-following. In a way, that fuzziness is its strength because it generalizes instead of collapsing under contradiction.
Everyday reasoning is full of heuristics, analogies, and pattern matches: we jump to conclusions, then backfill justification afterward. Psychologists call this “post hoc rationalization,” and there’s plenty of evidence that people form beliefs first and then search for logical scaffolding to support them. In fact, that’s how we manage to think fluidly at all; the world is too noisy and underspecified for purely deductive inference to function outside of controlled systems.
Even mathematicians, our best examples of deliberate, formal thinkers, often work this way. Many major proofs have been discovered intuitively and later found to contain errors that didn’t actually invalidate the final result. The insight was right, even if the intermediate steps were shaky. When the details get repaired, the overall structure stands. That’s very much like an LLM producing a chain of reasoning tokens that might include small logical missteps yet still landing on the correct conclusion: the “thinking” process is not literal step-by-step deduction, but a guided traversal through a manifold of associations shaped by prior experience (or training data, in the model’s case).
So if an LLM doesn’t collapse under contradictions, that’s not necessarily a bug; it may reflect the same resilience we see in human reasoning. Our minds aren’t brittle theorem provers; they’re pattern-recognition engines that trade strict logical consistency for generalization and robustness. In that sense, the fuzziness is the strength.
Yes, but that seems like moving the goalposts.
The stricter blends of reasoning are what everybody is so desperate to evoke from LLMs, preferably along with inhuman consistency, endurance, and speed. Just imagine the repercussions if a slam-dunk paper came out tomorrow, which somehow proved the architectures and investments everyone is using for LLMs are a dead-end for that capability.
This is definitely not true for me. My prompts frequently contain instructions that aren't 100% perfectly clear, suggest what I want rather than formally specifying it, typos, mistakes, etc. The fact that the LLM usually figures out what I meant to say, like a human would, is a feature for me.
I don't want an LLM to act like an automated theorem prover. We already have those. Their strictness makes them extremely difficult to use, so their application is extremely limited.
On reasoning, I see LLMs and classic algorithms as complements. LLMs do robust manifold following and associative inference. Traditional programs do brittle rule following with guarantees. The promising path looks like a synthesis where models use tools, call code, and drive search and planning methods such as MCTS, the way AlphaGo did. Think agentic systems that can read, write, execute, and verify.
LLMs are strongest where the problem is language. Language co evolved with cognition as a way to model the world, not just to chat. We already use languages to describe circuits, specify algorithms, and even generate other languages. That makes LLMs very handy for specification, coordination, and explanation.
LLMs can also statistically simulate algorithms, which is useful for having them think about these algorithms. But when you actually need the algorithm, it is most efficient to run the real thing in software or on purpose built hardware. Let the model write the code, compose the tools, and verify the output, rather than pretending to be a CPU.
To me the risk is not that LLMs are a dead end, but that people who do not understand them have unreasonable expectations. Real progress looks like building systems that use language to invent and implement better tools and route work to the right place. If a paper lands tomorrow that shows pure next token prediction is not enough for formal reasoning, that would be an example of misunderstanding LLMs, not a stop sign. We already saw something similar when Minsky and Papert highlighted that single layer perceptrons could not represent XOR, and the field later moved past that with multilayer networks. Hopefully we remember that and learn the right lesson this time.
An LLM is a noise generator. It generates tokens without logic, arithmetic, or any "reason" whatsoever. The noise that an LLM generates is not truly random. Instead, the LLM is biased to generate familiar noise. The LLM itself is nothing more than a model of token familiarity. Nothing about that model can tell you why some tokens are more familiar with others, just like an accounting spreadsheet can't tell you why it contains a list of charges and a summation next to the word "total". It could just as easily contain the same kind of data with an entirely different purpose.
What an LLM models is written human text. Should we really expect to not be surprised by the power and versatility of human-written text?
---
It's clear that these statistical models are very good at thoughtless tasks, like perception and hallucination. It's also clear that they are very bad at thoughtful tasks like logic and arithmetic - the things that traditional software is made of. What no one has really managed to figure out is how to bridge that gap.
My experience, though, is that they aren't good at defining the task to be coded, or thinking about some unexpected side-effects. Code that will be left for them to develop freely will likely become bloated quite fast.
If I had to guess, the current leading LLMs consciousness is most comparable to a small fish, with a conscious lifespan of a few seconds to a few minutes. Instead of perceiving water, nutrient gradients, light, heat, etc. it's perceiving tokens. It's conscious, but it's consciousness is so foreign to us it doesn't seem like consciousness. In the same way to an amoeba is conscious or a blade of grass is conscious but very different kind than we experience. I suspect LLMs are a new type of consciousness that's probably more different from ours than most if not all known forms of life.
I suspect the biggest change that would bring LLM consciousness closer to us would be some for of continuous learning/model updating.
Until then, even with RAG, and other clever teghniques I consider these models as having this really foreign slices of consciousness where they "feel" tokens and "act" out tokens, and they have perception, but their perception of the tokens is nothing like ours.
If one looks closely at simple organisms with simple sensory organs and nervous systems its hard not to see some parallels. It's just that the shape of consciousness is extremely different than any life form. (perception bandwidth, ability to act, temporality, etc)
Karl friston free energy principle gives a really interesting perspective on this I think.
To put it another way: lots of things are conscious, but humans are definitely the most conscious beings on Earth.
I am not sure if chimps can do the same.
If one is to quantify consciousness it would probably make sense to think of it as an area of awareness and cognizance across time.
Awareness scales with sensory scale and resolution (sensory receptors vs input token limits and token resolution). E.g. 128k tokens and tokens too coarse to count rs in strawberry.
Cognizance scales with internal representations of awareness (probably some relation to vector space resolution and granularity, though I suspect there is more to it than just vector space)
And the third component is time, how long the agent is conscious for.
So something like...
Time * awareness (receptors) * internal representations (cell diversity * # cells * connection diversity * # connections)
There is no way this equation is right but I suspect it's sort of directionally correct.
I'm deep in the subject but just riffing here, so take this with a lot of salt.
It's why anyone who's ever taken care of a needy pet will inevitably reach the comparison that taking care of a pet is similar to taking care of a very young child; it's needy, it experiences emotions but it can't quite figure out on its own how to adapt to an environment besides what it grew up around/it's own instincts. They experience some sort of qualia (a lot of animals are pretty family-minded), but good luck teaching a monkey to read. The closest we've gotten is teaching them that if they press the right button, they get food, but they take basically their entire lifespan to understand a couple hundred words, while humans easily surpass that.
IIRC some of the smartest animals in the world are actually rats. They experience a qualia very close to humans to the point that psychology experiments are often easily observable in rats.
Totally possible that we're talking past each other.
Why?
Very short answer is Karl Friston's free energy pricniple
851 more comments available on Hacker News