logoalt Hacker News

tkz1312last Monday at 10:13 PM37 repliesview on HN

Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

Consciousness or self awareness is of course a different question, and ones whose answer seems less clear right now.

Knee jerk dismissing the evidence in front of your eyes because you find it unbelievable that we can achieve true reasoning via scaled matrix multiplication is understandable, but also betrays a lack of imagination and flexibility of thought. The world is full of bizarre wonders and this is just one more to add to the list.


Replies

keiferskiyesterday at 8:58 AM

I don’t see how being critical of this is a knee jerk response.

Thinking, like intelligence and many other words designating complex things, isn’t a simple topic. The word and concept developed in a world where it referred to human beings, and in a lesser sense, to animals.

To simply disregard that entire conceptual history and say, “well it’s doing a thing that looks like thinking, ergo it’s thinking” is the lazy move. What’s really needed is an analysis of what thinking actually means, as a word. Unfortunately everyone is loathe to argue about definitions, even when that is fundamentally what this is all about.

Until that conceptual clarification happens, you can expect endless messy debates with no real resolution.

“For every complex problem there is an answer that is clear, simple, and wrong.” - H. L. Mencken

show 13 replies
notepad0x90yesterday at 7:13 AM

I don't get why you would say that. it's just auto-completing. It cannot reason. It won't solve an original problem for which it has no prior context to "complete" an approximated solution with. you can give it more context and more data,but you're just helping it complete better. it does not derive an original state machine or algorithm to solve problems for which there are no obvious solutions. it instead approximates a guess (hallucination).

Consciousness and self-awareness are a distraction.

Consider that for the exact same prompt and instructions, small variations in wording or spelling change its output significantly. If it thought and reasoned, it would know to ignore those and focus on the variables and input at hand to produce deterministic and consistent output. However, it only computes in terms of tokens, so when a token changes, the probability of what a correct response would look like changes, so it adapts.

It does not actually add 1+2 when you ask it to do so. it does not distinguish 1 from 2 as discrete units in an addition operation. but it uses descriptions of the operation to approximate a result. and even for something so simple, some phrasings and wordings might not result in 3 as a result.

show 11 replies
simulator5gyesterday at 6:38 AM

Having seen photocopiers so many times produce coherent, sensible, and valid chains of words on a page, I am at this point in absolutely no doubt that they are thinking.

show 6 replies
ben_wyesterday at 11:06 AM

> Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

While I'm not willing to rule *out* the idea that they're "thinking" (nor "conscious" etc.), the obvious counter-argument here is all the records we have of humans doing thinking, where the records themselves are not doing the thinking that went into creating those records.

And I'm saying this as someone whose cached response to "it's just matrix multiplication it can't think/be conscious/be intelligent" is that, so far as we can measure all of reality, everything in the universe including ourselves can be expressed as matrix multiplication.

Falsification, not verification. What would be measurably different if the null hypothesis was wrong?

show 1 reply
layer8yesterday at 12:04 AM

Sometimes after a night’s sleep, we wake up with an insight on a topic or a solution to a problem we encountered the day before. Did we “think” in our sleep to come up with the insight or solution? For all we know, it’s an unconscious process. Would we call it “thinking”?

The term “thinking” is rather ill-defined, too bound to how we perceive our own wakeful thinking.

When conversing with LLMs, I never get the feeling that they have a solid grasp on the conversation. When you dig into topics, there is always a little too much vagueness, a slight but clear lack of coherence, continuity and awareness, a prevalence of cookie-cutter verbiage. It feels like a mind that isn’t fully “there” — and maybe not at all.

I would agree that LLMs reason (well, the reasoning models). But “thinking”? I don’t know. There is something missing.

show 7 replies
geonyesterday at 12:23 AM

Having seen LLMs so many times produce incoherent, nonsensical and invalid chains of reasoning...

LLMs are little more than RNGs. They are the tea leaves and you read whatever you want into them.

show 2 replies
marcus_holmesyesterday at 4:59 AM

Yes, I've seen the same things.

But; they don't learn. You can add stuff to their context, but they never get better at doing things, don't really understand feedback. An LLM given a task a thousand times will produce similar results a thousand times; it won't get better at it, or even quicker at it.

And you can't ask them to explain their thinking. If they are thinking, and I agree they might, they don't have any awareness of that process (like we do).

I think if we crack both of those then we'd be a lot closer to something I can recognise as actually thinking.

show 3 replies
josefxyesterday at 7:40 AM

Counterpoint: The seahorse emoji. The output repeats the same simple pattern of giving a bad result and correcting it with another bad result until it runs out of attempts. There is no reasoning, no diagnosis, just the same error over and over again within a single session.

show 2 replies
techblueberryyesterday at 12:39 PM

Isn’t anthropomorphizing LLMs rather than understanding their unique presence in the world a “ lack of imagination and flexibility of thought”? It’s not that I can’t imagine applying the concept “thinking” to the output on the screen, I just don’t think it’s an accurate description.

show 1 reply
ryanackleyyesterday at 11:42 AM

I think we can call it "thinking" but it's dangerous to anthropomorphize LLMs. The media and AI companies have an agenda when doing so.

didibusyesterday at 3:22 AM

I guess it depends if you definite thinking thinking as chaining coherent reasoning sentences together 90-some% of the time.

But if you define thinking as the mechanism and process we mentally undergo and follow mentally... I don't think we have any clue if that's the same. Do we also just vector-map attention tokens and predict the next with a softmax? I doubt, and I don't think we have any proof that we do.

show 1 reply
noivyesterday at 7:12 AM

Different PoV: You have a local bug and ask the digital hive mind for a solution, but someone already solved the issue and their solution was incorporated... LLMs are just very effficient at compressing billions of solutions into a few GB.

Try to ask something no one ever came up with a solution so far.

show 1 reply
conartist6last Monday at 11:57 PM

Yeah but if I assign it a long job to process I would also say that an x86 CPU is "thinking" about a problem for me.

What we really mean in both cases is "computing," no?

khafrayesterday at 7:01 AM

"Consciousness" as in subjective experience, whatever it is we mean by "the hard problem," is very much in doubt.

But "self-awareness," as in the ability to explicitly describe implicit, inner cognitive processes? That has some very strong evidence for it: https://www.anthropic.com/research/introspection

darthvadenyesterday at 9:13 AM

If AI is thinking if slavery is bad then how can somebody own AI. How can investors can shares from AI profits? We are ok with slavery now. Ok i will have two black slaves now. Who can ask me? Why shld that be illegal?

show 2 replies
lispybananayesterday at 9:31 AM

Would they have diagnosed an issue if you hadn't presented it to them?

Life solves problems itself poses or collides with. Tools solve problems only when applied.

conartist6yesterday at 11:06 AM

So an x86 CPU is thinking?

So many times I've seen it produce sensible, valid chains of results.

Yes, I see evidence in that outcome that a person somewhere thought and understood. I even sometimes say that a computer is "thinking hard" about something when it freezes up.

...but ascribing new philosophical meaning to this simple usage of the word "thinking" is a step too far. It's not even a new way of using the word!

show 1 reply
raincolelast Monday at 11:20 PM

I'd represent the same idea but in a different way:

I don't know what the exact definition of "thinking" is. But if a definition of thinking rejects the possibility of that current LLMs think, I'd consider that definition useless.

show 1 reply
mlsuyesterday at 5:26 AM

They remind me of the apparitions in Solaris. They have this like mechanical, almost player-piano like quality to them. They both connect with and echo us at the same time. It seems crazy to me and very intellectually uncreative to not think of this as intelligence.

donkeybeeryesterday at 8:35 AM

Its overt or unaware religion. The point when you come down to the base of it is that these people believe in "souls".

johnnienakedyesterday at 3:06 AM

If you understand how they operate and you are reasonable and unbiased there is no way you could consider it thinking

ForHackernewsyesterday at 12:56 PM

But all those times the same system produces irrational gibberish don't count? GPT-5 will commonly make mistakes no thinking human could ever make.

Human: I'm trying to get my wolf, sheep and cabbage across the river in this boat, but the wolf keeps eating the sheep or the sheep eats the cabbage

Bot: You should put the sheep in the boat and take it across — if we delve into the biology of Canis lupus we discover that wolves don't eat cabbage!

H: Ok, so that worked great so far, the sheep is on one side and the wolf/cabbage is on the other.

B: Now, Option 1 is to bring the wolf across, or Option 2 you can bring the cabbage. I recommend (2) taking the cabbage as cabbages are smaller and easier to transport in a boat.

H: But then the sheep eats the cabbage, right? Remember that?

B: Exactly, that's sharp thinking. If you put the sheep and the cabbage together on the same side of the river, the sheep is sure to devour the cabbage. We need to not just separate sheep from cabbages — we need to separate cabbages from sheep! :rocketship:

triyambakamyesterday at 12:37 AM

> Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

People said the same thing about ELIZA

> Consciousness or self awareness is of course a different question,

Then how do you define thinking if not a process that requires consciousness?

show 1 reply
satisficeyesterday at 4:26 AM

I think you are the one dismissing evidence. The valid chains of reasoning you speak of (assuming you are talking about text you see in a “thinking model” as it is preparing its answer) are narratives, not the actual reasoning that leads to the answer you get.

I don’t know what LLMs are doing, but only a little experimentation with getting it to describe its own process shows that it CAN’T describe its own process.

You can call what a TI calculator does “thinking” if you want. But what people are interested in is human-like thinking. We have no reason to believe that the “thinking” of LLMs is human-like.

show 1 reply
lordnachoyesterday at 8:29 AM

I agree with you.

If you took a Claude session into a time machine to 2019 and called it "rent a programmer buddy," how many people would assume it was a human? The only hint that it wasn't a human programmer would be things where it was clearly better: it types things very fast, and seems to know every language.

You can set expectations in the way you would with a real programmer: "I have this script, it runs like this, please fix it so it does so and so". You can do this without being very precise in your explanation (though it helps) and you can make typos, yet it will still work. You can see it literally doing what you would do yourself: running the program, reading the errors, editing the program, and repeating.

People need to keep in mind two things when they compare LLMs to humans: you don't know the internal process of a human either, he is also just telling you that he ran the program, read the errors, and edited. The other thing is the bar for thinking: a four-year old kid who is incapable of any of these things you would not deny as a thinking person.

show 1 reply
Zardoz84yesterday at 7:56 AM

Having seen parrots so many times produce coherent, sensible, and valid chains of sounds and words, I am at this point in absolutely no doubt that they are thinking.

show 1 reply
intendedyesterday at 11:52 AM

what sound does a falling tree make if no one is listening?

I’ve asked LLMs to write code for me in fields I have little background knowledge, and then had to debug the whole thing after essentially having to learn the language and field.

On the other hand, for things I am well versed in, I can debug the output and avoid entire swathes of failed states, by having a clear prompt.

Its why I now insist that any discussion on GenAI projects also have the speaker mention the level of seniority they have ( proxy for S/W eng experience), Their familiarity with the language, the project itself (level of complexity) - more so than the output.

I also guarantee - that most people have VERY weak express knowledge of how their brains actually work, but deep inherent reflexes and intuitions.

flanked-everglyesterday at 10:55 AM

"Convince" the stock Claude Sonnet 4.5 that it's a sentient human being hooked up to Neuralink and then tell me again it's thinking. It's just not.

hagbarthyesterday at 9:21 AM

I'm not so sure. I, for one, do not think purely by talking to myself. I do that sometimes, but a lot of the time when I am working through something, I have many more dimensions to my thought than inner speech.

belteryesterday at 9:33 AM

Apparent reasoning can emerge from probabilistic systems that simply reproduce statistical order not genuine understanding.

Weather models sometimes “predict” a real pattern by chance, yet we don’t call the atmosphere intelligent.

If LLMs were truly thinking, we could enroll one at MIT and expect it to graduate, not just autocomplete its way through the syllabus or we could teach one how to drive.

yawpitchyesterday at 8:12 AM

You’re assuming the issues and bugs you’ve been addressing don’t already exist, already encoding human chain of reasoning, in the training data.

NoMoreNicksLeftyesterday at 5:51 AM

>Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking.

If one could write a quadrillion-line python script of nothing but if/elif/else statements nested 1 million blocks deep that seemingly parsed your questions and produced seemingly coherent, sensible, valid "chains of reasoning"... would that software be thinking?

And if you don't like the answer, how is the LLM fundamentally different from the software I describe?

>Knee jerk dismissing the evidence in front of your eyes because

There is no evidence here. On the very remote possibility that LLMs are at some level doing what humans are doing, I would then feel really pathetic that humans are as non-sapient as the LLMs. The same way that there is a hole in your vision because of a defective retina, there is a hole in your cognition that blinds you to how cognition works. Because of this, you and all the other humans are stumbling around in the dark, trying to invent intelligence by accident, rather than just introspecting and writing it out from scratch. While our species might someday eventually brute force AGI, it would be many thousands of years before we get there.

show 2 replies
smohareyesterday at 1:56 PM

[dead]

veegeeyesterday at 8:12 AM

[dead]

absurd1styesterday at 9:59 AM

[dead]

ath3ndlast Monday at 11:20 PM

[dead]