I think you missed the point of that quote. Birds fly, and airplanes fly; fish swim but submarines don't. It's an accident of language that we define "swim" in a way that excludes what submarines do. They move about under their own power under the water, so it's not very interesting to ask whether they "swim" or not.
Most people I've talked to who insist that LLMs aren't "thinking" turn out to have a similar perspective: "thinking" means you have to have semantics, semantics require meaning, meaning requires consciousness, consciousness is a property that only certain biological brains have. Some go further and claim that reason, which (in their definition) is something only human brains have, is also required for semantics. If that's how we define the word "think", then of course computers cannot be thinking, because you've defined the word "think" in a way that excludes them.
And, like Dijkstra, I find that discussion uninteresting. If you want to define "think" that way, fine, but then using that definition to insist LLMs can't do a thing because it can't "think" is like insisting that a submarine can't cross the ocean because it can't "swim".
Then you're missing the point of my rebuttal. You say submarines don't swim [like fish] despite both moving through water, the only distinction is mechanism. Can AI recursively create new capabilities like thinking does, or just execute tasks like submarines do? That's the question.
Reading the quote in context seems to indicate Dijkstra meant something else. His article is a complaint about overselling computers as doing or augmenting the thinking for humans. It's funny how the quote was lifted out of an article and became famous on it's own.
https://www.cs.utexas.edu/~EWD/transcriptions/EWD08xx/EWD867...