You’re assuming that thinking requires learning, which I don’t necessarily agree with. Humans can have brain damage which inhibits the formation of long term memories, but such people can still function in the world. Would you say the thing such a person’s brain is doing is something other than thinking?
At any rate, just because the architecture of current LLMs doesn’t support learning at inference time does not constitute a fundamental limit that can never be changed, just a local maximum that has worked well to productize the approach.
And I’m quite certain that once systems that include post-training learning exist people like you will find a way to distinguish that from human learning, moving the goalposts again. You’re not arguing in good faith, you have an essentially religious opinion and you will stick to it as long as you are able.
> but generation of output is stachostic selection of most common (/highly ranked if tuned) following patterns
This is not an accurate description of the transformer architecture. I’m not surprised that you are misinformed about this.