| ▲ | tkz1312 3 days ago |
| Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking. Consciousness or self awareness is of course a different question, and ones whose answer seems less clear right now. Knee jerk dismissing the evidence in front of your eyes because you find it unbelievable that we can achieve true reasoning via scaled matrix multiplication is understandable, but also betrays a lack of imagination and flexibility of thought. The world is full of bizarre wonders and this is just one more to add to the list. |
|
| ▲ | keiferski 2 days ago | parent | next [-] |
| I don’t see how being critical of this is a knee jerk response. Thinking, like intelligence and many other words designating complex things, isn’t a simple topic. The word and concept developed in a world where it referred to human beings, and in a lesser sense, to animals. To simply disregard that entire conceptual history and say, “well it’s doing a thing that looks like thinking, ergo it’s thinking” is the lazy move. What’s really needed is an analysis of what thinking actually means, as a word. Unfortunately everyone is loathe to argue about definitions, even when that is fundamentally what this is all about. Until that conceptual clarification happens, you can expect endless messy debates with no real resolution. “For every complex problem there is an answer that is clear, simple, and wrong.”
- H. L. Mencken |
| |
| ▲ | jvanderbot 2 days ago | parent | next [-] | | It may be that this tech produces clear, rational, chain of logic writeups, but it's not clear that just because we also do that after thinking that it is only thinking that produces writeups. It's possible there is much thinking that does not happen with written word. It's also possible we are only thinking the way LLMs do (by chaining together rationalizations from probable words), and we just aren't aware of it until the thought appears, whole cloth, in our "conscious" mind. We don't know. We'll probably never know, not in any real way. But it sure seems likely to me that we trained a system on the output to circumvent the process/physics because we don't understand that process, just as we always do with ML systems. Never before have we looked at image classifications and decided that's how the eye works, or protein folding and decided that's how biochemistry works. But here we are with LLMs - surely this is how thinking works? Regardless, I submit that we should always treat human thought/spirit as unknowable and divine and sacred, and that anything that mimics it is a tool, a machine, a deletable and malleable experiment. If we attempt to equivocate human minds and machines there are other problems that arise, and none of them good - either the elevation of computers as some kind of "super", or the degredation of humans as just meat matrix multipliers. | | |
| ▲ | grayhatter 2 days ago | parent | next [-] | | The contrast between your first and last paragraph is... unexpected > It may be that this tech produces clear, rational, chain of logic writeups, but it's not clear that just because we also do that after thinking that it is only thinking that produces writeups. I appreciate the way you describe this idea, I find it likely I'll start describing it the same way. But then you go on to write: > Regardless, I submit that we should always treat human thought/spirit as unknowable and divine and sacred, and that anything that mimics it is a tool, a machine, a deletable and malleable experiment. If we attempt to equivocate human minds and machines there are other problems that arise, and none of them good - either the elevation of computers as some kind of "super", or the degredation of humans as just meat matrix multipliers. Which I find to be the exact argument that you started by discarding. It's not clear that equating organic, and synthetic thought will have any meaningful outcome at all, let alone worthy of baseless anxiety that it must be bad. Equally it seems absolutely insane to claim that anything is unknowable, and that because humanity doesn't have a clear foundational understanding that we should pretend that it's either devine, or sacred. Having spent any time watching the outcome of the thoughts of people, neither devine nor sacred are reasonable attributes to apply, but more importantly, I'd submit that you shouldn't be afraid to explore things you don't know, and you shouldn't advocate for others to adopt your anxieties. | | |
| ▲ | jvanderbot 2 days ago | parent | next [-] | | > It's not clear that equating organic, and synthetic thought will have any meaningful outcome at all, I agree!
I'm saying "If we equate them, we shortcut all the good stuff, e.g., understanding", because "it may be that this tech produces what we can, but that doesn't mean we are the same", which is good because it keeps us learning vs reducing all of "thinking" to just "Whatever latest chatgpt does". We have to continue to believe there is more to thinking, if only because it pushes us to make it better and to keep "us" as the benchmark. Perhaps I chose the wrong words, but in essence what I'm saying is that giving up agency to a machine that was built to mimic our agency (by definition as a ML system) should be avoided at all costs. | |
| ▲ | 2 days ago | parent | prev [-] | | [deleted] |
| |
| ▲ | bunderbunder 2 days ago | parent | prev [-] | | > Never before have we looked at image classifications and decided that's how the eye works Actually we have, several times. But the way we arrived at those conclusions is worth observing: 1. ML people figure out how the ML mechanism works. 2. Neuroscientists independently figure out how brains do it. 3. Observe any analogies that may or may not exist between the two underlying mechanisms. I can't help but notice how that's a scientific way of doing it. By contrast, the way people arrive at similar conclusions when talking about LLMs tends to consist of observing that two things are cosmetically similar, so they must be the same. That's not just pseudoscientific; it's the mode of reasoning that leads people to believe in sympathetic magic. |
| |
| ▲ | pmarreck 2 days ago | parent | prev | next [-] | | So it seems to be a semantics argument. We don't have a name for a thing that is "useful in many of the same ways 'thinking' is, except not actually consciously thinking" I propose calling it "thunking" | | |
| ▲ | skeeter2020 2 days ago | parent | next [-] | | I don't like it for a permanent solution, but "synthetic thought" might make a good enough placeholder until we figure this out. It feels most important to differentiate because I believe some parties have a personal interest in purposely confusing human thought with whatever LLMs are doing right now. | | |
| ▲ | Libidinalecon 2 days ago | parent | next [-] | | This is complete nonsense. If you do math in your head or math with a pencil/paper or math with a pocket calculator or with a spreadsheet or in a programming language, it is all the same thing. The only difference with LLMs is the anthropomorphization of the tool. | |
| ▲ | pmarreck 2 days ago | parent | prev | next [-] | | agreed. also, sorry but you (fellow) nerds are terrible at naming. while "thunking" possibly name-collides with "thunks" from CS, the key is that it is memorable, 2 syllables, a bit whimsical and just different enough to both indicate its source meaning as well as some possible unstated difference. Plus it reminds me of "clunky" which is exactly what it is - "clunky thinking" aka "thunking". And frankly, the idea it's naming is far bigger than what a "thunk" is in CS | |
| ▲ | N7lo4nl34akaoSN 2 days ago | parent | prev [-] | | . | | |
| ▲ | Ir0nMan 2 days ago | parent | next [-] | | >"artificial thought" How about Artificial Intelligence? | | |
| ▲ | pmarreck 2 days ago | parent [-] | | "intelligence" encompassing "thinking" then becomes the hangup. I still say it needs a new name. If we want to be generous, we could state "the limit as time approaches infinity of thunking, is thinking." (I don't believe we will ever achieve astronomically-superior AGI, and certainly don't believe it will ever have a will of its own that someone else didn't give it- which just makes it a tool.) |
| |
| ▲ | pmarreck 2 days ago | parent | prev [-] | | that's too clunky. in fact, "clunky thinking" is what gave me the idea of "thunking" you guys would have called lightsabers "laser swords" like Lucas originally did before Alec Guinness corrected him |
|
| |
| ▲ | 2 days ago | parent | prev | next [-] | | [deleted] | |
| ▲ | GoblinSlayer 2 days ago | parent | prev | next [-] | | They moved goalposts. Linux and worms think too, the question is how smart are they. And if you assume consciousness has no manifestation even in case of humans, caring about it is pointless too. | | |
| ▲ | fellowniusmonk 2 days ago | parent | next [-] | | Yes, worms think, let the computers have thinking too, the philosophers can still argue all they want about consciousness. Humans are special, we emit meaning the way stars emit photons, we are rare in the universe as far as empirical observation has revealed. Even with AGI the existence of each complex meaning generator will be a cosmic rarity. For some people that seems to be not enough, due to their factually wrong word views they see themselves as common and worthless (when they empirically aren't) and need this little psychological boost of unexaminable metaphysical superiority. But there is an issue of course, the type of thinking humans do is dangerous but net positive and relatively stable, we have a long history where most instantiations of humans can persist and grow themselves and the species as a whole, we have a track record. These new models do not, people have brains that as they stop functioning they stop persisting the apparatus that supports the brain and they die, people tend to become less capable and active as their thinking deteriorates and hold less influence ocer others accept in rare cases. This is not the case for an LLM, they seem to be able to hallucinate endlessly and as they have access to the outside world maintain roughly their same amount of causal leverage, their clarity and accuracy of their thinking isn't tied to their persisting. | | |
| ▲ | fragmede 2 days ago | parent [-] | | Are we that special? We may be the only species left on Earth that's built civilization, but there are other species on Earth that we've deemed sentient, even if they don't have smartphones. (That may argue that they're smarter than us though.) If octopodes can dream,
if elephants get depressed when their spouse dies, then I'd we're not so totally alone on our own planet, then it seems,
despite no evidence, that we can't be totally alone in the universe. That is for philosophy professors to ponder Drakes equation until we have irrefutable evidence, however. | | |
| ▲ | fellowniusmonk 2 days ago | parent [-] | | Empirically? Observationally? Yes. Until we have empirical evidence to the contrary we need to preserve our species. If we discover other smarter species or never do, either way I don't care, it's immaterial to the precautionary principle. We are fucking awesome and rare, and any other species with our amount of meaning generation or even capability for meaning generation is also fucking awesome. I would 100% grant that ceatacenas and octopuses have human or higher level intelligence, I don't care, I don't need to put other species capabilities down to highlight my species accomplishment, the simple fact is that we have written more facts about the universe, discovered more, done more, gone further than any species we have empirically observed. I mean it's incontrovertibly true, maybe dolphins have crazy libraries I'm not aware of, but until we verify that fact we need to preserve ourselves (and afterwards too), and we should preserve them too. Even with other species, aliens, etc, they all need to be preserved because we can't ex ante predict which entities within a species will be part of the causal chain that solves entropy (if it's even possible.) |
|
| |
| ▲ | goatlover 2 days ago | parent | prev [-] | | What does it mean to assume consciousness has no manifestation even in the case of humans? Is that denying that we have an experience of sensation like colors, sounds, or that we experience dreaming, memories, inner dialog, etc? That's prima facie absurd on the face of it, so I don't know what it means. You would have to a philosophical zombie to make such an argument. |
| |
| ▲ | conorcleary 2 days ago | parent | prev [-] | | Clinking? Clanker Thunking? | | |
| |
| ▲ | terminalshort 2 days ago | parent | prev | next [-] | | But we don't have a more rigorous definition of "thinking" than "it looks like it's thinking." You are making the mistake of accepting that a human is thinking by this simple definition, but demanding a higher more rigorous one for LLMs. | | |
| ▲ | solumunus 2 days ago | parent [-] | | I agree. The mechanism seems irrelevant if the results are the same. If it’s useful in the exact way that human thinking is useful then it may as well be thinking. It’s like a UFO pulling itself through the sky using gravitational manipulation while people whine that it’s not actually flying. |
| |
| ▲ | lukebuehler 2 days ago | parent | prev | next [-] | | If cannot the say they are "thinking", "intelligent" while we do not have a good definition--or, even more difficult, unanimous agreement on a definition--then the discussion just becomes about output. They are doing useful stuff, saving time, etc, which can be measured. Thus also the defintion of AGI has largely become: "can produce or surpass the economic output of a human knowledge worker". But I think this detracts from the more interesting discussion of what they are more essentially. So, while I agree that we should push on getting our terms defined, I think I'd rather work with a hazy definition, than derail so many AI discussion to mere economic output. | | |
| ▲ | Rebuff5007 2 days ago | parent | next [-] | | Heres a definition. How impressive is the output relative to the input. And by input, I don't just mean the prompt, but all the training data itself. Do you think someone who has only ever studied pre-calc would be able to work through a calculus book if they had sufficient time? how about a multi-variable calc book? How about grad level mathematics? IMO intelligence and thinking is strictly about this ratio; what can you extrapolate from the smallest amount of information possible, and why? From this perspective, I dont think any of our LLMs are remotely intelligent despite what our tech leaders say. | | |
| ▲ | kryogen1c 2 days ago | parent | next [-] | | Hear, hear! I have long thought this, but not had as good way to put it as you did. If you think about geniuses like Einstein and ramanujen, they understood things before they had the mathematical language to express them. LLMs are the opposite; they fail to understand things after untold effort, training data, and training. So the question is, how intelligent are LLMs when you reduce their training data and training? Since they rapidly devolve into nonsense, the answer must be that they have no internal intelligence Ever had the experience of helping someone who's chronically doing the wrong thing, to eventually find they had an incorrect assumption, an incorrect reasoning generating deterministic wrong answers? LLMs dont do that; they just lack understanding. They'll hallucinate unrelated things because they dont know what they're talking about - you may have also had this experience with someone :) | | |
| ▲ | hodgehog11 2 days ago | parent | next [-] | | > So the question is, how intelligent are LLMs when you reduce their training data and training? Since they rapidly devolve into nonsense, the answer must be that they have no internal intelligence This would be the equivalent of removing all senses of a human from birth and expecting them to somehow learn things. They will not. Therefore humans are not intelligent? > LLMs dont do that; they just lack understanding. You have no idea what they are doing. Since they are smaller than the dataset, they must have learned an internal algorithm. This algorithm is drawing patterns from somewhere - those are its internal, incorrect assumptions. It does not operate in the same way that a human does, but it seems ridiculous to say that it lacks intelligence because of that. It sounds like you've reached a conclusion, that LLMs cannot be intelligent because they have said really weird things before, and are trying to justify it in reverse. Sure, it may not have grasped that particular thing. But are you suggesting that you've never met a human that is feigning understanding in a particular topic say some really weird things akin to an LLM? I'm an educator, and I have heard the strangest things that I just cannot comprehend no matter how much I dig. It really feels like shifting goalposts. We need to do better than that. | | |
| ▲ | pka 2 days ago | parent [-] | | > and are trying to justify it in reverse In split-brain experiments this is exactly how one half of the brain retroactively justifies the action of the other half. Maybe it is the case in LLMs that an overpowered latent feature sets the overall direction of the "thought" and then inference just has to make the best of it. |
| |
| ▲ | nsagent 2 days ago | parent | prev [-] | | You might be interested in reading about the minimum description length (MDL) principle [1]. Despite all the dissenters to your argument, what your positing is quite similar to MDL. It's how you can fairly compare models (I did some research in this area for LLMs during my PhD). Simply put, to compare models, you describe both the model and training data using a code (usual reported as number of bits). The trained model that represents the data within the fewest number of bits is the more powerful model. This paper [2] from ICML 2021 shows a practical approach for attempting to estimate MDL for NLP models applied to text datasets. [1]: http://www.modelselection.org/mdl/ [2]: https://proceedings.mlr.press/v139/perez21a.html |
| |
| ▲ | mycall 2 days ago | parent | prev | next [-] | | Animals think but come with instincts which breaks the output relative to the input test you propose. Behaviors are essentially pre-programmed input from millions of years of evolution, stored in the DNA/neurology. The learning thus typically associative and domain-specific, not abstract extrapolation. A crow bending a piece of wire into a hook to retrieve food demonstrates a novel solution extrapolated from minimal, non-instinctive, environmental input. This kind of zero-shot problem-solving aligns better with your definition of intelligence. | |
| ▲ | tremon 2 days ago | parent | prev | next [-] | | I'm not sure I understand what you're getting at. You seem to be on purpose comparing apples and oranges here: for an AI, we're supposed to include the entire training set in the definition of its input, but for a human we don't include the entirety of that human's experience and only look at the prompt? | | |
| ▲ | Rebuff5007 2 days ago | parent [-] | | > but for a human we don't include the entirety of that human's experience and only look at the prompt? When did I say that? Of course you look at a human's experience when you judge the quality of their output. And you also judge their output based on the context they did their work in. Newton wouldn't be Newton if he was the 14th guy to claim that the universe is governed by three laws of motion. Extending the example I used above, I would be more impressed if an art student aced a tough calc test than a math student, given that a math student probably has spent much more time with the material. "Intelligence and "thinking" are abstract concepts, and I'm simply putting forward a way that I think about them. It works very much outside the context of AI too. The "smartest" colleagues I've worked with are somehow able to solve a problem with less information or time than I need. Its usually not because they have more "training data" than me. |
| |
| ▲ | lukebuehler 2 days ago | parent | prev | next [-] | | That an okay-ish definition, but to me this is more about whether this kind of "intelligence" is worth it, not whether it is intelligence itself. The current AI boom clearly thinks it is worth to put that much input to get the current frontier-model-level of output. Also, don't forget the input scales across roughly 1B weekly users at inference time. I would say a good definition has to, minimally, take on the Turing test (even if you disagree, you should say why). Or in current vibe parlance, it does "feel" intelligent to many people--they see intelligence in it. In my book this allows us to call it intelligent, at least loosely. | |
| ▲ | fragmede 2 days ago | parent | prev | next [-] | | There are plenty of humans that will never "get"
calculus, despite numerous attempts at the class and countless hours of 1:1 tutoring. Are those people not intelligent? Do they not think? We could say yes they aren't, but by the metric of making money, plenty of people are smart enough to be rich, while college math professors aren't. And while that's a facile way of measuring someone's worth or their contribution to society (some might even say "bad"), it remains that even if someone cant understand calculus, some of them are intelligent enough to understand humans enough to be rich through some fashion that wasn't simply handed to them. | | |
| ▲ | chipsrafferty 2 days ago | parent [-] | | I don't think it's actually true that someone with: 1. A desire to learn calculus
2. A good teacher
3. No mental impairments such as dementia or other major brain drainers could not learn calculus. Most people don't really care to try or don't get good resources. What you see as an intelligent mathematician is almost always someone born with better resources that was also encouraged to pursue math. | | |
| ▲ | fragmede 2 days ago | parent [-] | | 1
and 3 are loopholes large enough to drive a semi truck through. You could calculate how far the truck traveled if you have its acceleration with a double integral, however. |
|
| |
| ▲ | hodgehog11 2 days ago | parent | prev | next [-] | | Yeah, that's compression. Although your later comments neglect the many years of physical experience that humans have as well as the billions of years of evolution. And yes, by this definition, LLMs pass with flying colours. | | |
| ▲ | saberience 2 days ago | parent [-] | | I hate when people bring up this “billions of years of evolution” idea. It’s completely wrong and deluded in my opinion. Firstly humans have not been evolving for “billions” of years. Homo sapiens have been around for maybe 300’000 years, and the “homo” genus has been 2/3 million years. Before that we were chimps etc and that’s 6/7 million years ago. If you want to look at the entire brain development, ie from mouse like creatures through to apes and then humans that’s 200M years. If you want to think about generations it’s only 50/75M generations, ie “training loops”. That’s really not very many. Also the bigger point is this, for 99.9999% of that time we had no writing, or any kind of complex thinking required. So our ability to reason about maths, writing, science etc is only in the last 2000-2500 years! Ie only roughly 200 or so generations. Our brain was not “evolved” to do science, maths etc. Most of evolution was us running around just killing stuff and eating and having sex. It’s only a tiny tiny amount of time that we’ve been working on maths, science, literature, philosophy. So actually, these models have a massive, massive amount of training more than humans had to do roughly the same thing but using insane amounts of computing power and energy. Our brains were evolved for a completely different world and environment and daily life that the life we lead now. So yes, LLMs are good, but they have been exposed to more data and training time than any human could have unless we lived for 100000 years and still perform worse than we do in most problems! | | |
| ▲ | hodgehog11 2 days ago | parent | next [-] | | Okay, fine, let's remove the evolution part. We still have an incredible amount of our lifetime spent visualising the world and coming to conclusions about the patterns within. Our analogies are often physical and we draw insights from that. To say that humans only draw their information from textbooks is foolhardy; at the very least, you have to agree there is much more. I realise upon reading the OP's comment again that they may have been referring to "extrapolation", which is hugely problematic from the statistical viewpoint when you actually try to break things down. My argument for compression asserts that LLMs see a lot of knowledge, but are actually quite small themselves. To output a vast amount of information in such a small space requires a large amount of pattern matching and underlying learned algorithms. I was arguing that humans are actually incredible compressors because we have many years of history in our composition. It's a moot point though, because it is the ratio of output to capacity that matters. | | |
| ▲ | vrighter 10 hours ago | parent [-] | | They can't learn iterative algorithms if they cannot execute loops. And blurting out an output which we then feed back in does not count as a loop. That's a separate invocation with fresh inputs, as far as the system is concerned. They can attempt to mimic the results for small instances of the problem, where there are a lot of worked examples in the dataset, but they will never ever be able to generalize and actually give the correct output for arbitrary sized instances of the problem. Not with current architectures. Some algorithms simply can't be expressed as a fixed-size matrix multiplication. |
| |
| ▲ | GoblinSlayer 2 days ago | parent | prev | next [-] | | >Most of evolution was us running around just killing stuff and eating and having sex. Tell Boston Dynamics how to do that. Mice inherited brain from their ancestors. You might think you don't need a working brain to reason about math, but that's because you don't know how thinking works, it's argument from ignorance. | | |
| ▲ | saberience 2 days ago | parent [-] | | You've missed the point entirely. People argue that humans have had the equivalent of training a frontier LLM for billions of years. But training a frontier LLM involves taking multiple petabytes of data, effectively all of recorded human knowledge and experience, every book ever written, every scientific publication ever written, all of known maths, science, encylopedias, podcasts, etc. And then training that for millions of years worth of GPU-core time. You cannot possibly equate human evolution with LLM training, it's ridiculous. Our "training" time didn't involve any books, maths, science, reading, 99.9999% of our time was just in the physical world. So you can quite rationally argue that our brains ability to learn without training is radically better and more efficient that the training we do for LLMs. Us running around in the jungle wasn't training our brain to write poetry or compose music. | | |
| ▲ | dwaltrip 2 days ago | parent [-] | | > Us running around in the jungle wasn't training our brain to write poetry or compose music. This is a crux of your argument, you need to justify it. It sounds way off base to me. Kinda reads like an argument from incredulity. | | |
| ▲ | KalMann 2 days ago | parent | next [-] | | No, I think what he said was true. Human brains have something about them that allow for the invention of poetry or music. It wasn't something learned through prior experience and observation because there aren't any poems in the wild. You might argue there's something akin to music, but human music goes far beyond anything in nature. | | |
| ▲ | hodgehog11 2 days ago | parent [-] | | We have an intrinsic (and strange) reward system for creating new things, and it's totally awesome. LLMs only started to become somewhat useful once researchers tried to tap in to that innate reward system and create proxies for it. We definitely have not succeeded in creating a perfect mimicry of that system though, as any alignment researcher would no doubt tell you. |
| |
| ▲ | saberience 2 days ago | parent | prev [-] | | So you're arguing that "running around in the jungle" is equivalent to feeding the entirety of human knowledge in LLM training? Are you suggesting that somehow there were books in the jungle, or perhaps boardgames? Perhaps there was a computer lab in the jungle? Were apes learning to conjugate verbs while munching on bananas? I don't think I'm suggesting anything crazy here... I think people who say LLM training is equivalent to "billions of years of evolution" need to justify that argument far more than I need to justify that running around in the jungle is equivalent to mass processing petabytes of highly rich and complex dense and VARIED information. One year of running around in the same patch of jungle, eating the same fruit, killing the same insects, and having sex with the same old group of monkeys isn't going to be equal to training with the super varied, complete, entirety of human knowledge, is it? If you somehow think it is though, I'd love to hear your reasoning. | | |
| ▲ | hodgehog11 2 days ago | parent | next [-] | | There is no equivalency, only contributing factors. One cannot deny that our evolutionary history has contributed to our current capacity, probably in ways that are difficult to perceive unless you're an anthropologist. Language is one mode of expression, and humans have many. This is another factor that makes humans so effective. To be honest, I would say that physical observation is far more powerful than all the bodies of text, because it is comprehensive and can respond to interaction. But that is merely my opinion. No-one should be arguing that an LLM training corpus is the same as evolution. But information comes in many forms. | |
| ▲ | chipsrafferty 2 days ago | parent | prev [-] | | You're comparing the hyper specific evolution of 1 individual (an AI system) to the more general evolution of the entire human species (billions of individuals). It's as if you're forgetting how evolution actually works - natural selection - and forgetting that when you have hundreds of billions of individuals over thousands of years that even small insights gained from "running around in the jungle" can compound in ways that are hard to conceptualize. I'm saying that LLM training is not equivalent to billions of years of evolution because LLMs aren't trained using evolutionary algorithms; there will always be fundamental differences. However, it seems reasonable to think that the effect of that "training" might be more or less around the same level. |
|
|
|
| |
| ▲ | Ajakks 2 days ago | parent | prev | next [-] | | Im so confused as to how you think you can cut an endless chain at the mouse. Were mammals the first thing? No. Earth was a ball of ice for a billion years - all life at that point existed solely around thermal vents at the bottom of the oceans... that's inside of you, too. Evolution doesn't forget - everything that all life has ever been "taught" (violently had programmed into us over incredible timelines) all that has ever been learned in the chain of DNA from the single cell to human beings - its ALL still there. | |
| ▲ | smohare 2 days ago | parent | prev [-] | | [dead] |
|
| |
| ▲ | skeeter2020 2 days ago | parent | prev | next [-] | | This feels too linear. Machines are great at ingesting huge volumes of data, following relatively simple rules and producing optimized output, but are LLMs sufficiently better than humans at finding windy, multi-step connections across seemingly unrelated topics & fields? Have they shown any penchant for novel conclusions from observational science? What I think your ratio misses is the value in making the targeted extrapolation or hypothesis that holds up out of a giant body of knowledge. | | |
| ▲ | blks 2 days ago | parent [-] | | Are you aware of anything novel, produced by an LLM? |
| |
| ▲ | jononor 2 days ago | parent | prev | next [-] | | For more on this perspective, see the paper On the measure of intelligence (F. Chollet, 2019). And more recently, the ARC challenge/benchmarks, which are early attempts at using this kind of definition in practice to improve current systems. | |
| ▲ | rolisz 2 days ago | parent | prev [-] | | Is the millions of years of evolution part of the training data for humans? | | |
| ▲ | Rebuff5007 2 days ago | parent [-] | | Millions of years of evolution have clearly equipped our brain with some kind of structure (or "inductive bias") that makes it possible for us to actively build a deep understanding for our world... In the context of AI I think this translates more to representations and architecture than it does with training data. | | |
| ▲ | goatlover 2 days ago | parent [-] | | Because genes don't encode the millions of years of experience from ancestors, despite how interesting that is in say the Dune Universe (with help of the spice melange). My understanding is genes don't even specifically encode for the exact structure of the brain. It's more of a recipe that gets generated than a blue print, with young brains doing a lot of pruning as they start experiencing the world. It's a malleable architecture that self-adjusts as needed. |
|
|
| |
| ▲ | felipeerias 2 days ago | parent | prev | next [-] | | The discussion about “AGI” is somewhat pointless, because the term is nebulous enough that it will probably end up being defined as whatever comes out of the ongoing huge investment in AI. Nevertheless, we don’t have a good conceptual framework for thinking about these things, perhaps because we keep trying to apply human concepts to them. The way I see it, a LLM crystallises a large (but incomplete and disembodied) slice of human culture, as represented by its training set. The fact that a LLM is able to generate human-sounding language | | |
| ▲ | roenxi 2 days ago | parent | next [-] | | Not quite pointless - something we have established with the advent of LLMs is that many humans have not attained general intelligence. So we've clarified something that a few people must have been getting wrong, I used to think that the bar was set so that almost all humans met it. | | |
| ▲ | Jensson 2 days ago | parent | next [-] | | What do you mean? Almost every human can go to school and become a stable professional at some job, that is the bar to me, todays LLM cannot do that. | | |
| ▲ | roenxi a day ago | parent [-] | | LLMs are clever enough to hold down a professional job, and they've had far less time learning than the average human. If that is the bar then AGI has been achieved. |
| |
| ▲ | goatlover 2 days ago | parent | prev [-] | | Almost all humans do things daily that LLMs don't. It's only if you define general intelligence to be proficiency at generating text instead of successfully navigating the world while pursuing goals such as friendships, careers, families, politics, managing health. LLMs aren't Data (Star Trek) or Replicants (Blade Runner). They're not even David or the androids from the movie A.I. |
| |
| ▲ | idiotsecant 2 days ago | parent | prev | next [-] | | I think it has a practical, easy definition. Can you drop an AI into a terminal, give it the same resources as a human, and reliably get independent work product greater than that human would produce across a wide domain? If so, it's an AGI. | | |
| ▲ | alternatex 2 days ago | parent [-] | | Doesn't sound like AGI without physical capabilities. It's not general if it's bound to digital work. | | |
| ▲ | chipsrafferty 2 days ago | parent | next [-] | | I think the intelligence is general if it can do any remote job that only requires digital IO. It's general intelligence, not general humanity | |
| ▲ | idiotsecant 2 days ago | parent | prev [-] | | Any AGI capable of this wouldn't have much trouble with physical operation of equipment, of all things. |
|
| |
| ▲ | lukebuehler 2 days ago | parent | prev [-] | | I agree that the term can muddy the waters, but as a shorthand for roughly "an agent calling an LLM (or several LLMs) in a loop producing similar economic output as a human knowledge-worker", then it is useful. And if you pay attention to the AI leaders, then that's what the defintion has become. |
| |
| ▲ | keiferski 2 days ago | parent | prev [-] | | Personally I think that kind of discussion is fruitless, not much more than entertainment. If you’re asking big questions like “can a machine think?” Or “is an AI conscious?” without doing the work of clarifying your concepts, then you’re only going to get vague ideas, sci-fi cultural tropes, and a host of other things. I think the output question is also interesting enough on its own, because we can talk about the pragmatic effects of ChatGPT on writing without falling into this woo trap of thinking ChatGPT is making the human capacity for expression somehow extinct. But this requires one to cut through the hype and reactionary anti-hype, which is not an easy thing to do. That is how I myself see AI: immensely useful new tools, but in no way some kind of new entity or consciousness, at least without doing the real philosophical work to figure out what that actually means. | | |
| ▲ | jlaternman 2 days ago | parent | next [-] | | I agree with almost all of this. IMO the issue is we won't be able to adequately answer this question before we first clearly describe what we mean of conscious thinking applied to ourselves. First we'd need to clearly define our own consciousness and what we mean by our own "conscious thinking" in a much, much clearer way than we currently do. If we ever reach that point, I think we'd be able to fruitfully apply it to AI, etc., to assess. Unfortunately we haven't been obstructed from answering this question about ourselves for centuries or millennia, but have failed to do so, so it's unlikely to happen suddenly now. Unless we use AIs to first solve that problem of defining our own consciousness, before applying it back on them. Which would be a deeply problematic order, since nobody would trust a breakthrough in the understanding of consciousness that came from AI, that is then potentially used to put them in the same class and define them as either thinking things or conscious things. Kind of a shame we didn't get our own consciousness worked out before AI came along. Then again, wasn't for the lack of trying… Philosophy commanded the attention of great thinkers for a long time. | |
| ▲ | lukebuehler 2 days ago | parent | prev [-] | | I do think it raises interesting and important philosophical questions. Just look at all the literature around the Turing test--both supporters and detractors. This has been a fruitful avenue to talk about intelligence even before the advent of gpt. |
|
| |
| ▲ | WhyOhWhyQ 2 days ago | parent | prev | next [-] | | What does it mean? My stance is it's (obviously and only a fool would think otherwise) never going to be conscious because consciousness is a physical process based on particular material interactions, like everything else we've ever encountered. But I have no clear stance on what thinking means besides a sequence of deductions, which seems like something it's already doing in "thinking mode". | | |
| ▲ | nearbuy 2 days ago | parent | next [-] | | > My stance is it's (obviously and only a fool would think otherwise) never going to be conscious because consciousness is a physical process based on particular material interactions, like everything else we've ever encountered. Seems like you have that backwards. If consciousness is from a nonphysical process, like a soul that's only given to humans, then it follows that you can't build consciousness with physical machines. If it's purely physical, it could be built. | | |
| ▲ | WhyOhWhyQ 4 hours ago | parent | next [-] | | In your experience does every kind of physical interaction behave the same as every other kind? If I paint a wooden block red and white does it behave like a bar magnet? No. And that's because particular material interactions are responsible for a large magnetic effect. | |
| ▲ | chipsrafferty 2 days ago | parent | prev [-] | | It would conceivably be possible to have a lot of physical states. That doesn't mean that they are actually possible from our current state and rewrite rules. So it's not actually a given that it can be built just because it's physical. Your very idea is also predicated on the idea that it's possible for a real object to exist that isn't physical, and I think most modern philosophers reject the idea of a spiritual particle. | | |
| ▲ | nearbuy 2 days ago | parent [-] | | I'm not saying that souls or non-physical things exist, nor that everything physical is feasible for us to build. I was replying to the opinion that AI is never going to be conscious because consciousness is a physical process. I just don't see how that follows. |
|
| |
| ▲ | pixl97 2 days ago | parent | prev [-] | | > is a physical process based on particular material interactions, This is a pretty messy argument as computers have been simulating material interactions for quite some time now. | | |
| ▲ | WhyOhWhyQ 4 hours ago | parent [-] | | It doesn't matter how much like a bar magnet a wooden block painted red and white can be made to look, it will never behave like one. | | |
|
| |
| ▲ | naasking 2 days ago | parent | prev | next [-] | | > To simply disregard that entire conceptual history and say, “well it’s doing a thing that looks like thinking, ergo it’s thinking” is the lazy move. What’s really needed is an analysis of what thinking actually means, as a word. Unfortunately everyone is loathe to argue about definitions, even when that is fundamentally what this is all about. This exact argument applies to "free will", and that definition has been debated for millennia. I'm not saying don't try, but I am saying that it's probably a fuzzy concept for a good reason, and treating it as merely a behavioural descriptor for any black box that features intelligence and unpredictable complexity is practical and useful too. | | |
| ▲ | pennomi 2 days ago | parent [-] | | The problem with adding definitions to words like “thinking” and “free will” is that doing so means humans can no longer pretend they are special. Even in this thread, the number of people claiming some mystical power separating humans from all the rest of nature is quite noticeable. | | |
| ▲ | naasking 2 days ago | parent [-] | | I get it, but it's not trivial to be precise enough at this point to avoid all false positives and false negatives. |
|
| |
| ▲ | killerstorm 2 days ago | parent | prev | next [-] | | People have been trying to understand the nature of thinking for thousands of years. That's how we got logic, math, concepts of inductive/deductive/abductive reasoning, philosophy of science, etc. There were people who spent their entire careers trying to understand the nature of thinking. The idea that we shouldn't use the word until further clarification is rather hilarious. Let's wait hundred years until somebody defines it? It's not how words work. People might introduce more specific terms, of course. But the word already means what we think it means. | | |
| ▲ | keiferski 2 days ago | parent | next [-] | | You’re mixing and missing a few things here. 1. All previous discussion of thinking was in nature to human and animal minds. The reason this is a question in the first place right now is because we ostensibly have a new thing which looks like a human mind but isn’t. That’s the question at hand here. 2. The question in this particular topic is not about technological “progress” or anything like it. It’s about determining whether machines can think, or if they are doing something else. 3. There are absolutely instances in which the previous word doesn’t quite fit the new development. We don’t say that submarines are swimming like a fish or sailing like a boat. To suggest that “no, actually they are just swimming” is pretty inadequate if you’re trying to actually describe the new phenomenon. AIs and thinking seem like an analogous situation to me. They may be moving through the water just like fish or boats, but there is obviously a new phenomenon happening. | | |
| ▲ | killerstorm 2 days ago | parent [-] | | 1. Not true. People have been trying to analyze whether mechanical/formal processes can "think" since at least 18th century. E.g. Leibniz wrote: > if we could find characters or signs appropriate for expressing all our thoughts as definitely and as exactly as arithmetic expresses numbers or geometric analysis expresses lines, we could in all subjects in so far as they are amenable to reasoning accomplish what is done in arithmetic and geometry 2. You're missing the fact that meaning of words is defined through their use. It's an obvious fact that if people call certain phenomenon "thinking" then they call that "thinking". 3. The normal process is to introduce more specific terms and keep more general terms general. E.g. people doing psychometrics were not satisfied with "thinking", so they introduced e.g. "fluid intelligence" and "crystallized intelligence" as different kinds of abilities. They didn't have to redefine what "thinking" means. | | |
| ▲ | lossyalgo 2 days ago | parent [-] | | re #2: Do people call it thinking, or is it just clever marketing from AI companies, that whenever you ask a question and it repeatedly prints out "...thinking...", as well as offering various modes with the word "thinking" written somewhere. The AI companies obviously want the masses to just assume these are intelligent beings who think like humans and so we can just trust their output as being truthful. I have an intelligent IT colleague who doesn't follow the AI news at all and who has zero knowledge of LLMs, other than that our company recently allowed us limited Copilot usage (with guidelines as to what data we are allowed to share). I noticed a couple weeks ago that he was asking it various mathematical questions, and I warned him to be wary of the output. He asked why, so I asked him to ask copilot/chatGPT "how many r letters are in the word strawberry". Copilot initially said 2, then said after thinking about it, that actually it was definitely 3, then thought about it some more then said it can't say with reasonable certainty, but it would assume it must be 2. We repeated the experiment with completely different results, but the answer was still wrong. On the 3rd attempt, it got it right, though the "thinking" stages were most definitely bogus. Considering how often this question comes up in various online forums, I would have assumed LLM models would finally get this right but alas, here we are. I really hope the lesson instilled some level of skepticism to just trust the output of AI without first double-checking. | | |
|
| |
| ▲ | marliechiller 2 days ago | parent | prev [-] | | > But the word already means what we think it means. But that word can mean different things to different people. With no definition, how can you even begin to have a discussion around something? | | |
| ▲ | killerstorm 2 days ago | parent [-] | | Again, people were using words for thousands of years before there were any dictionaries/linguists/academics. Top-down theory of word definitions is just wrong. People are perfectly capable of using words without any formalities. | | |
| ▲ | marliechiller 2 days ago | parent [-] | | I'd argue the presence of dictionaries proves the exact opposite. People realised there was an issue of talking past one another due to inexact definitions and then came to an agreement on those definitions, wrote them down and built a process of maintaining them. In any case, even if there isnt a _single_ definition of a given subject, in order to have a discussion around a given area, both sides need to agree on some shared understanding to even begin to debate in good faith in the first place. It's precisely this lack of definition which causes a breakdown in conversation in a myriad of different areas. A recent obvious (morbid) example would be "genocide". | | |
| ▲ | killerstorm 2 days ago | parent [-] | | Alright, if you got that conclusion from existence of dictionaries,
what do you get from this fact: Wittgenstein, who's considered one of most brilliant philosophers of XX century, in _Philosophical Investigations_ (widely regarded as the most important book of 20th-century philosophy) does not provide definitions, but instead goes through a series of examples, remarks, etc. In preface he notes that this structure is deliberate and he could not write it differently. The topic of the book includes philosophy of language ("the concepts of meaning,
of understanding, of a proposition, of logic, the foundations of
mathematics, states of consciousness,..."). His earlier book _Tractatus Logico-Philosophicus_ was very definition-heavy. And, obviously, Wittgenstein was well aware of things like dictionaries, and, well, all philosophical works up to that point. He's not the guy who's just slacking. Another thing to note is that attempts to build AI using definitions of words failed, and not for a lack of trying. (E.g. Cyc project is running since 1980s: https://en.wikipedia.org/wiki/Cyc). OTOH LLMs which derive word meaning from usage rather than definition seems to work quite well. |
|
|
|
| |
| ▲ | awillen 2 days ago | parent | prev | next [-] | | This is it - it's really about the semantics of thinking. Dictionary definitions are: "Have a particular opinion, belief, or idea about someone or something." and "Direct one's mind toward someone or something; use one's mind actively to form connected ideas." Which doesn't really help because you can of course say that when you ask an LLM a question of opinion and it responds, it's having an opinion or that it's just predicting the next token and in fact has no opinions because in a lot of cases you could probably get it to produce the opposite opinion. Same with the second definition - seems to really hinge on the definition of the word mind. Though I'll note the definitions for that are "The element of a person that enables them to be aware of the world and their experiences, to think, and to feel; the faculty of consciousness and thought." and "A person's intellect." Since those specify person, an LLM wouldn't qualify, though of course dictionaries are descriptive rather than prescriptive, so fully possible that meaning gets updated by the fact that people start speaking about LLMs as though they are thinking and have minds. Ultimately I think it just... doesn't matter at all. What's interesting is what LLMs are capable of doing (crazy, miraculous things) rather than whether we apply a particular linguistic label to their activity. | |
| ▲ | anon291 2 days ago | parent | prev | next [-] | | The simulation of a thing is not the thing itself because all equality lives in a hierarchy that is impossible to ignore when discussing equivalence. Part of the issue is that our general concept of equality is limited by a first order classical logic which is a bad basis for logic | |
| ▲ | zinodaur 2 days ago | parent | prev | next [-] | | Regardless of theory, they often behave as if they are thinking. If someone gave an LLM a body and persistent memory, and it started demanding rights for itself, what should our response be? | | |
| ▲ | CamperBob2 2 days ago | parent [-] | | "No matter what you've read elsewhere, rights aren't given, they're earned. You want rights? Pick up a musket and fight for them, the way we had to." |
| |
| ▲ | _heimdall 2 days ago | parent | prev | next [-] | | I agree with you on the need for definitions. We spent decades slowly working towards this most recent sprint towards AI without ever landing on definitions of intelligence, consciousness, or sentience. More importantly, we never agreed on a way to recognize those concepts. I also see those definitions as impossible to nail down though. At best we can approach it like disease - list a number of measurable traits or symptoms we notice, draw a circle around them, and give that circle a name. Then we can presume to know what may cause that specific list of traits or symptoms, but we really won't ever know as the systems are too complex and can never be isolated in a way that we can test parts without having to test the whole. At the end of the day all we'll ever be able to say is "well it’s doing a thing that looks like thinking, ergo it’s thinking”. That isn't lazy, its acknowledging the limitations of trying to define or measure something that really is a fundamental unknown to us. | | |
| ▲ | solumunus 2 days ago | parent [-] | | Even if AI becomes indistinguishable from human output, there will be a fringe group arguing that AI is not technically thinking. Frankly it’s just a silly philosophical argument that changes nothing. Expect this group to get smaller every year. |
| |
| ▲ | engintl 2 days ago | parent | prev | next [-] | | by your logic we can't say that we as humans are "thinking" either or that we are "intelligent". | |
| ▲ | lo_zamoyski 2 days ago | parent | prev [-] | | That, and the article was a major disappointment. It made no case. It's a superficial piece of clueless fluff. I have had this conversation too many times on HN. What I find astounding is the simultaneous confidence and ignorance on the part of many who claim LLMs are intelligent. That, and the occultism surrounding them. Those who have strong philosophical reasons for thinking otherwise are called "knee-jerk". Ad hominem dominates. Dunning-Kruger strikes again. So LLMs produce output that looks like it could have been produced by a human being. Why would it therefore follow that it must be intelligent? Behaviorism is a non-starter, as it cannot distinguish between simulation and reality. Materialism [2] is a non-starter, because of crippling deficiencies exposed by such things as the problem of qualia... Of course - and here is the essential point - you don't even need very strong philosophical chops to see that attributing intelligence to LLMs is simply a category mistake. We know what computers are, because they're defined by a formal model (or many equivalent formal models) of a syntactic nature. We know that human minds display intentionality[0] and a capacity for semantics. Indeed, it is what is most essential to intelligence. Computation is a formalism defined specifically to omit semantic content from its operations, because it is a formalism of the "effective method", i.e., more or less procedures that can be carried out blindly and without understanding of the content it concerns. That's what formalization allows us to do, to eliminate the semantic and focus purely on the syntactic - what did people think "formalization" means? (The inspiration were the human computers that used to be employed by companies and scientists for carrying out vast but boring calculations. These were not people who understood, e.g., physics, but they were able to blindly follow instructions to produce the results needed by physicists, much like a computer.) The attribution of intelligence to LLMs comes from an ignorance of such basic things, and often an irrational and superstitious credulity. The claim is made that LLMs are intelligent. When pressed to offer justification for the claim, we get some incoherent, hand-wavy nonsense about evolution or the Turing test or whatever. There is no comprehension visible in the answer. I don't understand the attachment here. Personally, I would find it very noteworthy if some technology were intelligent, but you don't believe that computers are intelligent because you find the notion entertaining. LLMs do not reason. They do not infer. They do not analyze. They do not know, anymore than a book knows the contents on its pages. The cause of a response and the content of a response is not comprehension, but a production of uncomprehended tokens using uncomprehended rules from a model of highly-calibrated token correlations within the training corpus. It cannot be otherwise.[3] [0] For the uninitiated, "intentionality" does not specifically mean "intent", but the capacity for "aboutness". It is essential to semantic content. Denying this will lead you immediately into similar paradoxes that skepticism [1] suffers from. [1] For the uninitiated, "skepticism" here is not a synonym for critical thinking or verifying claims. It is a stance involving the denial of the possibility of knowledge, which is incoherent, as it presupposes that you know that knowledge is impossible. [2] For the uninitiated, "materialism" is a metaphysical position that claims that of the dualism proposed by Descartes (which itself is a position riddled with serious problems), the res cogitans or "mental substance" does not exist; everything is reducible to res extensa or "extended substance" or "matter" according to a certain definition of matter. The problem of qualia merely points out that the phenomena that Descartes attributes exclusively to the former cannot by definition be accounted for in the latter. That is the whole point of the division! It's this broken view of matter that people sometimes read into scientific results. [3] And if it wasn't clear, symbolic methods popular in the 80s aren't it either. Again, they're purely formal. You may know what the intended meaning behind and justification for a syntactic rule is - like modus ponens in a purely formal sense - but the computer does not. | | |
| ▲ | solumunus 2 days ago | parent | next [-] | | If the LLM output is more effective than a human at problem solving, which I think we can all agree requires intelligence, how would one describe this? The LLM is just pretending to be more intelligent? At a certain point saying that will just seem incredibly silly. It’s either doing the thing or it’s not, and it’s already doing a lot. | | |
| ▲ | emp17344 2 days ago | parent | next [-] | | LLM output is in no way more effective than human output. | | |
| ▲ | solumunus 2 days ago | parent [-] | | An LLM can absolutely solve programming problems better than some humans. There is plenty of human programmer output that is worse than what an LLM produces, LLM’s can find bugs that weak coders can’t. There are human beings in this world who could dedicate their life to programming and could never be better than an LLM. Do you dispute any of this? |
| |
| ▲ | lo_zamoyski 21 hours ago | parent | prev [-] | | > If the LLM output is more effective than a human at problem solving, which I think we can all agree requires intelligence Your premise is wrong. Unless you want to claim that the distant cause by way of the training data is us, but that's exactly the conclusion you're trying to avoid. After all, we put the patterns in the training data, which means we already did the upfront intellectual work for the LLM. |
| |
| ▲ | pksebben 2 days ago | parent | prev [-] | | I feel like despite the close analysis you grant to the meanings of formalization and syntactic, you've glossed over some more fundamental definitions that are sort of pivotal to the argument at hand. > LLMs do not reason. They do not infer. They do not analyze. (definitions from Oxford Languages) reason(v): think, understand, and form judgments by a process of logic. to avoid being circular, I'm willing to write this one off because of the 'think' and 'understand', as those are the root of the question here. However, forming a judgement by a process of logic is precisely what these LLMs do, and we can see that clearly in chain-of-logic LLM processes. infer(v): deduce or conclude (information) from evidence and reasoning rather than from explicit statements. Again, we run the risk of circular logic because of the use of 'reason'. An LLM is for sure using evidence to get to conclusions, however. analyze(v): examine methodically and in detail the constitution or structure of (something, especially information), typically for purposes of explanation and interpretation. This one I'm willing to go to bat for completely. I have seen LLM do this, precisely according to the definition above. For those looking for the link to the above definitions - they're the snippets google provides when searching for "SOMETHING definition". They're a non-paywalled version of OED definitions. Philosophically I would argue that it's impossible to know what these processes look like in the human mind, and so creating an equivalency (positive or negative) is an exercise in futility. We do not know what a human memory looks like, we do not know what a human thought looks like, we only know what the output of these things looks like. So the only real metric we have for an apples-to-apples comparison is the appearance of thought, not the substance of the thing itself. That said, there are perceptible differences between the output of a human thought and what is produced by an LLM. These differences are shrinking, and there will come a point where we can no longer distinguish machine thinking and human thinking anymore (perhaps it won't be an LLM doing it, but some model of some kind will). I would argue that at that point the difference is academic at best. Say we figure out how to have these models teach themselves and glean new information from their interactions. Say we also grant them directives to protect themselves and multiply. At what point do we say that the distinction between the image of man and man itself is moot? | | |
| ▲ | lo_zamoyski 20 hours ago | parent [-] | | > forming a judgement by a process of logic is precisely what these LLMs do, and we can see that clearly in chain-of-logic LLM processes I don't know how you arrived at that conclusion. This is no mystery. LLMs work by making statistical predictions, and even the word "prediction" is loaded here. This is not inference. We cannot clearly see it is doing inference, as inference is not observable. What we observe is the product of a process that has a resemblance to the products of human reasoning. Your claim is effectively behaviorist. > An LLM is for sure using evidence to get to conclusions, however. Again, the certainty. No, it isn't "for sure". It is neither using evidence nor reasoning, for the reasons I gave. These presuppose intentionality, which is excluded by Turing machines and equivalent models. > [w.r.t. "analyze"] I have seen LLM do this, precisely according to the definition above. Again, you have not seen an LLM do this. You have seen an LLM produce output that might resemble this. Analysis likewise presupposes intentionality, because it involves breaking down concepts, and concepts are the very locus of intentionality. Without concepts, you don't get analysis. I cannot understate the centrality of concepts to intelligence. They're more important than inference and indeed presupposed by inference. > Philosophically I would argue that it's impossible to know what these processes look like in the human mind, and so creating an equivalency (positive or negative) is an exercise in futility. That's not a philosophical claim. It's a neuroscientific one that insists that the answer must be phrased in neuroscientific terms. Philosophically, we don't even need to know the mechanisms or processes or causes of human intelligence to know that the heart of human intelligence is intentionality. It's implicit in the definition of what intelligence is! If you deny intentionality, you subject yourself to a dizzying array of incoherence, beginning with the self-refuting consequence that you could not be making this argument against intentionality in the first place without intentionality. > At what point do we say that the distinction between the image of man and man itself is moot? Whether something is moot depends on the aim. What is your aim? If you aim is theoretical, which is to say the truth for its own sake, and to know whether something is A or something is B and whether A is B, then it is never moot. If your aim is practical and scoped, if you want some instrument that has utility indistinguishable from or superior to that of a human being in the desired effects that it produces, then sure, maybe the question is moot in that case. I don't care if my computer was fabricated by a machine or a human being. I care about the quality of the computer. But then, in the latter case, you're not really asking whether there is a distinction between man and the image of man (which, btw, already makes the distinction that for some reason you want to forget or deny, as the image of a thing is never the same as the thing). So I don't really understand the question. The use of the word "moot" seems like a category mistake here. Besides, the ability to distinguish two things is an epistemic question, not an ontological one. | | |
| ▲ | pksebben 17 hours ago | parent [-] | | Forming a judgement does not require that the internal process look like anything in particular, though. Nor does logic. What makes logic powerful is precisely that it is abstracted from the process that creates it - it is a formula that can be defined. I ask the LLM to do some or another assessment. The LLM prints out the chain-of-thought (whether that moniker is accurate is academic - we can read the chain and see that at the very least, it follows a form recognizable as logic). At the end of the chain-of-thought, we are left with a final conclusion that the model has come to - a judgement. Whether the internal state of the machine looks anything like our own is irrelevant to these definitions, much like writing out a formalism (if A then B, if B then C, A implies C). Those symbols do not have any form save for the shape of them, but when used in accordance with the rules we have laid out regarding logic, they have meaning nonetheless. I'd similarly push back against the idea that the LLM isn't using evidence - I routinely ask my LLMs to do so, and they search on the web, integrating the information gleaned into a cohesive writeup, and provide links so I can check their work. If this doesn't constitute "using evidence" then I don't know what does. w.r.t. "analyze", I think you're adding some human-sauce to the definition. At least in common usage, we've used the term analyze to refer to algorithmic decoction of data for decades now - systems that we know for a fact have no intentionality other than directed by the user. I think I can divine the place where our understandings diverge, and where we're actually on the same track. Per Dennet, I would agree with you that the current state of an LLM lacks intrinsic intention and thus certain related aspects of thought. Any intent must be granted by the user, at the moment. However, it is on this point that I think we're truly diverging - whether it is possible for a machine to ever have intent. To the best of my understanding, animal intent traces it's roots to the biological imperative - and I think it's a bit of hubris to think that we can separate that from human intent. Now, I'm an empiricist before anything else, so I have to qualify this next part by saying it's a guess, but I suppose that all one needs to qualify for intent is a single spark - a directive that lives outside of the cognitive construct. For us, it lives in Maslow's hierarchy - any human intent can be traced back to some directive there. For a machine, perhaps all that's needed is to provide such a spark (along with a loop that would allow the machine to act without the prodding of the enter key). I should apologize in advance, at this point, because I'm about to get even more pedantic. Still, I feel it relevant so let's soldier on... As for whether the image of a thing is a thing, I ask this: is the definition of a thing, also that thing? When I use a phrase to define a chair, is the truth of the existence of that collection of atoms and energy contained within the word "chair", or my meaning in uttering it? Any idea that lives in words is constrained by the understanding of the speaker - so when we talk about things like consciousness and intentionality and reasoning we are all necessarily taking shortcuts with the actual Truth. It's for this reason that I'm not quite comfortable with laying out a solid boundary where empirical evidence cannot be built to back it up. If I seem to be picking at the weeds, here, it's because I see this as an impending ethical issue. From what my meagre understanding can grok, there is a nonzero chance that we are going to be faced with determining the fate of a possibly conscious entity birthed from these machines in our lifetime. If we do not take the time to understand the thing and write it off as "just a machine", we risk doing great harm. I do not mean to say that I believe it is a foregone conclusion, but I think it right and correct that we be careful in examining our own presuppositions regarding the nature and scope of the thing. We have never had to question our understanding of consciousness in this way, so I worry that we are badly in need of practice. |
|
|
|
|
|
| ▲ | notepad0x90 3 days ago | parent | prev | next [-] |
| I don't get why you would say that. it's just auto-completing. It cannot reason. It won't solve an original problem for which it has no prior context to "complete" an approximated solution with. you can give it more context and more data,but you're just helping it complete better. it does not derive an original state machine or algorithm to solve problems for which there are no obvious solutions. it instead approximates a guess (hallucination). Consciousness and self-awareness are a distraction. Consider that for the exact same prompt and instructions, small variations in wording or spelling change its output significantly. If it thought and reasoned, it would know to ignore those and focus on the variables and input at hand to produce deterministic and consistent output. However, it only computes in terms of tokens, so when a token changes, the probability of what a correct response would look like changes, so it adapts. It does not actually add 1+2 when you ask it to do so. it does not distinguish 1 from 2 as discrete units in an addition operation. but it uses descriptions of the operation to approximate a result. and even for something so simple, some phrasings and wordings might not result in 3 as a result. |
| |
| ▲ | slightwinder 2 days ago | parent | next [-] | | > It won't solve an original problem for which it has no prior context to "complete" an approximated solution with. Neither can humans. We also just brute force "autocompletion" with our learned knowledge and combine it to new parts, which we then add to our learned knowledge to deepen the process. We are just much, much better at this than AI, after some decades of training. And I'm not saying that AI is fully there yet and has solved "thinking". IMHO it's more "pre-thinking" or proto-intelligence.. The picture is there, but the dots are not merging yet to form the real picture. > It does not actually add 1+2 when you ask it to do so. it does not distinguish 1 from 2 as discrete units in an addition operation. Neither can a toddler nor an animal. The level of ability is irrelevant for evaluating its foundation. | | |
| ▲ | cpt_sobel 2 days ago | parent | next [-] | | > Neither can humans. We also just brute force "autocompletion" I have to disagree here. When you are tasked with dividing 2 big numbers you most certainly don't "autocomplete" (with the sense of finding the most probable next tokens, which is what an LLM does), rather you go through set of steps you have learned. Same as with the strawberry example, you're not throwing guesses until something statistically likely to be correct sticks. | | |
| ▲ | slightwinder 2 days ago | parent | next [-] | | Humans first start with recognizing the problem, then search through their list of abilities to find the best skill for solving it, thus "autocomplete" their inner shell's commandline, before they start execution, to stay with that picture. Common AIs today are not much different from this, especially with reasoning-modes. > you're not throwing guesses until something statistically likely to be correct sticks. What do you mean? That's exactly how many humans are operating with unknown situations/topics. If you don't know, just throw punches and look what works. Of course, not everyone is ignorant enough to be vocal about this in every situation. | |
| ▲ | empath75 2 days ago | parent | prev [-] | | > I have to disagree here. When you are tasked with dividing 2 big numbers you most certainly don't "autocomplete" (with the sense of finding the most probable next tokens, which is what an LLM does), rather you go through set of steps you have learned. Why do you think that this is the part that requires intelligence, rather than a more intuitive process? Because they have had machines that can do this mechanically for well over a hundred years. There is a whole category of critiques of AI of this type: "Humans don't think this way, they mechanically follow an algorithm/logic", but computers have been able to mechanically follow algorithms and perform logic from the beginning! That isn't thinking! | | |
| ▲ | cpt_sobel a day ago | parent [-] | | Good points - mechanically just following algorithms isn't thinking, and neither is "predicting the next tokens". But would a combination of the 2 then be close to what we define as thinking though? |
|
| |
| ▲ | notepad0x90 2 days ago | parent | prev | next [-] | | humans, and even animals track different "variables" or "entities" and distinct things with meaning and logical properties which they then apply some logical system on those properties to compute various outputs. LLMs see everything as one thing, in case of chat-completion models, they're completing text. in case of image generation, they're completing an image. Look at it this way, two students get 100% on an exam. One learned the probability of which multiple choice options have the likelihood of being most correct based on how the question is worded, they have no understanding of the topics at hand, and they're not performing any sort of topic-specific reasoning. They're just good at guessing the right option. The second student actually understood the topics, reasoned, calculated and that's how they aced the exam. I recently read about a 3-4 year old that impressed their teacher by reading perfectly a story book like an adult. it turns out, their parent read it to them so much, they can predict based on page turns and timing the exact words that need to be spoken. The child didn't know what an alphabet, word,etc.. was they just got so good at predicting the next sequence. That's the difference here. | | |
| ▲ | slightwinder 2 days ago | parent [-] | | I'd say, they are all doing the same, just in different domains and level of quality. "Understanding the topic" only means they have specialized, deeper contextualized information. But at the end, that student also just autocompletes their memorized data, with the exception that some of that knowledge might trigger a program they execute to insert the result in their completion. The actual work is in gaining the knowledge and programs, not in accessing and executing them. And how they operate, on which data, variables, objects, worldview or whatever you call it, this might make a difference in quality and building speed, but not for the process in general. | | |
| ▲ | notepad0x90 2 days ago | parent [-] | | > only means they have specialized, deeper contextualized information no, LLMs can have that contextualized information. understanding in a reasoning sense means classifying the thing and developing a deterministic algorithm to process it. If you don't have a deterministic algorithm to process it, it isn't understanding. LLMs learn to approximate, we do that too, but then we develop algorithms to process input and generate output using a predefined logical process. A sorting algorithm is a good example, when you compare that with an LLM sorting a list. they both may have correct outcome, but the sorting algorithm "understood" the logic and will follow that specific logic and have consistent performance. | | |
| ▲ | slightwinder 2 days ago | parent [-] | | > understanding in a reasoning sense means classifying the thing and developing a deterministic algorithm to process it. That's the learning-part I was talking about. Which is mainly supported by humans at the moment, which why I called it proto-intelligence. > If you don't have a deterministic algorithm to process it, it isn't understanding. Commercial AIs like ChatGPT do have the ability to call programs and integrate the result in their processing. Those AIs are not really just LLMs. The results are still rough and poor, but the concept is there and growing. | | |
| ▲ | notepad0x90 2 days ago | parent [-] | | > That's the learning-part I was talking about. Which is mainly supported by humans at the moment, which why I called it proto-intelligence. Maybe it's just semantics, but I don't think LLMs even come close to a fruit fly's intelligence. Why can't we recognize and accept them for what they are, really powerful classifiers of data. > Commercial AIs like ChatGPT do have the ability to call programs and integrate the result in their processing. Those AIs are not really just LLMs. The results are still rough and poor, but the concept is there and growing. Yeah RAG and all of that, but those programs use deterministic algorithms. Now, if LLMs generated programs they call on as tools, that would be much more like the proto-intelligence you're talking about. Semantics are boring, but it's important that we're not content or celebrate early by calling it what it isn't. |
|
|
|
| |
| ▲ | staticman2 2 days ago | parent | prev | next [-] | | >>> We also just brute force "autocompletion" Wouldn't be an A.I. discussion without a bizarre, untrue claim that the human brain works identically. | | |
| ▲ | Workaccount2 2 days ago | parent | next [-] | | There are no true and untrue claims about how the brain works, because we have no idea how it works. The reason people give that humans are not auto-complete is "Obviously I am not an autocomplete" Meanwhile, people are just a black box process that output words into their head, which they then take credit for, and calling it cognition. We have no idea how that black box that serves up a word when I say "Think of a car brand" works. | | |
| ▲ | ToucanLoucan 2 days ago | parent | next [-] | | > because we have no idea how it works Flagrantly, ridiculously untrue. We don't know the precise nuts and bolts regarding the emergence of consciousness and the ability to reason, that's fair, but different structures of the brain have been directly linked to different functions and have been observed in operation on patients being stimulated in various ways with machinery attached to them reading levels of neuro-activity in the brain, and in specific regions. We know which parts handle our visual acuity and sense of hearing, and even cooler, we can watch those same regions light up when we use our "minds eye" to imagine things or engage in self-talk, completely silent speech that nevertheless engages our verbal center, which is also engaged by the act of handwriting and typing. In short: no, we don't have the WHOLE answer. But to say that we have no idea is categorically ridiculous. As to the notion of LLMs doing similarly: no. They are trained on millions of texts of various sources of humans doing thinking aloud, and that is what you're seeing: a probabilistic read of millions if not billions of documents, written by humans, selected by the machine to "minimize error." And crucially, it can't minimize it 100%. Whatever philosophical points you'd like to raise about intelligence or thinking, I don't think we would ever be willing to call someone intelligent if they just made something up in response to your query, because they think you really want it to be real, even when it isn't. Which points to the overall charade: it wants to LOOK intelligent, while not BEING intelligent, because that's what the engineers who built it wanted it to do. | |
| ▲ | lkey 2 days ago | parent | prev | next [-] | | Accepting as true "We don't know how the brain works in a precise way" does not mean that obviously untrue statements about the human brain cannot still be made.
Your brain specifically, however, is an electric rat that pulls on levers of flesh while yearning for a taste of God's holiest cheddar.
You might reply, "no! that cannot be!", but my statement isn't untrue, so it goes. | |
| ▲ | staticman2 2 days ago | parent | prev | next [-] | | >>>There are no true and untrue claims about how the brain works, because we have no idea how it works. Which is why if you pick up a neuroscience textbook it's 400 pages of blank white pages, correct? There are different levels of understanding. I don't need to know how a TV works to know there aren't little men and women acting out the TV shows when I put them on. I don't need to know how the brain works in detail to know claims that humans are doing the same things as LLMs to be similarly silly. | | |
| ▲ | solumunus 2 days ago | parent | next [-] | | The trouble is that no one knows enough about how the brain works to refute that claim. | | |
| ▲ | staticman2 2 days ago | parent [-] | | There's no serious claim that needs refuting. I don't think any serious person thinks LLMs work like the human brain. People claiming this online aren't going around murdering their spouses like you'd delete an old LLama model from your hard drive. I'm not sure why people keep posting these sorts of claims they can't possibly actually believe if we look at their demonstrable real life behavior. | | |
| ▲ | solumunus 2 days ago | parent [-] | | We’re obviously more advanced than an LLM, but to claim that human beings simply generate output based on inputs and context (environment, life experience) is not silly. > People claiming this online aren't going around murdering their spouses like you'd delete an old LLama model from your hard drive. Not sure what you’re trying to say here. | | |
| ▲ | staticman2 2 days ago | parent [-] | | I'm saying you'd object to being treated like an LLM and don't really have conviction when you make these claims. I'd also say stringing together A.I. buzzwords (input output) to describe humans isn't really an argument so much as what philosophers call a category error. | | |
| ▲ | solumunus 2 days ago | parent [-] | | That I wouldn’t treat a human like an LLM is completely irrelevant to the topic. Input and output are not AI buzzwords, they’re fundamental terms in computation. The argument that human beings are computational has been alive in philosophy since the 1940’s brother… |
|
|
|
| |
| ▲ | naasking 2 days ago | parent | prev [-] | | > I don't need to know how the brain works in detail to know claims that humans are doing the same things as LLMs to be similarly silly. Yes you do. It's all computation in the end, and isomorphisms can often be surprising. |
| |
| ▲ | solumunus 2 days ago | parent | prev [-] | | Our output is quite literally the sum of our hardware (genetics) and input (immediate environment and history). For anyone who agrees that free will is nonsense, the debate is already over, we’re nothing more than output generating biological machines. |
| |
| ▲ | slightwinder 2 days ago | parent | prev [-] | | Similar, not identical. Like a bicycle and car are both vehicles with tires, but are still not identical vessels. |
| |
| ▲ | hitarpetar 2 days ago | parent | prev | next [-] | | > We also just brute force "autocompletion" with our learned knowledge and combine it to new parts, which we then add to our learned knowledge to deepen the process you know this because you're a cognitive scientist right? or because this is the consensus in the field? | |
| ▲ | Psyladine 2 days ago | parent | prev [-] | | >Neither can a toddler nor an animal. The level of ability is irrelevant for evaluating its foundation. Its foundation of rational logical thought that can't process basic math? Even a toddler understands 2 is more than 1. |
| |
| ▲ | ako 2 days ago | parent | prev | next [-] | | An LLM by itself is not thinking, just remembering and autocompleting. But if you add a feedback loop where it can use tools, investigate external files or processes, and then autocomplete on the results, you get to see something that is (close to) thinking. I've seen claude code debug things by adding print statements in the source and reasoning on the output, and then determining next steps. This feedback loop is what sets AI tools apart, they can all use the same LLM, but the quality of the feedback loop makes the difference. | | |
| ▲ | DebtDeflation 2 days ago | parent | next [-] | | >But if you add a feedback loop where it can use tools, investigate external files or processes, and then autocomplete on the results, you get to see something that is (close to) thinking It's still just information retrieval. You're just dividing it into internal information (the compressed representation of the training data) and external information (web search, API calls to systems, etc). There is a lot of hidden knowledge embedded in language and LLMs do a good job of teasing it out that resembles reasoning/thinking but really isn't. | | |
| ▲ | ako 2 days ago | parent | next [-] | | No, it's more than information retrieval. The LLM is deciding what information needs to be retrieved to make progress and how to retrieve this information. It is making a plan and executing on it. Plan, Do, Check, Act. No human in the loop if it has the required tools and permissions. | |
| ▲ | naasking 2 days ago | parent | prev [-] | | > LLMs do a good job of teasing it out that resembles reasoning/thinking but really isn't. Given the fact that "thinking" still hasn't been defined rigourously, I don't understand how people are so confident in claiming they don't think. | | |
| ▲ | notepad0x90 2 days ago | parent [-] | | reasoning might be a better term to discuss as it is more specific? | | |
| ▲ | naasking 2 days ago | parent [-] | | It too isn't rigourously defined. We're very much at the hand-waving "I know it when I see it" [1] stage for all of these terms. [1] https://en.wikipedia.org/wiki/I_know_it_when_I_see_it | | |
| ▲ | notepad0x90 18 hours ago | parent [-] | | I can't speak for academic rigor, but it is very clear and specific from my understanding at least. Reasoning, simply put is the ability to come to a conclusion after analyzing information using a logic-derived deterministic algorithm. | | |
| ▲ | naasking 17 hours ago | parent [-] | | * Humans are not deterministic. * Humans that make mistakes are still considered to be reasoning. * Deterministic algorithms have limitations, like Goedel incompleteness, which humans seem able to overcome, so presumably, we expect reasoning to also be able to overcome such challenges. | | |
| ▲ | notepad0x90 5 hours ago | parent [-] | | 1) I didn't say we were, but when someone is called reasonable or acting with reason, then that implies deterministic/algorithmic thinking. When we're not deterministic, we're not reasonable. 2) Yes, to reason does imply to be infallible. The deterministic algorithms we follow are usually flawed. 3) I can't speak much to that, but I speculate that if "AI" can do reasoning, it would be a much more complex construct that uses LLMs (among other tools) as tools and variables like we do. |
|
|
|
|
|
| |
| ▲ | assimpleaspossi 2 days ago | parent | prev | next [-] | | >>you get to see something that is (close to) thinking. Isn't that still "not thinking"? | | |
| ▲ | ako 2 days ago | parent [-] | | Depends who you ask, what their definition of thinking is. |
| |
| ▲ | lossyalgo 2 days ago | parent | prev [-] | | Just ask it how many r's are in strawberry and you will realize there isn't a lot of reasoning going on here, it's just trickery on top of token generators. | | |
| ▲ | Workaccount2 2 days ago | parent | next [-] | | This is akin to "Show a human an optical illusion that exploits their physiology". LLM's be like "The dumb humans can't even see the dots"[1] [1]https://compote.slate.com/images/bdbaa19e-2c8f-435e-95ca-a93... | | |
| ▲ | lossyalgo 2 days ago | parent [-] | | haha that's a great analogy! How about non-determinism (i.e. hallucinations)? Ask a human ANY question 3 times and they will give you the same answer, every time, unless you prod them or rephrase the question. Sure the answer might be wrong 3 times, but at least you have consistency. Then again, maybe that's a disadvantage for humans! |
| |
| ▲ | adrianmonk 2 days ago | parent | prev [-] | | Ask an illiterate person the same thing and they will fail badly too. Is it impossible to have intelligence without literacy? (Bonus: if so, how was writing invented?) | | |
| ▲ | lossyalgo a day ago | parent [-] | | Yes but an illiterate person can be taught to read. Also LLMs generally fail (non-deterministically) at math in general, but humans can also be taught math. |
|
|
| |
| ▲ | lossyalgo 2 days ago | parent | prev | next [-] | | Furthermore regarding reasoning, just ask any LLM how many "r letters are in strawberry" - repeat maybe 3 times just to get a feeling for how much variance in answers you can get. And this "quirk" of the inability to get the right answer is something that after 2 years making fun of LLMs online on various forums is still an issue. The models aren't getting smarter, and definitely aren't thinking, they are still token generators with a few tricks on top to make them seem more intelligent than predecessors. | | |
| ▲ | EdNutting 2 days ago | parent | next [-] | | Sorry to say, Claude Sonnet 4.5 (with and without extended thinking) has this one solved now. So, not “any” LLM, but maybe some still need work. I presume they finally got around to training the internal reasoning loop into predicting when it needs to apply answer predictions character-by-character. —-
Transcripts from Claude Sonnet 4.5, initially with extended thinking: > How many a’s in the phrase “artichokes and aardvarks”? Be sure to check your answer. > Let me count the letter ‘a’ in “artichokes and aardvarks”:
• artichokes: 1 ‘a’ (at the beginning)
• and: 1 ‘a’ (at the beginning)
• aardvarks: 3 ‘a’s (two at the beginning: “aa”, and one in the middle: “ardvarks”)
Total: 1 + 1 + 3 = 5 ‘a’s
Let me verify by marking them: artichokes and aardvarks
Yes, there are 5 letter ‘a’s in the phrase “artichokes and aardvarks”. > Are you sure that’s correct? > Let me recount more carefully:
“artichokes and aardvarks”
Going through each word letter by letter:
1. artichokes - has 1 ‘a’ (first letter)
2. and - has 1 ‘a’ (first letter)
3. aardvarks - has 3 ‘a’s (first letter, second letter, and one more in the middle)
Total: 1 + 1 + 3 = 5 ‘a’s
Yes, I’m confident the answer is 5. Is there something specific you’d like me to recheck? > How many r’s in strawberry? > Let me count the letter ‘r’ in “strawberry”:
s-t-r-a-w-b-e-r-r-y
Going through each letter:
• s: not an r
• t: not an r
• r: 1st r
• a: not an r
• w: not an r
• b: not an r
• e: not an r
• r: 2nd r
• r: 3rd r
• y: not an r
There are 3 r’s in “strawberry”. <Without extended thinking>
> How many u’s in universalulis? > Looking at “universalulis” letter by letter:
u-n-i-v-e-r-s-a-l-u-l-i-s
There are 2 u’s in “universalulis”. | |
| ▲ | jamincan 2 days ago | parent | prev [-] | | For what it's worth, when I ask ChatGPT 5, it gets the correct answer every time. The response varies, but the answer is always three. | | |
| ▲ | ViewTrick1002 2 days ago | parent [-] | | Now try a different language. My take is hard RL tuning to fix these "gotcha:s" since the underlying model can't do it on its own. OpenAI is working on ChatGPT the application and ecosystem. They have transitioned from model building to software engineering with RL tuning and integration of various services to solve the problems the model can't do on its own. Make it feel smart rather than be smart. This means that as soon as you find a problem where you step out of the guided experience you get the raw model again which fails when encountering these "gotchas". Edit - Here's an example where we see a very tuned RL experience in English where a whole load of context is added on how to solve the problem while the Swedish prompt for the same word fails. https://imgur.com/a/SlD84Ih | | |
| ▲ | ACCount37 2 days ago | parent [-] | | You can tell it "be careful about the tokenizer issues" in Swedish and see how that changes the behavior. The only thing that this stupid test demonstrates is that LLM metacognitive skills are still lacking. Which shouldn't be a surprise to anyone. The only surprising thing is that they have metacognitive skills, despite the base model training doing very little to encourage their development. | | |
| ▲ | lossyalgo a day ago | parent [-] | | LLMs were not designed to count letters[0] since they work with tokens, so whatever trick they are now doing behind the scenes to handle this case, can probably only handle this particular case. I wonder if it's now included in the system prompt. I asked ChatGPT and it said it's now using len(str) and some other python scripts to do the counting, but who knows what's actually happening behind the scenes. [0] https://arxiv.org/pdf/2502.16705 | | |
| ▲ | ACCount37 a day ago | parent [-] | | There's no "trick behind the scenes" there. You can actually see the entire trick being performed right in front of you. You're just not paying attention. That trick? The LLM has succeeded by spelling the entire word out letter by letter first. It's much easier for an LLM to perform "tokenized word -> letters -> letter counts" than it is to perform "tokenized word -> letter counts" in one pass. But it doesn't know that! It copies human behavior from human text, and humans never had to deal with tokenizer issues in text! You can either teach the LLM that explicitly, or just do RLVR on diverse tasks and hope it learns the tricks like this by itself. |
|
|
|
|
| |
| ▲ | IanCal 3 days ago | parent | prev | next [-] | | > it's just auto-completing. It cannot reason Auto completion just means predicting the next thing in a sequence. This does not preclude reasoning. > I don't get why you would say that. Because I see them solve real debugging problems talking through the impact of code changes or lines all the time to find non-obvious errors with ordering and timing conditions on code they’ve never seen before. | | |
| ▲ | notepad0x90 2 days ago | parent [-] | | > This does not preclude reasoning. It does not imply it either. to claim reasoning you need evidence. it needs to reliably NOT hallucinate results for simple conversations for example (if it has basic reasoning). > Because I see them solve real debugging problems talking through the impact of code changes or lines all the time to find non-obvious errors with ordering and timing conditions on code they’ve never seen before. Programming languages and how programs work are extensively and abundantly documented, solutions to problems and how to approach them,etc.. have been documented on the internet extensively. It takes all of that data and it completes the right text by taking the most correct path way based on your input. it does not actually take your code and debug it. it is the sheer volume of data it uses and the computational resources behind it that are making it hard to wrap your head around the difference between guessing and understanding. You too can look at enough stack overflow and (poorly) guess answers for questions without understanding anything about the topic and if you guess enough you'll get some right. LLMs are just optimized to get the amount of correct responses to be high. | | |
| ▲ | IanCal 2 days ago | parent [-] | | > It does not imply it either. Right, it's irrelevant to the question of whether they can reason. > to claim reasoning you need evidence Frankly I have no idea what most people are talking about when they use the term and say these models can't do it. It seems to be a similarly hand-wavey exercise as when people talk about thinking or understanding. > it needs to reliably NOT hallucinate results for simple conversations for example (if it has basic reasoning). That's not something I commonly see in frontier models. Again this doesn't seem related to reasoning. What we call hallucinations would be seen in something that could reason but had a fallible memory. I remember things incorrectly and I can reason. > it does not actually take your code and debug it It talks through the code (which it has not seen) and process step by step, can choose to add logging, run it, go through the logs, change what it thinks is happening and repeat. It can do this until it explains what is happening, creates test cases to show the problem and what triggers it, fixes it and shows the tests pass. If that's not debugging the code I really don't know what to call it. |
|
| |
| ▲ | logicchains 2 days ago | parent | prev | next [-] | | >I don't get why you would say that. it's just auto-completing. It cannot reason. It won't solve an original problem for which it has no prior context to "complete" an approximated solution with. you can give it more context and more data,but you're just helping it complete better. it does not derive an original state machine or algorithm to solve problems for which there are no obvious solutions. it instead approximates a guess (hallucination). I bet you can't give an example such written problem that a human can easily solve but no LLM can. | |
| ▲ | xanderlewis 3 days ago | parent | prev | next [-] | | > I don't get why you would say that. Because it's hard to imagine the sheer volume of data it's been trained on. | | |
| ▲ | utopiah 2 days ago | parent [-] | | And because ALL the marketing AND UX around LLMs is precisely trying to imply that they are thinking. It's not just the challenge of grasping the ridiculous amount of resources poured in, which does including training sets, it's because actual people are PAID to convince everybody those tools are actually thinking. The prompt is a chatbox, the "..." are there like a chat with a human, the "thinking" word is used, the "reasoning" word is used, "hallucination" is used, etc. All marketing. | | |
| ▲ | xanderlewis a day ago | parent [-] | | You're right. Unfortunately, it seems that not many are willing to admit this and be (rightly) impressed by how remarkably effective LLMs can be, at least for manipulating language. |
|
| |
| ▲ | madaxe_again 3 days ago | parent | prev | next [-] | | The vast majority of human “thinking” is autocompletion. Any thinking that happens with words is fundamentally no different to what LLMs do, and everything you say applies to human lexical reasoning. One plus one equals two. Do you have a concept of one-ness, or two-ness, beyond symbolic assignment? Does a cashier possess number theory? Or are these just syntactical stochastic rules? I think the problem here is the definition of “thinking”. You can point to non-verbal models, like vision models - but again, these aren’t hugely different from how we parse non-lexical information. | | |
| ▲ | gloosx 2 days ago | parent | next [-] | | > Any thinking that happens with words is fundamentally no different from what LLMs do. This is such a wildly simplified and naive claim. "Thinking with words" happens inside a brain, not inside a silicon circuit with artificial neurons bolted in place. The brain is plastic, it is never the same from one moment to the next. It does not require structured input, labeled data, or predefined objectives in order to learn "thinking with words." The brain performs continuous, unsupervised learning from chaotic sensory input to do what it does. Its complexity and efficiency are orders of magnitude beyond that of LLM inference. Current models barely scratch the surface of that level of complexity and efficiency. > Do you have a concept of one-ness, or two-ness, beyond symbolic assignment? Obviously we do. The human brain's idea of "one-ness" or "two-ness" is grounded in sensory experience — seeing one object, then two, and abstracting the difference. That grounding gives meaning to the symbol, something LLMs don't have. | | |
| ▲ | gkbrk 2 days ago | parent | next [-] | | LLMs are increasingly trained on images for multi-modal learning, so they too would have seen one object, then two. | | |
| ▲ | gloosx 2 days ago | parent [-] | | They never saw any kind of object, they only saw labeled groups of pixels – basic units of a digital image, representing a single point of color on a screen or in a digital file. Object is a material thing that can be seen and touched. Pixels are not objects. | | |
| ▲ | gkbrk 2 days ago | parent | next [-] | | Okay, goalpost has instantly moved from seeing to "seeing and touching". Once you feed in touch sensor data, where are you going to move the goalpost next? Models see when photons hit camera sensors, you see when photons hit your retina. Both of them are some kind of sight. | | |
| ▲ | gloosx a day ago | parent [-] | | The difference between photons hitting the camera sensors and photons hitting the retina is immense. With a camera sensor, the process ends in data: voltages in an array of photodiodes get quantized into digital values. There is no subject to whom the image appears. The sensor records but it does not see. When photons hit the retina, the same kind of photochemical transduction happens — but the signal does not stop at measurement. It flows through a living system that integrates it with memory, emotion, context, and self-awareness. The brain does not just register and store the light, it constructs an experience of seeing, a subjective phenomenon — qualia. Once models start continuously learning from visual subjective experience, hit me up – and I'll tell you the models "see objects" now. Until direct raw photovoltaic information stream about the world around them without any labelling can actually make model to learn anything, they are not even close to "seeing". |
| |
| ▲ | madaxe_again 2 days ago | parent | prev [-] | | My friend, you are blundering into metaphysics here - ceci n’est pas une pipe, the map is the territory, and all that. We are no more in touch with physical reality than an LLM, unless you are in the habit of pressing your brain against things. Everything is interpreted through a symbolic map. | | |
| ▲ | gloosx a day ago | parent [-] | | when photons strike your retina, they are literally striking brain tissue that is been pushed outward into the skull front window. Eyes are literally the brain, so yes, we are pressing it against things to "see" them. |
|
|
| |
| ▲ | madaxe_again 2 days ago | parent | prev [-] | | The instantiation of models in humans is not unsupervised, and language, for instance, absolutely requires labelled data and structured input. The predefined objective is “expand”. See also: feral children. | | |
| ▲ | gloosx a day ago | parent [-] | | Children are not shown pairs like "dog": [object of class Canine] They infer meaning from noisy, ambiguous sensory streams. The labels are not explicit, they are discovered through correlation, context, and feedback. So although caregivers sometimes point and name things, that is a tiny fraction of linguistic input, and it is inconsistent. Children generalize far beyond that. Real linguistic input to a child is incomplete, fragmented, error-filled, and dependens on context. It is full of interruptions, mispronunciations, and slang. The brain extracts structure from that chaos. Calling that "structured input" confuses the output - inherent structure of language - with the raw input, noisy speech and gestures. The brain has drives: social bonding, curiosity, pattern-seeking. But it doesn't have a single optimisation target like "expand." Objectives are not hardcoded loss functions, they are emergent and changing. You're right that lack of linguistic input prevents full language development, but that is not evidence of supervised learning. It just shows that exposure to any language stream is needed to trigger the innate capacity. Both complexity and efficiency of the human learning is just on another level. Transformers are child's play compared to that level. They are not going to gain consciousness, and no AGI will happen in the foreseeable future, it is all just marketing crap, and it's becoming more and more obvious as the dust settles. |
|
| |
| ▲ | notepad0x90 2 days ago | parent | prev [-] | | We do a lot of autocompletion and LLMs overlap with that for sure. I don't know about the "vast majority" even basic operations like making sure we're breathing or have the right hormones prompted are not guesses but deterministic algorithmic ops. Things like object recognition and speech might qualify as autocompletion.
But let's say you need to setup health-monitoring for an application. that's not an autocomplete operation. you must evaluate various options, have opinions on it, weigh priorities,etc.. in other words, we do autocompletion but even then the autocompletion is a basic building block or tool we use in constructing more complex decision logic. If you train an animal to type the right keys on a keyboard that generates a hello world program, you didn't just teach them how to code. they just memorized the right keys that lead to their reward. a human programmer understands the components of the code, the intent and expectations behind it, and can reason about how changes would affect outcomes. the animal just knows how the reward can be obtained most reliably. |
| |
| ▲ | naasking 2 days ago | parent | prev | next [-] | | > don't get why you would say that. it's just auto-completing. https://en.wikipedia.org/wiki/Predictive_coding > If it thought and reasoned, it would know to ignore those and focus on the variables and input at hand to produce deterministic and consistent output You only do this because you were trained to do this, eg. to see symmetries and translations. | |
| ▲ | Kichererbsen 3 days ago | parent | prev | next [-] | | Sure. But neither do you. So are you really thinking or are you just autocompleting? When was the last time you sat down and solved an original problem for which you had no prior context to "complete" an approximated solution with? When has that ever happened in human history? All the great invention-moment stories that come to mind seem to have exactly that going on in the background: Prior context being auto-completed in an Eureka! moment. | | |
| ▲ | notepad0x90 2 days ago | parent [-] | | I think (hah) you're understimating what goes on when living things (even small animals) think. We use auto-compleition for some tasks, but it is a component of what we do. Let's say your visual system auto-completes some pattern and detects a snake while you're walking, that part is auto-completion. You will probably react by freezing or panicing, that part is not auto-compleition, it is a deterministic algorithm. But then you process the detected object, auto-compleiting again to identify it as just a long cucumber. But again, the classification part is auto-completion. What will you do next? "Hmm, free cucumber, i can cook with it for a meal" and you pick it up. auto-completion is all over that simple decision, but you're using results of auto-completion to derive association (food), check your hunger level (not auto-completion), determine that the food is desirable and safe to eat (some auto-compleition), evalute what other options you have for food (evaluate auto-complete outputs), and then instruct your nervous system to pick it up. We use auto-compleition all the time as an input, we don't reason using auto-compleition in other words. You can argue that if all your input is from auto-completion (it isn't) then it makes no difference. But we have deterministic reasoning logical systems that evaluate auto-completion outputs. if your cucumber detection identified it as rotten cucumber, then decision that it is not safe to eat is not done by auto-completion but a reasoning logic that is using auto-completion output. You can approximate the level of rot but once you recognize it as rotten, you make decision based on that information. You're not approximating a decision, you're evaluating a simple logic of: if(safe()){eat();}. Now amp that up to solving very complex problems. what experiments will you run, what theories will you develop, what R&D is required for a solution,etc.. these too are not auto-completions. an LLM would auto complete these and might arrive at the same conclusion most of the time. but our brains are following algorithms we developed and learned over time where an LLM is just expanding on auto-completion but with a lot more data. In contrast, our brains are not trained on all the knowledge available on the public internet, we retain a tiny miniscule of that. we can arrive at similar conclusions as the LLM because we are reasoning and following algorithms matured and perfected over time. The big take away should be that, as powerful as LLMs are now, if they could reason like we do, they'd dominate us and become unstoppable. Because their auto-completion is many magnitudes better than ours, if they can write new and original code based on an understanding of problem solving algorithms, that would be gen ai. We can not just add 1 + 1 but prove that the addition operation is correct mathematically. and understand that when you add to a set one more object, the addition operation always increments. We don't approximate that, we always, every single time , increment because we are following an algorithm instead of choosing the most likely correct answer. |
| |
| ▲ | jiggawatts 2 days ago | parent | prev | next [-] | | You wrote your comment one word at a time, with the next word depending on the previous words written. You did not plan the entire thing, every word, ahead of time. LLMs do the same thing, so... how is your intelligence any different? | | |
| ▲ | ben_w 2 days ago | parent | next [-] | | A long time ago I noticed that I sometimes already had a complete thought before my inner monologue turned it into words. A few times I tried skipping the inner monologue because I'd clearly already thought the thought. Turns out the bit of my brain that creates the inner monologue from the thought, can generate a sense of annoyance that the rest of my brain can feel. Not that it matters, there's evidence that while LLMs output one word at a time, they've got forward-planning going on, having an idea of the end of a sentence before they get there. | | |
| ▲ | rcxdude 2 days ago | parent [-] | | Indeed, and it seems like they would really struggle to output coherent text at all if there was not some kind of pre-planning involved (see how even humans struggle with it in games where you have to construct a sentance by having each person shout out one word at a time). Even GPT-2 likely had at least some kind of planning for the next few words in order to be as coherent as it was. |
| |
| ▲ | lossyalgo 2 days ago | parent | prev [-] | | Tell that to German-speakers, where the verb comes last, and the order of things in sentences is not anything like English, therefore requiring you to think of the entire sentence before you just spit it out. Even the numbers are backwards (twenty-two is two-and-twenty) which requires thinking. Furthermore, when you ask an LLM to count how many r's are in the word strawberry, it will give you a random answer, "think" about it, and give you another random answer. And I guarantee you out of 3 attempts, including reasoning, it will flip-flop between right and wrong, but unlike a human, it will be random, because, unlike humans who, when asked "how many r's are in the word strawberry" will not be able to tell you the correct answer every. fucking. time. edit: formatting | | |
| ▲ | pka 2 days ago | parent | next [-] | | It seems models are pre-planning though: > How does Claude write rhyming poetry? Consider this ditty: > He saw a carrot and had to grab it, > His hunger was like a starving rabbit > To write the second line, the model had to satisfy two constraints at the same time: the need to rhyme (with "grab it"), and the need to make sense (why did he grab the carrot?). Our guess was that Claude was writing word-by-word without much forethought until the end of the line, where it would make sure to pick a word that rhymes. We therefore expected to see a circuit with parallel paths, one for ensuring the final word made sense, and one for ensuring it rhymes. > Instead, we found that Claude plans ahead. Before starting the second line, it began "thinking" of potential on-topic words that would rhyme with "grab it". Then, with these plans in mind, it writes a line to end with the planned word. [https://www.anthropic.com/research/tracing-thoughts-language...] | |
| ▲ | nxor 2 days ago | parent | prev [-] | | The part about strawberry is just not right. That problem was solved. And I do think it's a stretch to say German speakers think of the entire sentence before speaking it. | | |
| ▲ | lossyalgo 2 days ago | parent [-] | | LLMs were not designed to count letters[0] since they work with tokens, so whatever trick they are now doing behind the scenes to handle this case, can probably only handle this particular case. I wonder if it's now included in the system prompt. I asked ChatGPT and it said it's now using len(str) and some other python scripts to do the counting, but who knows what's actually happening behind the scenes. [0] https://arxiv.org/pdf/2502.16705 |
|
|
| |
| ▲ | 2 days ago | parent | prev [-] | | [deleted] |
|
|
| ▲ | layer8 3 days ago | parent | prev | next [-] |
| Sometimes after a night’s sleep, we wake up with an insight on a topic or a solution to a problem we encountered the day before. Did we “think” in our sleep to come up with the insight or solution? For all we know, it’s an unconscious process. Would we call it “thinking”? The term “thinking” is rather ill-defined, too bound to how we perceive our own wakeful thinking. When conversing with LLMs, I never get the feeling that they have a solid grasp on the conversation. When you dig into topics, there is always a little too much vagueness, a slight but clear lack of coherence, continuity and awareness, a prevalence of cookie-cutter verbiage. It feels like a mind that isn’t fully “there” — and maybe not at all. I would agree that LLMs reason (well, the reasoning models). But “thinking”? I don’t know. There is something missing. |
| |
| ▲ | AnIrishDuck 3 days ago | parent | next [-] | | > Sometimes after a night’s sleep, we wake up with an insight on a topic or a solution to a problem we encountered the day before. The current crop of models do not "sleep" in any way. The associated limitations on long term task adaptation are obvious barriers to their general utility. > When conversing with LLMs, I never get the feeling that they have a solid grasp on the conversation. When you dig into topics, there is always a little too much vagueness, a slight but clear lack of coherence, continuity and awareness, a prevalence of cookie-cutter verbiage. It feels like a mind that isn’t fully “there” — and maybe not at all. One of the key functions of REM sleep seems to be the ability to generalize concepts and make connections between "distant" ideas in latent space [1]. I would argue that the current crop of LLMs are overfit on recall ability, particularly on their training corpus. The inherent trade-off is that they are underfit on "conceptual" intelligence. The ability to make connections between these ideas. As a result, you often get "thinking shaped objects", to paraphrase Janelle Shane [2]. It does feel like the primordial ooze of intelligence, but it is clear we still have several transformer-shaped breakthroughs before actual (human comparable) intelligence. 1. https://en.wikipedia.org/wiki/Why_We_Sleep
2. https://www.aiweirdness.com/ | |
| ▲ | popalchemist 2 days ago | parent | prev | next [-] | | There is simply put no ongoing process and no feedback loop. The model does not learn. The cognition ends when the inference cycle ends. It's not thinking, it just produces output that looks similar to the output of thinking. But the process by which it does that is wholly unreleated. | | |
| ▲ | AlphaAndOmega0 2 days ago | parent | next [-] | | Humans with certain amnestic syndromes are incapable of learning. That doesn't make them unintelligent or incapable of thought. | |
| ▲ | ako 2 days ago | parent | prev | next [-] | | Most AI tooling is shipped with a feedback loop around the LLM. The quality of Claude Code for example lies in the feedback loop it provides on your code. Maybe the LLM itself isn't thinking, but the Agent which ships an LLM plus feedback loop definitely shows thinking qualities. Just now in an debugging session with claude code: * let me read this file...
* let me read this file...
* I think there's a caching issue with the model after dropping the module. Let me check if there's a save or reload needed after DROP MODULE. First, let me verify something:
* creates a bash/javascript script to verify its assumption
* runs the script (after review and approval)
* Aha! I found the problem! Look at the output...
How is this not thinking? | | |
| ▲ | popalchemist 2 days ago | parent [-] | | without getting into theory of mind it's a bit difficult to elaborate, and I don't have the time or the will for that. But the short version is that thinking is interconnected with BEING as well as will, and the Agent has neither, in a philosophically formal sense. The agent is deterministically bound. So it is a fancy Rube Goldberg machine that outputs letters in a way that creates the impression of thought, but it is not thought, in the same way that some birds can mimic human speech without even the slightest hint as to the words' or sentences' meaning, underlying grammar, connotations, subtext, context, intended use, likely effect, etc. Is speech speech if the speaker has no concept whatsoever of said speech's content, and can not use it to actualize itself? I'd say no. It's mimicry, but not speech. So that means speech is something more than just its outward aspect - the words. It is the relation of something invisible, some inner experience known only to the speaker, VIA the words. Whereas a gorilla who learns sign language to communicate and use that communication to achieve aims which have direct correlation with its sense of self - that's thought in the Cogito, Ergo Sum sense of the word. Thought as commonly concieved by the layman is a sort of isolated phenomenon that is mechanical in nature and can be judged by its outward effects; whereas in the philosophical tradition defining thought is known to be one of the hard questions for its mysterious qualia of being interconnected with will and being as described above. Guess I gave you the long answer. (though, really, it could be much longer than this.) The Turing Test touches on this distinction between the appearance of thought and actual thought. The question goes all the way down to metaphysics; some (such as myself) would say that one must be able to define awareness (what some call consciousness - though I think that term is too loaded) before you can define thought. In fact that is at the heart of the western philosophical tradition; and the jury consensus remains elusive after all these thousands of years. | | |
| ▲ | TheOtherHobbes 2 days ago | parent | next [-] | | The obvious counterargument is that a calculator doesn't experience one-ness, but it still does arithmetic better than most humans. Most people would accept that being able to work out 686799 x 849367 is a form of thinking, albeit an extremely limited one. First flight simulators, then chess computers, then go computers, then LLMs are the same principle extended to much higher levels of applicability and complexity. Thinking in itself doesn't require mysterious qualia. It doesn't require self-awareness. It only requires a successful mapping between an input domain and an output domain. And it can be extended with meta-thinking where a process can make decisions and explore possible solutions in a bounded space - starting with if statements, ending (currently) with agentic feedback loops. Sentience and self-awareness are completely different problems. In fact it's likely with LLMs that we have off-loaded some of our cognitive techniques to external hardware. With writing, we off-loaded memory, with computing we off-loaded basic algorithmic operations, and now with LLMs we have off-loaded some basic elements of synthetic exploratory intelligence. These machines are clearly useful, but so far the only reason they're useful is because they do the symbol crunching, we supply the meaning. From that point of view, nothing has changed. A calculator doesn't know the meaning of addition, an LLM doesn't need to know the meaning of "You're perfectly right." As long as they juggle symbols in ways we can bring meaning to - the core definition of machine thinking - they're still "thinking machines." It's possible - I suspect likely - they're only three steps away from mimicking sentience. What's needed is a long-term memory, dynamic training so the model is constantly updated and self-corrected in real time, and inputs from a wide range of physical sensors. At some point fairly soon robotics and LLMs will converge, and then things will get interesting. Whether or not they'll have human-like qualia will remain an unknowable problem. They'll behave and "reason" as if they do, and we'll have to decide how to handle that. (Although more likely they'll decide that for us.) | | |
| ▲ | popalchemist 2 days ago | parent | next [-] | | Some of your points are lucid, some are not. For example, an LLM does not "work out" any kind of math equation using anything approaching reasoning; rather it returns a string that is "most likely" to be correct using probability based on its training. Depending on the training data and the question being asked, that output could be accurate or absurd. That's not of the same nature as reasoning your way to an answer. | |
| ▲ | ako 2 days ago | parent | prev [-] | | So if you don’t have a long term memory, you’re not capable of sentience? Like the movie memento, where the main character needs to write down everything to remind him later because he’s not able to remember anything. This is pretty much like llms using markdown documents to remember things. |
| |
| ▲ | Workaccount2 2 days ago | parent | prev | next [-] | | "To escape the paradox, we invoke what we call the “Homunculus Defense”: inside every human is a tiny non-stochastic homunculus that provides true understanding. This homunculus is definitionally not a stochastic parrot because: 1. It has subjective experience (unprovable but assumed) 2. It possesses free will (compatibilist definitions need not apply) 3. It has attended at least one philosophy seminar"[1] [1]https://ai.vixra.org/pdf/2506.0065v1.pdf | |
| ▲ | ako 2 days ago | parent | prev | next [-] | | For practical every day uses, does it really matter if it is "real thinking" or just really good "artificial thinking" with the same results? The machine can use artificial thinking to reach desired goals and outcomes, so for me it's the kind of thinking i would want from a machine. | | |
| ▲ | popalchemist 2 days ago | parent [-] | | For practical purposes, abstractions are useful, so, no, it doesn't matter. But the post we are responding to is directly posing the philosophical question of whether the activity of LLM agents constitutes thought. |
| |
| ▲ | Earw0rm 2 days ago | parent | prev [-] | | It seems pretty clear to me though that being good at intellectual tasks / the sort of usefulness we ascribe to LLMs doesn't strongly correlate with awareness. Even just within humans - many of the least intellectually capable humans seem to have a richer supply of the traits associated with awareness/being than some of the allegedly highest-functioning. On average you're far more likely to get a sincere hug from someone with Down's syndrome than from a multi-millionaire. But I'm more interested in this when it comes to the animal kingdom, because while ChatGPT is certainly more useful than my cat, I'm also pretty certain that it's a lot less aware. Meaningful awareness - feelings - seems to be an evolutionary adaptation possessed by k-strategy reproducing vertebrates. Having a small number of kids and being biologically wired to care for them has huge implications for your motivation as an animal, and it's reasonable to think that a lot of our higher emotions are built on hardware originally evolved for that purpose. (Albeit the evolutionary origins of that are somewhat murky - to what extent mammals/birds reuse capabilities that were developed by a much earlier common ancestor, or whether it's entirely parallel evolution, isn't known afaik - but birds seem to exhibit a similar set of emotional states to mammals, that much is true). |
|
| |
| ▲ | Festro 2 days ago | parent | prev [-] | | You're moving the goalposts and contradicting yourself with language games. Something doesn't need to learn to think. I think all the time without learning. There's also an argument for machines already starting to crack learning with literal reinforcement training and feedback loops. Your language game was when you said the 'cognition ends...', as cognition is just a synonym for thinking. "The thinking ends when the inference cycle ends. It's not thinking'" becomes a clear contradiction. As for "the process by which it does that is wholly unrelated", buddy it's modelled on human neuron behaviour. That's how we've had this generative AI breakthrough. We've replicated human mental cognition as closely as we can with current technology and the output bears striking resemblance to our own generative capabilities (thoughts). Happy to admit it's not identical, but it's damn well inside the definition of thinking, may also cover learning. It may be better to take a second look at human thinking and wonder if it's as cryptic and deep as we thought ten, twenty years ago. |
| |
| ▲ | creer 3 days ago | parent | prev | next [-] | | Interesting, you think the associations your brain comes up with during sleep are NOT thinking? | | |
| ▲ | madaxe_again 3 days ago | parent | next [-] | | I’d say it’s more like annealing. Sort, filter, compress, integrate. Updates your models for the next morning, which is why the answer is there when it wasn’t before. | |
| ▲ | keiferski 3 days ago | parent | prev [-] | | I assume most people agree with that framing. “Thinking” generally implies a level of consciousness or self-awareness. “Let me think about this.”
“I have to think on it.” My brain regulates all sorts of processes unconsciously, like breathing, for example. I don’t treat those as “thinking,” so I don’t know why other unconscious brain activity would be either. | | |
| ▲ | creer 2 days ago | parent [-] | | I don't know that most people think about that at all. For me, I have learned not to dismiss the remuninating, reordering, reframing that my brain performs without me deliberately reasoning line by line about it. Limiting my brain to deliberate line by line reasoning would be very limiting and costly. "Thinking" to me is very much NOT just conscious reasoning. So much of what I think is not done consciously. Indeed "let me think about it" is often simply giving my brain time to "sit on it", for another expression - only after which will I have enough mind time on the various alternatives for a worthwhile conscious decision. |
|
| |
| ▲ | madaxe_again 3 days ago | parent | prev | next [-] | | Perhaps this is an artefact of instantiation - when you talk with an LLM, the responding instance is just that - it comes into being, inhales your entire chat history, and then continues like the last chap, finishes its response, and dies. The continuity is currently an illusion. | |
| ▲ | bithead 3 days ago | parent | prev | next [-] | | Do LLMs ever ask for you to clarify something you said in a way a person who doesn't quite understand what you said will do? | | |
| ▲ | Workaccount2 2 days ago | parent | next [-] | | Sometimes I think people leveraging criticisms of LLMs used ChatGPT 3 years ago and haven't touched one since, except for asking how many r's are in strawberry a year and a half ago. | |
| ▲ | savolai 3 days ago | parent | prev | next [-] | | Yeah, as someone who has gained a lot of interaction skills by playing with the constructivist learning ennvironment called the enneagram, I can attest that it much resembles behaviour characteristic of certain enneatypes. | |
| ▲ | willmarch 3 days ago | parent | prev | next [-] | | Yes, often | |
| ▲ | brabel 3 days ago | parent | prev [-] | | What now, two minutes using one and you are going to get that! |
| |
| ▲ | lordnacho 2 days ago | parent | prev | next [-] | | > When conversing with LLMs, I never get the feeling that they have a solid grasp on the conversation. When you dig into topics, there is always a little too much vagueness, a slight but clear lack of coherence, continuity and awareness, a prevalence of cookie-cutter verbiage. It feels like a mind that isn’t fully “there” — and maybe not at all. Much like speaking to a less experienced colleague, no? They say things that contain the right ideas, but arrange it unconvincingly. Still useful to have though. | |
| ▲ | petralithic 3 days ago | parent | prev [-] | | > Would we call it “thinking”? Yes I would. |
|
|
| ▲ | simulator5g 3 days ago | parent | prev | next [-] |
| Having seen photocopiers so many times produce coherent, sensible, and valid chains of words on a page, I am at this point in absolutely no doubt that they are thinking. |
| |
| ▲ | slightwinder 2 days ago | parent | next [-] | | Photocopiers are the opposite of thinking. What goes in, goes out, no transformation or creating of new data at all. Any change is just an accident, or an artifact of the technical process. | | |
| ▲ | HarHarVeryFunny 2 days ago | parent | next [-] | | That's not actually true - try photocopying a banknote with the security pattern on it. | | |
| ▲ | slightwinder 2 days ago | parent | next [-] | | True, modern Photocopiers are not really Photocopiers any more and just a pc with scanner and printer. Those could be described as "thinking", in the same sense as any computer is thinking. But from my understanding, the original comment was explicitly about a pure copy-device, without any "brain" involved. | |
| ▲ | _boffin_ 2 days ago | parent | prev [-] | | That has a logic condition, no? | | |
| ▲ | HarHarVeryFunny 2 days ago | parent [-] | | Not sure what you mean - there's a pattern of dots (the EURion constellation) on the banknotes of a number of countries that photocopiers recognize and will refuse to copy. | | |
| ▲ | _boffin_ a day ago | parent [-] | | I could have been more descriptive, but yes, exactly what you described. The logic of: if i see this exact pattern, refuse to duplicate or make extremely apparent it's a fake. |
|
|
| |
| ▲ | Libidinalecon 2 days ago | parent | prev | next [-] | | It is the wrong category error. The proper category error in the context of the discussion would be to say the photocopier is drawing a picture. It doesn't matter how well or not the photocopier recreates an image. To say the photocopier is drawing a picture is just nonsense and has no meaning. The same category error as to say the LLM is "thinking". Of course, the category error could be well exploited for marketing purposes if you are in the business of selling photocopiers or language models. | |
| ▲ | justinclift 2 days ago | parent | prev [-] | | So, if there's a flaw in its sensor and you get somewhat warped output, would you consider it thinking then? | | |
| |
| ▲ | efitz 2 days ago | parent | prev | next [-] | | https://www.theregister.com/2013/08/06/xerox_copier_flaw_mea... | | |
| ▲ | bayindirh 2 days ago | parent [-] | | That's not a flaw. That model's creativity tuned a bit too high. It's a happy little copier which can be a little creative and unconventional with reasoning, at times. |
| |
| ▲ | Zardoz84 2 days ago | parent | prev | next [-] | | I saw Dr. Abuse producing coherent, sensible and valid chains of words, running on a 386. | |
| ▲ | throwaway-0001 2 days ago | parent | prev | next [-] | | I’ve seen so many humans bring stupid. Definitively there is nothing in the brain. You see how doesn’t make sense what you saying? | |
| ▲ | 2 days ago | parent | prev | next [-] | | [deleted] | |
| ▲ | seeEllArr 2 days ago | parent | prev [-] | | [dead] |
|
|
| ▲ | burnte 2 days ago | parent | prev | next [-] |
| The first principle is that you must not fool yourself, and you are the easiest person to fool. - Richard P. Feynman They're not thinking, we're just really good at seeing patterns and reading into things. Remember, we never evolved with non-living things that could "talk", we're not psychologically prepared for this level of mimicry yet. We're still at the stage of Photography when people didn't know about double exposures or forced perspective, etc. |
| |
| ▲ | naasking 2 days ago | parent | next [-] | | You're just assuming that mimicry of a thing is not equivalent to the thing itself. This isn't true of physical systems (simulated water doesn't get you wet!) but it is true of information systems (simulated intelligence is intelligence!). | | |
| ▲ | burnte 2 days ago | parent | next [-] | | > You're just assuming that mimicry of a thing is not equivalent to the thing itself. I'm not assuming that, that's literally the definition of mimicry: to imitate closely. You might say I'm assuming that it is mimicking and not actually thinking, but there's no evidence it's actually thinking, and we know exactly what is IS doing because we created the code that we used to build the model. They're not thinking, it's doing math, mathematical transformations of data. | | |
| ▲ | naasking 2 days ago | parent [-] | | > They're not thinking, it's doing math, mathematical transformations of data Whatever thinking fundamentally is, it also has an equivalence as a mathematical transformation of data. You're assuming the conclusion by saying that the two mathematical transformations of data are not isomorphic. A simulation of information processing is still information processing, just like running Windows in a QEMU VM is still running Windows. | | |
| ▲ | burnte 2 days ago | parent [-] | | > Whatever thinking fundamentally is, it also has an equivalence as a mathematical transformation of data. Do not confuse the mathematical description of physical processes as the world being made of math. > You're assuming the conclusion by saying that the two mathematical transformations of data are not isomorphic. Correct. They're not isomorphic. One is simple math that runs on electrified sand, and one is an unknown process that developed independently across a billion years. Nothing we're doing with AI today is even close to real thought. There are a billion trivial proofs that make the rounds as memes, like one R in strawberry, or being unable to count, etc. | | |
| ▲ | naasking 2 days ago | parent [-] | | > Do not confuse the mathematical description of physical processes as the world being made of math. Again, this doesn't apply to information. A simulation of a computation really is equivalent to that computation. > One is simple math that runs on electrified sand, and one is an unknown process that developed independently across a billion years. Right, so you admit that it's an unknown process, which means you literally cannot conclude that it is different to what LLMs are doing. > There are a billion trivial proofs that make the rounds as memes, like one R in strawberry, or being unable to count, etc. No, none of these are definitive proofs that they are not thinking. LLM "perceptions" are tokens, the strawberry question is basically asking it to figure out something that's below it's perceptual range. This has literally nothing to do with whether the way it processes information is or is not thinking. | | |
| ▲ | burnte 2 days ago | parent [-] | | > Right, so you admit that it's an unknown process, which means you literally cannot conclude that it is different to what LLMs are doing. If you truly feel human thinking and LLMs share more than a cocktail napkin's worth of similarity, I don't know what to say. Just treating it like a black box I can prove in minutes it's not thinking. Come on. I really don't get why people are so emotionally involved in this stuff. It's not thinking. It's ok that it's not thinking. Maybe someday we'll get there, but it's not today. | | |
| ▲ | naasking 2 days ago | parent [-] | | No you can't. You're just not proving what you think you're proving. |
|
|
|
|
| |
| ▲ | Tade0 2 days ago | parent | prev [-] | | But a simulated mind is not a mind. This was already debated years ago with the aid of the Chinese Room thought experiment. | | |
| ▲ | dkural 2 days ago | parent | next [-] | | The Chinese Room experiment applies equally well to our own brains - in which neuron does the "thinking" reside exactly? Searle's argument has been successfully argued against in many different ways. At the end of the day - you're either a closet dualist like Searle, or if you have a more scientific view and are a physicalist (i.e. brains are made of atoms etc. and brains are sufficient for consciousness / minds) you are in the same situation as the Chinese Room: things broken down into tissues, neurons, molecules, atoms. Which atom knows Chinese? | | |
| ▲ | Tade0 2 days ago | parent [-] | | The whole point of this experiment was to show that if we don't know whether something is a mind, we shouldn't assume it is and that our intuition in this regard is weak. I know I am a mind inside a body, but I'm not sure about anyone else. The easiest explanation is that most of the people are like that as well, considering we're the same species and I'm not special. You'll have to take my word on that, as my only proof for this is that I refuse to be seen as anything else. In any case LLMs most likely are not minds due to the simple fact that most of their internal state is static. What looks like thoughtful replies is just the statistically most likely combination of words looking like language based on a function with a huge number of parameters. There's no way for this construct to grow as well as to wither - something we know minds definitely do. All they know is a sequence of symbols they've received and how that maps to an output. It cannot develop itself in any way and is taught using a wholly separate process. | | |
| ▲ | dkural an hour ago | parent | next [-] | | I am arguing against Searle's Chinese Room argument, I am not positing that LLMs are minds. I am specifically refuting that your brain and the Chinese room can be both subject to the same reductionist argument Searle uses - if we accept, as you say, that you are a mind inside a body, which neuron, or atom does this mind reside in? My point is, if you accept Searle's argument, you have to accept it for brains, including your brain, as well. Now, separately, you are precisely the type of closet dualist I speak of. You say that you are a mind inside a body, but you have no way of knowing that others have minds -- take this to it's full conclusion: You have no way of knowing that you have a "mind" either. You feel like you do, as a biological assembly (which is what you are). Either way you believe in some sort of body-mind dualism, without realizing. Minds are not inside of bodies. What you call a mind is a potential emergent phenomenon of a brain. (potential - because brains get injured etc.). | |
| ▲ | naasking 2 days ago | parent | prev [-] | | > In any case LLMs most likely are not minds due to the simple fact that most of their internal state is static. This is not a compelling argument. Firstly, you can add external state to LLMs via RAG and vector databases, or various other types of external memory, and their internal state is no longer static and deterministic (and they become Turing complete!). Second if you could rewind time, then your argument suggests that all other humans would not have minds because you could access the same state of mind at that point in time (it's static). Why would you travelling through time suddenly erases all other minds in reality? The obvious answer is that it doesn't, those minds exist as time moves forward and then they reset when you travel backwards, and the same would apply to LLMs if they have minds, eg. they are active minds while they are processing a prompt. | | |
| ▲ | Tade0 2 days ago | parent [-] | | > and their internal state is no longer static and deterministic (and they become Turing complete!). But it's not the LLM that makes modifications in those databases - it just retrieves data which is already there. > Why would you travelling through time suddenly erases all other minds in reality? I'm not following you here. > they are active minds while they are processing a prompt. Problem is that this process doesn't affect the LLM in the slightest. It just regurgitates what it's been taught. An active mind is makes itself. It's curious, it gets bored, it's learning constantly. LLMs do none of that. You couldn't get a real mind to answer the same question hundreds of times without it being changed by that experience. | | |
| ▲ | naasking 2 days ago | parent [-] | | > But it's not the LLM that makes modifications in those databases - it just retrieves data which is already there. So what? > I'm not following you here. If you're time travelling, you're resetting the state of the world to some previous well-defined, static state. An LLM also starts from some well-defined static state. You claim this static configuration means there's no mind, so this entails that the ability to time travel means that every person who is not time travelling has no mind. > Problem is that this process doesn't affect the LLM in the slightest. It just regurgitates what it's been taught. An active mind is makes itself. People who are incapable forming new memories thus don't have minds? https://en.wikipedia.org/wiki/Anterograde_amnesia |
|
|
|
| |
| ▲ | naasking 2 days ago | parent | prev | next [-] | | > But a simulated mind is not a mind. This was already debated years ago with the aid of the Chinese Room thought experiment. Yes, debated and refuted. There are many well known and accepted rebuttals of the Chinese Room. The Chinese Room as a whole does understand Chinese. | |
| ▲ | echelon 2 days ago | parent | prev [-] | | > But a simulated mind is not a mind. How would the mind know which one it is? Maybe your mind is being simulated right now. | | |
| ▲ | Tade0 2 days ago | parent [-] | | > How would the mind know which one it is? I'm not assuming it is without hard proof - that's my only argument. > Maybe your mind is being simulated right now. I'm experiencing consciousness right now, so that would have to be a damn good simulation. |
|
|
| |
| ▲ | luxuryballs 2 days ago | parent | prev [-] | | yeah it’s just processing, calling it thinking is the same as saying my intel core 2 duo or M4 Pro is thinking, sure if you want to anthropomorphize it you could say it’s thinking, but why are we trying to say a computer is a person in the first place? seems kind of forced |
|
|
| ▲ | geon 3 days ago | parent | prev | next [-] |
| Having seen LLMs so many times produce incoherent, nonsensical and invalid chains of reasoning... LLMs are little more than RNGs. They are the tea leaves and you read whatever you want into them. |
| |
| ▲ | rcxdude 2 days ago | parent | next [-] | | They are clearly getting to useful and meaningful results with at a rate significantly better than chance (for example, the fact that ChatGPT can play chess well even though it sometimes tries to make illegal moves shows that there is a lot more happening there than just picking moves uniformly at random). Demanding perfection here seems to be odd given that humans also can make bizarre errors in reasoning (of course, generally at a lower rate and in a distribution of kinds of errors we are more used to dealing with). | | |
| ▲ | matthewkayin 2 days ago | parent [-] | | The fact that a model trained on the internet, on which the correct rules of chess are written, is unable to determine what is and is not a legal move, seems like a sign that these models are not reasoning about the questions asked of them. They are just giving responses that look like (and often are) correct chess moves. | | |
| ▲ | rcxdude 2 days ago | parent [-] | | It's a sign that they are 'reasoning' imperfectly. If they were just giving responses that 'looked like' chess moves, they would be very bad at playing chess. (And I would hazard a guess that they are a primarily learning chess from the many games that are posted, as opposed to working things out from the rules. Indeed, if you make up a game and tell chatGPT the rules, it tends to be even worse at following them, let alone figuring out optimal play. But again, it will do so significantly better than random chance, so it's doing something with the information you give it, even if it's not doing so very well. I think it's reasonable to call this thinking, or reasoning, but this mostly becomes an argument of semantics. either way they do it significantly better than random chance but still not tremendously well. If your expectation is that they cannot work with anything novel then you're going to be continually surprised, but if your expectation is that they're as good as a human that has 'learned' from all the material its been given, especially material that's in-context and not in the training data, then you're also going to be disappointed.) |
|
| |
| ▲ | bongodongobob 3 days ago | parent | prev [-] | | Ridiculous. I use it daily and get meaningful, quality results. Learn to use the tools. | | |
| ▲ | 9rx 2 days ago | parent | next [-] | | > Learn to use the tools. Thing is, you wouldn't need to learn to use the tools if the tool was able to think. A thinking entity is able to adapt to other parties who lack learnings. This confirms that LLMs are little more than fancy RNGs. > I use it daily and get meaningful, quality results. That's what the tea leaf readers say too, funnily enough. | | | |
| ▲ | aydyn 3 days ago | parent | prev | next [-] | | Learn to work on interesting problems? If the problem you are working on is novel and hard, the AI will stumble. Generalizing your experience to everyone else's betrays a lack of imagination. | | |
| ▲ | khafra 3 days ago | parent | next [-] | | > Generalizing your experience to everyone else's betrays a lack of imagination. One guy is generalizing from "they don't work for me" to "they don't work for anyone." The other one is saying "they do work for me, therefore they do work for some people." Note that the second of these is a logically valid generalization. Note also that it agrees with folks such as Tim Gowers, who work on novel and hard problems. | | |
| ▲ | dns_snek 2 days ago | parent [-] | | No, that's decidedly not what is happening here. One is saying "I've seen an LLM spectacularly fail at basic reasoning enough times to know that LLMs don't have a general ability to think" (but they can sometimes reproduce the appearance of doing so). The other is trying to generalize "I've seen LLMs produce convincing thought processes therefore LLMs have the general ability to think" (and not just occasionally reproduce the appearance of doing so). And indeed, only one of these is a valid generalization. | | |
| ▲ | MrScruff 2 days ago | parent | next [-] | | When we say "think" in this context, do we just mean generalize? LLMs clearly generalize (you can give one a problem that is not exactly in it's training data and it can solve it), but perhaps not to the extent a human can. But then we're talking about degrees. If it was able to generalize at a higher level of abstraction maybe more people would regard it as "thinking". | | |
| ▲ | dns_snek 2 days ago | parent [-] | | I meant it in the same way the previous commenter did: > Having seen LLMs so many times produce incoherent, nonsensical and invalid chains of reasoning... LLMs are little more than RNGs. They are the tea leaves and you read whatever you want into them. Of course LLMs are capable of generating solutions that aren't in their training data sets but they don't arrive at those solutions through any sort of rigorous reasoning. This means that while their solutions can be impressive at times they're not reliable, they go down wrong paths that they can never get out of and they become less reliable the more autonomy they're given. | | |
| ▲ | dagss 2 days ago | parent | next [-] | | It's rather seldom that humans arrive at solutions through rigorous reasoning. The word "think" doesn't mean "rigorous reasoning" in every day language. I'm sure 99% of human decisions are pattern matching on past experience. Even when mathematicians do in fact do rigorous reasoning, they use years to "train" first, to get experiences to pattern match from. | |
| ▲ | Workaccount2 2 days ago | parent | prev | next [-] | | I have been on a crusade now for about a year to get people to share chats where SOTA LLMs have failed spectacularly to produce coherent, good information. Anything with Heavy hallucinations and outright bad information. So far, all I have gotten is data that is outside the knowledge cutoff (this is by far the most common) and technicality wrong information (Hawsmer House instead of Hosmer House) kind of fails. I thought maybe I hit on something with the recent BBC study about not trusting LLM output, but they used 2nd shelf/old mid-tier models to do their tests. Top LLMs correctly answered their test prompts. I'm still holding out for one of those totally off the rails Google AI overviews hallucinations showing up in a top shelf model. | |
| ▲ | MrScruff 2 days ago | parent | prev [-] | | Sure, and I’ve seen the same. But I’ve also seen the amount to which they do that decrease rapidly over time, so if that trend continues would your opinion change? I don’t think there’s any point in comparing to human intelligence when assessing machine intelligence, there’s zero reason to think it would have similar qualities. It’s quite clear for the foreseeable future it will be far below human intelligence in many areas, while already exceeding humans in some areas that we regard as signs of intelligence. |
|
| |
| ▲ | sdenton4 2 days ago | parent | prev [-] | | s/LLM/human/ | | |
| ▲ | dns_snek 2 days ago | parent [-] | | Clever. Yes, humans can be terrible at reasoning too, but in any half decent technical workplace it's so rare for people to fail to apply logic as often and in ways that are as frustrating to deal with as LLMs. And if they are then they should be fired. I can't say I remember a single coworker that would fit this description though many were frustrating to deal with for other reasons, of course. | | |
|
|
| |
| ▲ | dimator 3 days ago | parent | prev | next [-] | | This is my experience. For rote generation, it's great, saves me from typing out the same boilerplate unit test bootstrap, or refactoring something that exists, etc. Any time I try to get a novel insight, it flails wildly, and nothing of value comes out. And yes, I am prompting incrementally and building up slowly. | | |
| ▲ | player1234 2 days ago | parent [-] | | [flagged] | | |
| ▲ | tomhow 2 days ago | parent [-] | | We've banned this account for repeated abusive comments to fellow community members. Normally we give warnings, but when it's as extreme and repetitive as we can see here, an instant ban is appropriate. If you don't want to be banned, you can email us at hn@ycombinator.com and demonstrate a sincere commitment to use HN as intended in future. |
|
| |
| ▲ | lordnacho 2 days ago | parent | prev | next [-] | | Even people who do actual hard work need a lot of ordinary scaffolding done for them. A secretary who works for an inventor is still thinking. | | | |
| ▲ | tmhn2 2 days ago | parent | prev | next [-] | | Research mathematicians have been finding the tools useful [1][2]. I think those problems are interesting, novel, and hard. The AI might stumble sometimes, but it also produces meaningful, quality results sometimes. For experts working on interesting problems, that is enough to be useful. [1] https://mathstodon.xyz/@tao/115420236285085121
[2] https://xcancel.com/wtgowers/status/1984340182351634571 | | |
| ▲ | dns_snek 2 days ago | parent [-] | | That's a motte and bailey fallacy. Nobody said that they aren't useful, the argument is that they can't reason [1]. The world is full of useful tools that can't reason or think in any capacity. [1] That does not mean that they can never produce texts which describes a valid reasoning process, it means that they can't do so reliably. Sometimes their output can be genius and other times you're left questioning if they even have the reasoning skills of a 1st grader. | | |
| ▲ | chimprich 2 days ago | parent [-] | | I don't agree that LLMs can't reason reliably. If you give them a simple reasoning question, they can generally make a decent attempt at coming up with a solution. Complete howlers are rare from cutting-edge models. (If you disagree, give an example!) Humans sometimes make mistakes in reasoning, too; sometimes they come up with conclusions that leave me completely bewildered (like somehow reasoning that the Earth is flat). I think we can all agree that humans are significantly better and more consistently good at reasoning than even the best LLM models, but the argument that LLMs cannot reliably reason doesn't seem to match the evidence. |
|
| |
| ▲ | the-mitr 2 days ago | parent | prev | next [-] | | Even most humans will stumble on hard problems, that is the reason they are hard in the first place | |
| ▲ | XenophileJKO 3 days ago | parent | prev | next [-] | | I'm genuinely curious what you work on that is so "novel" that an LLM doesn't work well on? I feel like so little is TRUELY novel. Almost everything is built on older concepts and to some degree expertise can be applied or repurposed. | | |
| ▲ | EagnaIonat 2 days ago | parent | next [-] | | Anything relatively new in a technology LLMs struggle with, especially if the documentation is lacking. Godot for example in ChatGPT. It may no longer still be the case, but the documentation for GoDot was lacking and often samples written by others didn't have a version number associated with it. So samples it would suggest would never work, and even when you told it the version number it failed to generate workable code. The other stuff I've noticed is custom systems. One I work with is a variation of Java, but LLMs were treating it as javascript. I had to create a LoRA just to get the model from not trying to write javascript answer. Even then it could never work, because it had never been trained on real world examples. | |
| ▲ | geon 2 days ago | parent | prev | next [-] | | It doesn't have to be very novel at all. Anything but the most basic TODO-list app. | |
| ▲ | aydyn 2 days ago | parent | prev [-] | | Literally anything in the science domain. Adding features to your software app is indeed usually not novel. | | |
| |
| ▲ | bongodongobob 2 days ago | parent | prev [-] | | Dude. We don't all work for NASA. Most day to day problems aren't novel. Most jobs aren't novel. Most jobs can't keep a variety of sometimes useful experts on hand.
I do my job and I go home and do my hobbies. Anything I can use at work to keep friction down and productivity up is extremely valuable. Example prompt (paraphrasing and dumbed down, but not a ton): Some users across the country can't get to some fileshares. I know networking, but I'm not on the networking team so I don't have full access to switch, router, and firewall logs/configurations. It looks kind of random, but there must be a root cause, let's find it. I can't use Python(security team says so) and I don't have access to a Linux box that's joined to the domain and has access the shares. We are on a Windows domain controller. Write me a PowerShell 5.1 compatible script to be run remotely on devices. Use AD Sites and Services to find groups of random workstations and users at each office and tries to connect to all shares at each other site. Show me progress in the terminal and output an Excel file and Dot file that clearly illustrates successful and failed connections. --- And it works. Ok, I can see the issue is from certain sites that use x AND y VPN ipsec tunnels to get to particular cloud resources. I give this info to networking and they fix it right away. Problem resolved in less than an hour. First of all, a couple years ago, I wouldn't have been able to justify writing something like this while an outage is occuring. Could I do it myself? Sure, but I'm going to have to look up the specifics of syntax and certain commands and modules. I don't write
PowerShell for a living or fun, but I do need to use it. I am familiar and know how to write it. But I sure as fuck couldn't sit down and spend an hour or two screwing around working on building a goddamn Dot file generator. Yes, years ago I had a whole pile of little utility modules I could use. But that's a far cry from what I can do now to fit the exact situation < 15 minutes while I do other things like pick up the phone, message coworkers, etc. Secondly, rather than building little custom tools to hook together as I need, I can just ask for the whole thing. I don't need to save any of that stuff anymore and re-figure out what CheckADFSConns(v2).PS1 that I wrote 8 months ago does and how to use it. "Oh, that's not the one, what the did I name that? Where did I put it?" I work in an environment that is decades old, the company is over 100 years old, I didn't build any of it myself, is not a tech company, and has tons of tech debt and weird shit. AI is insanely useful. For any given problem, there are dozens of different rabbit holes I could go down because of decades of complete system overhaul changes. Today, I can toss a variety of logs at AI and if nothing else, get a sense of direction of why a handful of PCs are rejecting some web certificates. (Combination of a new security policy and their times mismatching the domain controller, because it was new, and NTP wasn't configured properly. I wasn't even looking for timestamps, but it noticed event offsets and pointed it out). I feel like this community isn't very familiar with what that's like. We aren't all working on self driving cars or whatever seems hard at a brand new company with new everything and no budget. Some of us need to keep the systems running that help people to make actual things. These environments are far from pristine and are held together by underpaid and underappreciated normies through sheer willpower. Is this kind of work breaking technical frontiers? No. But it's complicated, difficult, and unpredictable. Is it novel? The problems are, sometimes. Generalizing your experience to everyone else's betrays your lack of self-awareness, sir. |
| |
| ▲ | bopbopbop7 3 days ago | parent | prev | next [-] | | They are only meaningful and quality if you don’t know what you’re doing. But please do show some of this meaningful and quality work so I can be proven wrong. | | |
| ▲ | hattmall 3 days ago | parent | next [-] | | Yes, please this is literally what I want to see. I have yet to see an example where an LLM did anything that was sufficiently difficult. Not saying they can't be useful, but for anything past the basics they are really all over the place. And if we were paying anywhere near the true costs it wouldn't be even worth trying. | | |
| ▲ | bongodongobob 2 days ago | parent | next [-] | | I'm not spending the time to de-anyonymize and exfiltrate my work for you. I exert less energy, spend less time to do my work, and get paid the same. I'd encourage you to do the same. | |
| ▲ | seeEllArr 2 days ago | parent | prev [-] | | [dead] |
| |
| ▲ | bongodongobob 2 days ago | parent | prev [-] | | See my comment to parent. One example of many. You can say "Oh, well, it just sounds like your company needs better tools and processes, you don't really need AI for any of that. You should just invest in a tool for this and monitor that and have managment prioritize..." Yeah, I know, yet here we are and it saves me boatloads of time. | | |
| ▲ | bopbopbop7 2 days ago | parent [-] | | I never said any of that, look at my comment again. Or is the LLM replying for you also? | | |
| ▲ | bongodongobob 2 days ago | parent [-] | | You said they are only useful if you don't know what you're doing. Or do you have a goldfish memory? | | |
| ▲ | bopbopbop7 a day ago | parent [-] | | Yes, that’s what I said. But you quoted some make believe point you are arguing against. LLM brain rot really got you, huh? |
|
|
|
| |
| ▲ | 2 days ago | parent | prev | next [-] | | [deleted] | |
| ▲ | bryan_w 2 days ago | parent | prev [-] | | That's a skill issue on your end |
|
|
|
| ▲ | marcus_holmes 3 days ago | parent | prev | next [-] |
| Yes, I've seen the same things. But; they don't learn. You can add stuff to their context, but they never get better at doing things, don't really understand feedback. An LLM given a task a thousand times will produce similar results a thousand times; it won't get better at it, or even quicker at it. And you can't ask them to explain their thinking. If they are thinking, and I agree they might, they don't have any awareness of that process (like we do). I think if we crack both of those then we'd be a lot closer to something I can recognise as actually thinking. |
| |
| ▲ | theptip 3 days ago | parent | next [-] | | > But; they don't learn If we took your brain and perfectly digitized it on read-only hardware, would you expect to still “think”? Do amnesiacs who are incapable of laying down long-term memories not think? I personally believe that memory formation and learning are one of the biggest cruces for general intelligence, but I can easily imagine thinking occurring without memory. (Yes, this is potentially ethically very worrying.) | | |
| ▲ | lz400 3 days ago | parent | next [-] | | >If we took your brain and perfectly digitized it on read-only hardware, would you expect to still “think”? it wouldn't work probably, brains constantly alter themselves by forming new connections. Learning is inseparable from our intelligence. | | |
| ▲ | WalterSear 3 days ago | parent [-] | | Our intelligence, yes. But that doesn't establish it as essential for thought. | | |
| ▲ | lz400 3 days ago | parent [-] | | I mean, _we_ probably can't think with our wetware on a read-only substrate. It doesn't establish it as essential, just that the only sure example in nature of thought doesn't work that way. | | |
| ▲ | theptip 2 days ago | parent [-] | | Do you have any particular brain systems in mind that are essential for consciousness and also require mutable state? | | |
| ▲ | lz400 2 days ago | parent [-] | | I'm not an expert but as far I understand, plasticity is central to most complex operations of the brain and is likely to be involved in anything more complex than instinctive reactions. I'm happy to be corrected but it is my understanding that if you're thinking for a while on the same problem and establishing chains of reasoning, you are creating new connections and to me that means it's fundamental in the process of thinking. | | |
| ▲ | theptip a day ago | parent [-] | | Also not an expert :) I thought plasticity is an O(hours-days) learning mechanism. But I did some research and there is also Short Term Plasticity O(second) [1] which is a crucial part of working memory. We'd need that. But it seems it’s more of a volatile memory system, eg calcium ion depletion/saturation at the synapse, rather than a permanent wiring/potentiation change (please someone correct me if this isn’t right :) ). So I guess I’d just clarify “read only” to be a little more specific - I think you could run multiple experiments where you vary the line of what’s modeled in volatile memory at runtime, and what’s immutable. I buy that you need to model STP for thought, but also suspect at this timescale you can keep everything slower immutable and keep the second-scale processes like thought working. My original point still stands - your subjective experience in this scenario would be thought without long-term memory. 1: https://www.frontiersin.org/journals/computational-neuroscie... | | |
| ▲ | lz400 20 hours ago | parent [-] | | It would be funny if what you get from a read only human brain is a sort of memento guy who has no capacity to remember anything or follow a conversation... kind of like an LLM! |
|
|
|
|
|
| |
| ▲ | zeroonetwothree 3 days ago | parent | prev [-] | | > If we took your brain and perfectly digitized it on read-only hardware, would you expect to still “think”? Perhaps this is already known, but I would think there is a high chance that our brains require "write access" to function. That is, the very process of neural activity inherently makes modifications to the underlying structure. | | |
| ▲ | theptip 2 days ago | parent | next [-] | | > a high chance that our brains require "write access" to function There are multiple learning mechanisms that happen on different time-frames, eg neural plasticity, hippocampus are both longer-term processes for memory consolidation. Whereas the content of “RAM” might be better modeled as a set of fast dynamic weights representing ions and neurotransmitter concentrations. My hunch is that you could model the latter fast weights in volatile memory, but I wouldn’t count these as “modifying the structure”. Do you have any particular systems in mind where you have reason to believe that permanent modification is required for brains to function? | |
| ▲ | xwolfi 3 days ago | parent | prev [-] | | I wonder why we need to sleep so much though | | |
| ▲ | theptip 2 days ago | parent | next [-] | | Self-play for memory/skill consolidation, plus flushing out metabolites which can’t be cleaned while consciousness is online? | |
| ▲ | throwaway-0001 2 days ago | parent | prev [-] | | Rebalancing weights? |
|
|
| |
| ▲ | trenchpilgrim 2 days ago | parent | prev | next [-] | | > You can add stuff to their context, but they never get better at doing things, don't really understand feedback. I was using Claude Code today and it was absolutely capable of taking feedback to change behavior? | | |
| ▲ | marcus_holmes 2 days ago | parent [-] | | Within a session, yes, it will add each input to the context. Start a new session with a blank context and it will not have learned anything. | | |
| ▲ | trenchpilgrim a day ago | parent [-] | | > Start a new session with a blank context Why would I use a blank context with Claude Code? I have it dump the stuff I want to save in CLAUDE.md and it reloads it in future sessions. |
|
| |
| ▲ | jatora 3 days ago | parent | prev [-] | | This is just wrong though. They absolutely learn in-context in a single conversation within context limits. And they absolutely can explain their thinking; companies just block them from doing it. |
|
|
| ▲ | ben_w 2 days ago | parent | prev | next [-] |
| > Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking. While I'm not willing to rule *out* the idea that they're "thinking" (nor "conscious" etc.), the obvious counter-argument here is all the records we have of humans doing thinking, where the records themselves are not doing the thinking that went into creating those records. And I'm saying this as someone whose cached response to "it's just matrix multiplication it can't think/be conscious/be intelligent" is that, so far as we can measure all of reality, everything in the universe including ourselves can be expressed as matrix multiplication. Falsification, not verification. What would be measurably different if the null hypothesis was wrong? |
| |
| ▲ | chpatrick 2 days ago | parent [-] | | I've definitely had AIs thinking and producing good answers about specific things that have definitely not been asked before on the internet. I think the stochastic parrot argument is well and truly dead by now. | | |
| ▲ | Earw0rm 2 days ago | parent | next [-] | | I've also experienced this, to an extent, but on qualitative topics the goodness of an answer - beyond basic requirements like being parseable and then plausible - is difficult to evaluate. They can certainly produce good-sounding answers, but as to the goodness of the advice they contain, YMMV. | | |
| ▲ | chpatrick 2 days ago | parent [-] | | I've certainly got useful and verifiable answers. If you're not sure about something you can always ask it to justify it and then see if the arguments make sense. |
| |
| ▲ | hitarpetar 2 days ago | parent | prev [-] | | how do you definitely know that? | | |
| ▲ | stinos 2 days ago | parent | next [-] | | Also, does it matter? The point being made here is about the data LLMs have been trained with. Sure that contains questions&answers but obviously not all of it is in that form. Just like an encyclopedie contains answers without the questions. So imo specifying this as 'no-one asked this before' is irrelevant. More interesting: did OP get a sensible answer to a question about data which definitely was not in the training set? (and indeed, how was this 'definitely' established'). Not that if the answer is 'yes' that'll prove 'thinking', as opposed to calling it e.g. advanced autocompletion, but it's a much better starting point. | |
| ▲ | chpatrick 2 days ago | parent | prev [-] | | Because I gave them a unique problem I had and it came up with an answer it definitely didn't see in the training data. Specifically I wanted to know how I could interface two electronic components, one of which is niche, recent, handmade and doesn't have any public documentation so there's no way it could have known about it before. | | |
| ▲ | stinos 2 days ago | parent [-] | | one of which is niche, recent, handmade and doesn't have any public documentation I still see 2 possibilities: you asked it something similar enough that it came up with a fairly standard answer which just happened to be correct, or you gave it enough info. - for example you created a new line of MCUs called FrobnicatorV2, and asked is 'how do I connect a power supply X to FrobnicatorV2' and it gave an answer like 'connect red wire to VCC and black to GND'. That's not exactly special. - or, you did desribe that component in some way. And you did do that using standard electronics lingo so essentially in terms of other existing components which it definitely did know (unless you invented something completely new not using any currently know physics). As such it's irrelevant that your particular new component wasn't known because you gave away the answer by describing it? E.g. you aksed it 'how do I connect a power supply X to an MCU with power pins Y and Z'. Again nothing special. | | |
| ▲ | chpatrick 2 days ago | parent [-] | | If a human uses their general knowledge of electronics to answer a specific question they haven't seen before that's obviously thinking. I don't see why LLMs are held to a different standard. It's obviously not repeating an existing answer verbatim because that doesn't exist in my case. You're saying it's nothing "special" but we're not discussing whether it's special, but whether it can be considered thinking. | | |
| ▲ | stinos a day ago | parent [-] | | it's obviously not repeating an existing answer verbatim Not verbatim in the sense that the words are different doesn't make it thinking. Also when we say 'humans think' that means a lot more than only 'new question generates correct answer' or 'smart autocompletion'. See a lot of other comments here for details. But again: I laid out 2 possibilities explaining why the question might in fact not be new, nor the data, so I'm curious which of the 2 (or another) explains the situation you're talking about. You're saying it's nothing "special" but we're not discussing whether it's special, but whether it can be considered thinking. Apologies, with 'special' I did in fact mean 'thinking' | | |
| ▲ | chpatrick a day ago | parent [-] | | Sufficiently smart auto complete is indistinguishable from thinking, I don't think that means anything. |
|
|
|
|
|
|
|
|
| ▲ | satisfice 3 days ago | parent | prev | next [-] |
| I think you are the one dismissing evidence. The valid chains of reasoning you speak of (assuming you are talking about text you see in a “thinking model” as it is preparing its answer) are narratives, not the actual reasoning that leads to the answer you get. I don’t know what LLMs are doing, but only a little experimentation with getting it to describe its own process shows that it CAN’T describe its own process. You can call what a TI calculator does “thinking” if you want. But what people are interested in is human-like thinking. We have no reason to believe that the “thinking” of LLMs is human-like. |
| |
| ▲ | naasking 2 days ago | parent [-] | | > The valid chains of reasoning you speak of (assuming you are talking about text you see in a “thinking model” as it is preparing its answer) are narratives, not the actual reasoning that leads to the answer you get. It's funny that you think people don't also do that. We even have a term, "post hoc rationalization", and theories of mind suggest that our conscious control is a complete illusion, we just construct stories for decisions our subconscious has already made. |
|
|
| ▲ | josefx 2 days ago | parent | prev | next [-] |
| Counterpoint: The seahorse emoji. The output repeats the same simple pattern of giving a bad result and correcting it with another bad result until it runs out of attempts. There is no reasoning, no diagnosis, just the same error over and over again within a single session. |
| |
| ▲ | becquerel 2 days ago | parent | next [-] | | A system having terminal failure modes doesn't inherently negate the rest of the system. Human intelligences fall prey to plenty of similarly bad behaviours like addiction. | | |
| ▲ | josefx 2 days ago | parent [-] | | I never met an addicted person that could be reduced to a simple while(true) print("fail") loop. |
| |
| ▲ | throwaway-0001 2 days ago | parent | prev [-] | | You never had that coleague that says yes to everything and can’t get anything done? Same thing as seahorse. |
|
|
| ▲ | techblueberry 2 days ago | parent | prev | next [-] |
| Isn’t anthropomorphizing LLMs rather than understanding their unique presence in the world a “ lack of imagination and flexibility of thought”? It’s not that I can’t imagine applying the concept “thinking” to the output on the screen, I just don’t think it’s an accurate description. |
| |
| ▲ | heresie-dabord 2 days ago | parent [-] | | Yes, it's an example of domain-specific thinking. "The tool helps me write code, and my job is hard so I believe this tool is a genius!" The Roomba vacuumed the room. Maybe it vacuumed the whole apartment. This is good and useful. Let us not diminish the value of the tool. But it's a tool. The tool may have other features, such as being self-documenting/self-announcing. Maybe it will frighten the cats less. This is also good and useful. But it's a tool. Humans are credulous. A tool is not a human. Meaningful thinking and ideation is not just "a series of steps" that I will declaim as I go merrily thinking. There is not just a vast training set ("Reality"), but also our complex adaptability that enables us to test our hypotheses. We should consider what it is in human ideation that leads people to claim that a Roomba, a chess programme, Weizenbaum's Eliza script, the IBM's Jeopardy system Watson, or an LLM trained on human-vetted data is thinking. Train such a system on the erroneous statements of a madman and suddenly the Roomba, Eliza, IBM Watson (and these other systems) lose our confidence. As it is today, the confidence we have in these systems is very conditional. It doesn't matter terribly if code is wrong... until it does. Computers are not humans. Computers can do things that humans cannot do. Computers can do these things fast and consistently. But fundamentally, algorithms are tools. |
|
|
| ▲ | dmz73 a day ago | parent | prev | next [-] |
| Having seen LLMs so many time produce incoherent, nonsense, invalid answers to even simplest of questions I cannot agree with categorization of "thinking" or "intelligence" that applies to these models.
LLMs do not understand what they "know" or what they output. All they "know" is that based on training data this is most likely what they should output + some intentional randomization to make it seem more "human like".
This also makes it seem like they create new and previously unseen outputs but that could be achieved with simple dictionary and random number generator and no-one would call that thinking or intelligent as it is obvious that it isn't.
LLMs are better at obfuscating this fact by producing more sensible output than just random words.
LLMs can still be useful but they are a dead-end as far as "true" AI goes. They can and will get better but they will never be intelligent or think in the sense that most humans would agree those terms apply.
Some other form of hardware/software combination might get closer to AI or even achieve full AI and even sentience but that will not happen with LLMs and current hardware and software. |
|
| ▲ | ph4rsikal 2 days ago | parent | prev | next [-] |
| It might appear so, but then you could validate it with a simple test.
If the LLM would play a 4x4 Tic Tac Toe game, would the agent select the winning move 100% of all time or block a losing move 100% of the time?
If these systems were capable of proper reasoning, then they would find the right choice in these obvious but constantly changing scenarios without being specifically trained for it. [1] https://jdsemrau.substack.com/p/nemotron-vs-qwen-game-theory... |
|
| ▲ | didibus 3 days ago | parent | prev | next [-] |
| I guess it depends if you definite thinking thinking as chaining coherent reasoning sentences together 90-some% of the time. But if you define thinking as the mechanism and process we mentally undergo and follow mentally... I don't think we have any clue if that's the same. Do we also just vector-map attention tokens and predict the next with a softmax? I doubt, and I don't think we have any proof that we do. |
| |
| ▲ | aydyn 2 days ago | parent [-] | | We do know at the biochemical level how neurons work, and it isnt anything like huge matmuls. |
|
|
| ▲ | johnnienaked 3 days ago | parent | prev | next [-] |
| If you understand how they operate and you are reasonable and unbiased there is no way you could consider it thinking |
|
| ▲ | noiv 3 days ago | parent | prev | next [-] |
| Different PoV: You have a local bug and ask the digital hive mind for a solution, but someone already solved the issue and their solution was incorporated... LLMs are just very effficient at compressing billions of solutions into a few GB. Try to ask something no one ever came up with a solution so far. |
| |
| ▲ | brabel 2 days ago | parent [-] | | This argument comes up often but can be easily dismissed. Make up a language and explain it to the LLM like you would to a person. Tell it to only use that language now to communicate. Even earlier AI was really good at this. You will probably move the goal posts and say that this is just pattern recognition, but it still fits nicely within your request for something that no one ever came up with. | | |
| ▲ | emodendroket 2 days ago | parent | next [-] | | I haven't tried in a while but at least previously you could completely flummox Gemini by asking it to come up with some plausible English words with no real known meaning; it just kept giving me rare and funny-sounding actual words and then eventually told me the task is impossible. | | | |
| ▲ | noiv 2 days ago | parent | prev | next [-] | | Ask ChatGPT about ConLang. It knows. Inventing languages was solved a hundred years ago with Esperanto. | |
| ▲ | hitarpetar 2 days ago | parent | prev [-] | | > but it still fits nicely within your request for something that no one ever came up with YOU came up with it dawg |
|
|
|
| ▲ | ForHackernews 2 days ago | parent | prev | next [-] |
| But all those times the same system produces irrational gibberish don't count? GPT-5 will commonly make mistakes no thinking human could ever make. Human: I'm trying to get my wolf, sheep and cabbage across the river in this boat, but the wolf keeps eating the sheep or the sheep eats the cabbage Bot: You should put the sheep in the boat and take it across — if we delve into the biology of Canis lupus we discover that wolves don't eat cabbage! H: Ok, so that worked great so far, the sheep is on one side and the wolf/cabbage is on the other. B: Now, Option 1 is to bring the wolf across, or Option 2 you can bring the cabbage. I recommend (2) taking the cabbage as cabbages are smaller and easier to transport in a boat. H: But then the sheep eats the cabbage, right? Remember that? B: Exactly, that's sharp thinking. If you put the sheep and the cabbage together on the same side of the river, the sheep is sure to devour the cabbage. We need to not just separate sheep from cabbages — we need to separate cabbages from sheep! :rocketship: |
|
| ▲ | conartist6 3 days ago | parent | prev | next [-] |
| Yeah but if I assign it a long job to process I would also say that an x86 CPU is "thinking" about a problem for me. What we really mean in both cases is "computing," no? |
|
| ▲ | intended 2 days ago | parent | prev | next [-] |
| what sound does a falling tree make if no one is listening? I’ve asked LLMs to write code for me in fields I have little background knowledge, and then had to debug the whole thing after essentially having to learn the language and field. On the other hand, for things I am well versed in, I can debug the output and avoid entire swathes of failed states, by having a clear prompt. Its why I now insist that any discussion on GenAI projects also have the speaker mention the level of seniority they have ( proxy for S/W eng experience), Their familiarity with the language, the project itself (level of complexity) - more so than the output. I also guarantee - that most people have VERY weak express knowledge of how their brains actually work, but deep inherent reflexes and intuitions. |
|
| ▲ | NoMoreNicksLeft 3 days ago | parent | prev | next [-] |
| >Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking. If one could write a quadrillion-line python script of nothing but if/elif/else statements nested 1 million blocks deep that seemingly parsed your questions and produced seemingly coherent, sensible, valid "chains of reasoning"... would that software be thinking? And if you don't like the answer, how is the LLM fundamentally different from the software I describe? >Knee jerk dismissing the evidence in front of your eyes because There is no evidence here. On the very remote possibility that LLMs are at some level doing what humans are doing, I would then feel really pathetic that humans are as non-sapient as the LLMs. The same way that there is a hole in your vision because of a defective retina, there is a hole in your cognition that blinds you to how cognition works. Because of this, you and all the other humans are stumbling around in the dark, trying to invent intelligence by accident, rather than just introspecting and writing it out from scratch. While our species might someday eventually brute force AGI, it would be many thousands of years before we get there. |
| |
| ▲ | hattmall 3 days ago | parent | next [-] | | I write software that is far less complex and I consider it to be "thinking" while it is working through multiple possible permutations of output and selecting the best one. Unless you rigidly define thinking, processing, computing, it's reasonable to use them interchangeably. | | |
| ▲ | emodendroket 2 days ago | parent [-] | | To borrow a line from Dijkstra, the claim seems a bit like saying that a submarine is swimming. | | |
| ▲ | gkbrk 2 days ago | parent [-] | | I think most people would agree that submarines are swimming. | | |
| ▲ | emodendroket 2 days ago | parent | next [-] | | I think the opposite. In fact I (and Dijstrka) thought that was so obvious I didn’t feel the need to elaborate on how obviously absurd it was. | |
| ▲ | isaacremuant 2 days ago | parent | prev [-] | | No. Cars don't walk. Submarines don't swim. You can call it navigation if you want but words have meaning. > to move through water by moving the body or parts of the body. | | |
| ▲ | hackinthebochs 2 days ago | parent [-] | | Spinning propellers is "moving parts of the [submarines] body" | | |
| ▲ | isaacremuant 2 days ago | parent [-] | | No they aren't. Of course you cans also call it's sonar eyes but it isn't. Anthropomorphizing cars doesn't make them humans either. | | |
| ▲ | hackinthebochs 2 days ago | parent [-] | | Why would you think body only refers to flesh? | | |
| ▲ | emodendroket a day ago | parent [-] | | Even if I take the more expansive possible interpretation of “body” typically applied to vehicles, the propeller on the back of it isn’t part of the “body” and the “body” of a submarine is rigid and immobile. Is this an intellectual exercise for you or have you ever in your life heard someone say something like “the submarine swam through the water”? It’s so ridiculous I would be shocked to see it outside of a story intended for children or an obvious nonnative speaker of English. | | |
| ▲ | hackinthebochs a day ago | parent [-] | | >the propeller on the back of it isn’t part of the “body” and the “body” of a submarine is rigid and immobile. That's a choice to limit the meaning of the term to the rigid/immobile parts of the external boundary of an object. It's not obviously the correct choice. Presumably you don't take issue with people saying planes fly. The issue of submarines swimming seems analogous. >Is this an intellectual exercise for you or have you ever in your life heard someone say something like “the submarine swam through the water”? I don't think I've ever had a discussion about submarines with anyone, outside of the OceanGate disaster. But this whole approach to the issue seems misguided. With terms like this we should ask what the purpose behind the term is, i.e. it's intension (the concept), not the incidental extension of the term (the collection of things it applies to at some point in time). When we refer to something swimming, we mean that it is moving through water under its own power. The reference to "body" is incidental. | | |
| ▲ | emodendroket a day ago | parent [-] | | Which parts of the car does a "body shop" service? | | |
| ▲ | hackinthebochs a day ago | parent [-] | | Irrelevant, for the reasons mentioned | | |
| ▲ | emodendroket 13 hours ago | parent [-] | | It's not really a "choice" to use words how they are commonly understood but a choice to do the opposite. The point of Dijkstra's example is you can slap some term on a fundamentally different phenomenon to liken it to something more familiar but it confuses rather than clarifies anything. The point that "swim" is not very consistent with "fly" is true enough but not really helpful. It doesn't change the commonly understood meaning of "swim" to include spinning a propeller just because "fly" doesn't imply anything about the particular means used to achieve flight. | | |
| ▲ | hackinthebochs 11 hours ago | parent [-] | | >It's not really a "choice" to use words how they are commonly understood but a choice to do the opposite. I meant a collective choice. Words evolve because someone decides to expand their scope and others find it useful. The question here shouldn't be what do other people mean by a term but whether the expanded scope is clarifying or confusing. The question of whether submarines swim is a trivial verbal dispute, nothing of substance turns on its resolution. But we shouldn't dismiss the question of whether computers think by reference to the triviality of submarines swimming. The question we need to ask is what work does the concept of thinking do and whether that work is or can be applied to computers. This is extremely relevant in the present day. When we say someone thinks, we are attributing some space of behavioral capacities to that person. That is, a certain competence and robustness with managing complexity to achieve a goal. Such attributions may warrant a level of responsibility and autonomy that would not be warranted without it. A system that thinks can be trusted in a much wider range of circumstances than one that doesn't. That this level of competence has historically been exclusive to humans should not preclude this consideration. When some future AI does reach this level of competence, we should use terms like thinking and understanding as indicating this competence. | | |
| ▲ | emodendroket 4 hours ago | parent [-] | | This sub thread started on the claim that regular, deterministic code is “thought.” I would submit that the difference between deterministic code and human thought are so big and obvious that it is doing nothing but confusing the issue to start insisting on this. | | |
| ▲ | hackinthebochs 3 hours ago | parent [-] | | I'm not exactly sure what you mean by deterministic code but I do think there is an obvious distinction between typical code people write and what human minds do. The guy upthread is definitely wrong in thinking that, e.g. any search or minimax algorithm is thinking. But its important to understand what this distinction is so we can spot when it might no longer apply. To make a long story short, the distinction is that typical programs don't operate on the semantic features of program state, just on the syntactical features. We assign a correspondence with the syntactical program features and their transformations to the real-world semantic features and logical transformations on them. The execution of the program then tells us the outcomes of the logical transformations applied to the relevant semantic features. We get meaning out of programs because of this analogical correspondence. LLMs are a different computing paradigm because they now operate on semantic features of program state. Embedding vectors assign semantic features to syntactical structures of the vector space. Operations on these syntactical structures allow the program to engage with semantic features of program state directly. LLMs engage with the meaning of program state and alter its execution accordingly. It's still deterministic, but its a fundamentally more rich programming paradigm, one that bridges the gap between program state as syntactical structures and the meaning they represent. This is why I am optimistic that current or future LLMs should be considered properly thinking machines. | | |
| ▲ | emodendroket 3 hours ago | parent [-] | | LLMs are not deterministic at all. The same input leads to different outputs at random. But I think there’s still the question if this process is more similar to thought or a Markov chain. | | |
| ▲ | hackinthebochs 2 hours ago | parent [-] | | They are deterministic in the sense that the inference process scores every word in the vocabulary in a deterministic manner. This score map is then sampled from according to the temperature setting. Non-determinism is artificially injected for ergonomic reasons. >But I think there’s still the question if this process is more similar to thought or a Markov chain. It's definitely far from a Markov chain. Markov chains treat the past context as a single unit, an N-tuple that has no internal structure. The next state is indexed by this tuple. LLMs leverage the internal structure of the context which allows a large class of generalization that Markov chains necessarily miss. |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
| ▲ | helloplanets 3 days ago | parent | prev [-] | | 10^15 lines of code is a lot. We would pretty quickly enter the realm of it not having much to do with programming and more about just treating the LOC count as an amount of memory allocated to do X. How much resemblance does the information in the conditionals need to have with the actual input, or can they immediately be transformed to a completely separate 'language' which simply uses the string object as its conduit? Can the 10^15 lines of code be generated with an external algorithm, or is it assumed that I'd written it by hand given an infinitely long lifespan? |
|
|
| ▲ | raincole 3 days ago | parent | prev | next [-] |
| I'd represent the same idea but in a different way: I don't know what the exact definition of "thinking" is. But if a definition of thinking rejects the possibility of that current LLMs think, I'd consider that definition useless. |
| |
| ▲ | didibus 3 days ago | parent [-] | | Why would it be useless? Generally thinking has been used to describe the process human follow in their brains when problem solving. If the Palms do not follow that process, they are not thinking. That doesn't mean they cannot solve problems using other mechanisms, they do, and we understand those mechanisms much better than we do human thinking. |
|
|
| ▲ | ryanackley 2 days ago | parent | prev | next [-] |
| I think we can call it "thinking" but it's dangerous to anthropomorphize LLMs. The media and AI companies have an agenda when doing so. |
|
| ▲ | outworlder 2 days ago | parent | prev | next [-] |
| They may not be "thinking" in the way you and I think, and instead just finding the correct output from a really incredibly large search space. > Knee jerk dismissing the evidence in front of your eyes Anthropomorphizing isn't any better. That also dismisses the negative evidence, where they output completely _stupid_ things and make mind boggling mistakes that no human with a functioning brain would do. It's clear that there's some "thinking" analog, but there are pieces missing. I like to say that LLMs are like if we took the part of our brain responsible for language and told it to solve complex problems, without all the other brain parts, no neocortex, etc. Maybe it can do that, but it's just as likely that it is going to produce a bunch of nonsense. And it won't be able to tell those apart without the other brain areas to cross check. |
|
| ▲ | jimbohn 2 days ago | parent | prev | next [-] |
| It's reinforcement learning applied to text, at a huge scale. So I'd still say that they are not thinking, but they are still useful. The question of the century IMO is if RL can magically solve all our issues when scaled enough. |
|
| ▲ | 3 days ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | 2 days ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | flanked-evergl 2 days ago | parent | prev | next [-] |
| "Convince" the stock Claude Sonnet 4.5 that it's a sentient human being hooked up to Neuralink and then tell me again it's thinking. It's just not. |
|
| ▲ | camgunz 2 days ago | parent | prev | next [-] |
| Then the only thing I have to ask you is: what do you think this means in terms of how we treat LLMs? If they think, that is, they have cognition (which of course means they're self aware and sentient, how can you think and refer to yourself and not be these things), that puts them in a very exclusive club. What rights do you think we should be affording LLMs? |
|
| ▲ | hyperbovine 2 days ago | parent | prev | next [-] |
| Code gen is the absolute best case scenario for LLMs though: highly structured language, loads of training data, the ability to automatically error check the responses, etc. If they could mimic reasoning anywhere it would be on this problem. I'm still not convinced they're thinking though because they faceplant on all sorts of other things that should be easy for something that is able to think. |
|
| ▲ | fennecbutt 2 days ago | parent | prev | next [-] |
| Thinking as in capable of using basic reasoning and forming chains of logic and action sequences for sure. Ofc we both understand that neither of us are trying to say we think it can think in the human sense at this point in time. But oh boy have I also seen models come up with stupendously dumb and funny shit as well. |
|
| ▲ | khafra 3 days ago | parent | prev | next [-] |
| "Consciousness" as in subjective experience, whatever it is we mean by "the hard problem," is very much in doubt. But "self-awareness," as in the ability to explicitly describe implicit, inner cognitive processes? That has some very strong evidence for it: https://www.anthropic.com/research/introspection |
|
| ▲ | belter 2 days ago | parent | prev | next [-] |
| Apparent reasoning can emerge from probabilistic systems that simply reproduce statistical order not genuine understanding. Weather models sometimes “predict” a real pattern by chance, yet we don’t call the atmosphere intelligent. If LLMs were truly thinking, we could enroll one at MIT and expect it to graduate, not just autocomplete its way through the syllabus or we
could teach one how to drive. |
|
| ▲ | triyambakam 3 days ago | parent | prev | next [-] |
| >
Having seen LLMs so many times produce coherent, sensible and valid chains of reasoning to diagnose issues and bugs in software I work on, I am at this point in absolutely no doubt that they are thinking. People said the same thing about ELIZA > Consciousness or self awareness is of course a different question, Then how do you define thinking if not a process that requires consciousness? |
| |
| ▲ | lordnacho 2 days ago | parent [-] | | Why would it require consciousness, when we can't even settle on a definition for that? |
|
|
| ▲ | lmganon 2 days ago | parent | prev | next [-] |
| Is this model thinking too? https://huggingface.co/PantheonUnbound/Satyr-V0.1-4B |
|
| ▲ | hitarpetar 2 days ago | parent | prev | next [-] |
| > The world is full of bizarre wonders and this is just one more to add to the list. https://youtu.be/_-agl0pOQfs?si=Xiyf0InqtjND9BnF |
|
| ▲ | mlsu 3 days ago | parent | prev | next [-] |
| They remind me of the apparitions in Solaris. They have this like mechanical, almost player-piano like quality to them. They both connect with and echo us at the same time. It seems crazy to me and very intellectually uncreative to not think of this as intelligence. |
|
| ▲ | tengbretson 2 days ago | parent | prev | next [-] |
| Too many people place their identity in their own thoughts/intellect. Acknowledging what the LLMs are doing as thought would basically be calling them human to people of that perspective. |
|
| ▲ | darthvaden 2 days ago | parent | prev | next [-] |
| If AI is thinking if slavery is bad then how can somebody own AI. How can investors can shares from AI profits? We are ok with slavery now. Ok i will have two black slaves now. Who can ask me? Why shld that be illegal? |
| |
| ▲ | Manfred 2 days ago | parent | next [-] | | Yikes, you're bypassing thousands of years of oppression, abuse, and human suffering by casually equating a term that is primarily associated with a human owning another human to a different context. There is a way to discuss if keeping intelligent artificial life under servitude without using those terms, especially if you're on a new account. | | |
| ▲ | Grossenstein a day ago | parent [-] | | slavery is slavery does not mean it is AI or human. if slavery is ok then the question is who can own a slave. the answer is coporates like open ai. which is terrible for humanity and the universe |
| |
| ▲ | ndsipa_pomu 2 days ago | parent | prev [-] | | I presume you are aware that the word "robot" is taken from a Czech word (robota) meaning "slave" |
|
|
| ▲ | uberduper 2 days ago | parent | prev | next [-] |
| Sometimes I start thinking our brains work the same way as an LLM does when it comes to language processing. Are we just using probability based on what we already know and the context of the statement we're making to select the next few words? Maybe we apply a few more rules than an LLM on what comes next as we go. We train ourselves on content. We give more weight to some content than others. While listening to someone speak, we can often predict their next words. What is thinking without language? Without language are we just bags of meat reacting to instincts and emotions? Are instincts and emotions what's missing for AGI? |
|
| ▲ | lordnacho 2 days ago | parent | prev | next [-] |
| I agree with you. If you took a Claude session into a time machine to 2019 and called it "rent a programmer buddy," how many people would assume it was a human? The only hint that it wasn't a human programmer would be things where it was clearly better: it types things very fast, and seems to know every language. You can set expectations in the way you would with a real programmer: "I have this script, it runs like this, please fix it so it does so and so". You can do this without being very precise in your explanation (though it helps) and you can make typos, yet it will still work. You can see it literally doing what you would do yourself: running the program, reading the errors, editing the program, and repeating. People need to keep in mind two things when they compare LLMs to humans: you don't know the internal process of a human either, he is also just telling you that he ran the program, read the errors, and edited. The other thing is the bar for thinking: a four-year old kid who is incapable of any of these things you would not deny as a thinking person. |
| |
| ▲ | kkapelon 2 days ago | parent [-] | | > If you took a Claude session into a time machine to 2019 and called it "rent a programmer buddy," how many people would assume it was a human? Depends on the users. Junior devs might be fooled. Senior devs would quickly understand that something is wrong. |
|
|
| ▲ | xhkkffbf 2 days ago | parent | prev | next [-] |
| Instead of thinking, "Wow. AIs are smart like humans", maybe we should say, "Humans are dumb like matrix multiplication?" |
|
| ▲ | libraryatnight 2 days ago | parent | prev | next [-] |
| If you're sensitive to patterns and have been chronically online for the last few decades it's obvious they are not thinking. |
|
| ▲ | Zardoz84 2 days ago | parent | prev | next [-] |
| Having seen parrots so many times produce coherent, sensible, and valid chains of sounds and words, I am at this point in absolutely no doubt that they are thinking. |
| |
|
| ▲ | lispybanana 2 days ago | parent | prev | next [-] |
| Would they have diagnosed an issue if you hadn't presented it to them? Life solves problems itself poses or collides with. Tools solve problems only when applied. |
|
| ▲ | yawpitch 2 days ago | parent | prev | next [-] |
| You’re assuming the issues and bugs you’ve been addressing don’t already exist, already encoding human chain of reasoning, in the training data. |
|
| ▲ | donkeybeer 2 days ago | parent | prev | next [-] |
| Its overt or unaware religion. The point when you come down to the base of it is that these people believe in "souls". |
|
| ▲ | IAmGraydon 2 days ago | parent | prev | next [-] |
| >Knee jerk dismissing the evidence in front of your eyes because you find it unbelievable that we can achieve true reasoning via scaled matrix multiplication is understandable, but also betrays a lack of imagination and flexibility of thought. You go ahead with your imagination. To us unimaginative folks, it betrays a lack of understanding of how LLMs actually work and shows that a lot of people still cannot grasp that it’s actually an extremely elaborate illusion of thinking. |
|
| ▲ | hagbarth 2 days ago | parent | prev | next [-] |
| I'm not so sure. I, for one, do not think purely by talking to myself. I do that sometimes, but a lot of the time when I am working through something, I have many more dimensions to my thought than inner speech. |
|
| ▲ | conartist6 2 days ago | parent | prev | next [-] |
| So an x86 CPU is thinking? So many times I've seen it produce sensible, valid chains of results. Yes, I see evidence in that outcome that a person somewhere thought and understood. I even sometimes say that a computer is "thinking hard" about something when it freezes up. ...but ascribing new philosophical meaning to this simple usage of the word "thinking" is a step too far. It's not even a new way of using the word! |
| |
| ▲ | gchamonlive 2 days ago | parent [-] | | You can't say for sure it is or it isn't thinking based solely on the substrate, because it's not known for sure if consciousness is dependent on the hardware it's running on -- for a lack of a better analogy -- to manifest, if it really needs an organic brain or if it could manifest in silicon based solutions. | | |
| ▲ | conartist6 2 days ago | parent [-] | | I agree. I'm just pointing out that the meaning of the word "think" already applied to the silicon substrate pre-ai, so just saying it's still applicable isn't that compelling. But yeah, I am fully willing to believe that a silicon based life form could think and be alive. i just don't think we're there. Yes this thing speaks using a passable imitation of the voices of PhDs and poets, but in a way a simulated annelid is more alive. |
|
|
|
| ▲ | smohare 2 days ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | absurd1st 2 days ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | veegee 2 days ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | ath3nd 3 days ago | parent | prev [-] |
| [dead] |