▲ | ninetyninenine 2 days ago | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
I've been saying this to people. Tons of people don't realize that we have no idea how these things work. So questions like is it AGI? It is conscious? Where the questions themselves contain words with fuzzy and ill defined meanings are pointless. The core of the matter is: We don't know! So when someone says chatGPT could be conscious. Or when someone says it can't be conscious. We don't actually know! And we aren't even fully sure about the definition of consciousness. My problem with HN is that a ton of people here take the stance that we know how LLMs work and that we know definitively they aren't conscious and the people who say otherwise are alarmist and stupid. I think the fact of the matter is, if you're putting your foot down and saying LLMs aren't intelligent... you're wildly illogical and misinformed about the status quo of Artificial intelligence and a huge portion of the mob mentality on HN thinks this way. It's like there's these buzz phrases getting thrown around saying that the LLM is just a glorified auto complete (which it is) and people latch on to this buzz phrases and their entire understanding of LLMs becomes basically a composition of these buzz concepts like "transformers" and "LLMs" and "chain of thought" when in actuality real critical thinking about what's going on in these networks tells you we don't UNDERSTAND anything. Also the characterization in the article is mistaken. It says we understand LLMs in a limited way. Yeah sure. It's as limited as our understanding of the human brain. You know scientists found a way to stimulate the reward centers of the brain using electrodes and were able to cause a person to feel the utmost pleasure? The whole golden gate bridge thing is exactly that. You perturb something in the network and causes a semi predictable output. In the end we still generally don't get wtf is going on. We literally understand nothing. What we do understand is so minuscule compared with what we don't understand that it's negligible. | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | sirwhinesalot 2 days ago | parent [-] | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
This is such a weird take to me. We know exactly how LLMs and neural networks in general work. They're just highly scaled up versions of their smaller curve fitting cousins. For those we can even make pretty visualizations that show exactly what is happening as the network "learns". I don't mean "we can see parts of the brain light up", I mean "we can see the cuts and bends each ReLU is doing to the input". We built these things, we know exactly how they work. There's no surprise beyond just how good prediction accuracy gets with a big enough model. Deep Neural Networks are also a very different architecture from what is going on in our brains (which work more like Spiking Neural Networks) and our brains don't do backpropagation, so you can't even make direct comparisons. | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
|