| ▲ | lowsong 2 days ago | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
I'm impressed that such a short post can be so categorically incorrect. > For years, despite functional evidence and scientific hints accumulating, certain AI researchers continued to claim LLMs were stochastic parrots > In 2025 finally almost everybody stopped saying so. There is still no evidence that LLMs are anything beyond "stochastic parrots". There is no proof of any "understanding". This is seeing faces in clouds. > I believe improvements to RL applied to LLMs will be the next big thing in AI. With what proof or evidence? Gut feeling? > Programmers resistance to AI assisted programming has lowered considerably. Evidence is the opposite, most developers do not trust it. https://survey.stackoverflow.co/2025/ai#2-accuracy-of-ai-too... > It is likely that AGI can be reached independently with many radically different architectures. There continues to be no evidence beyond "hope" that AGI is even possible, yet alone that Transformer models are the path there. > The fundamental challenge in AI for the next 20 years is avoiding extinction. Again, nothing more than a gut feeling. Much like all the other AI hype posts this is nothing more than "well LLMs sure are impressive, people say they're not, but I think they're wrong and we will make a machine god any day now". | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| ▲ | crystal_revenge a day ago | parent [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Strongly agree with this comment. Decoder-only LLMs (the ones we use) are literally Markov Chains, the only (and major) difference is a radically more sophisticated state representation. Maybe "stochastic parrot" is overly dismissive sounding, but it's not a fundamentally wrong understanding of LLMs. The RL claims are also odd because, for starters, RLHF is not "reinforcement learning" based on any classical definition of RL (which almost always involve an online component). And further, you can chat with anyone who has kept up with the RL field, and quickly realize that this is also a technology that still hasn't quite delivered on the promises it's been making (despite being an incredibly interesting area of research). There's no reason to speculate that RL techniques will work with "agents" where they have failed to achieve wide spread success in similar domains. I continue to be confused why smart, very technical people can't just talk about LLMs honestly. I personally think we'd have much more progress if we could have conversations like "Wow! The performance of a Markov Chain with proper state representation is incredible, let's understand this better..." rather than "AI is reasoning intelligently!" I get why non-technical people get caught up in AI hype discussions, but for technical people that understand LLMs it seems counter productive. Even more surprising to me is that this hype has completely destroyed any serious discussions of the technology and how to use it. There's so much oppurtunity lost around practical uses of incorporating LLMs into software while people wait for agents to create mountains of slop. | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||