| ▲ | didip 3 hours ago |
| Indeed. I don't understand why Hacker News is so dismissive about the coming of LLMs, maybe HN readers are going through 5 stages of grief? But LLM is certainly a game changer, I can see it delivering impact bigger than the internet itself. Both require a lot of investments. |
|
| ▲ | crystal_revenge 2 hours ago | parent | next [-] |
| > I don't understand why Hacker News is so dismissive about the coming of LLMs I find LLMs incredibly useful, but if you were following along the last few years the promise was for “exponential progress” with a teaser world destroying super intelligence. We objectively are not on that path. There is no “coming of LLMs”. We might get some incremental improvement, but we’re very clearly seeing sigmoid progress. I can’t speak for everyone, but I’m tired of hyperbolic rants that are unquestionably not justified (the nice thing about exponential progress is you don’t need to argue about it) |
| |
| ▲ | viraptor an hour ago | parent | next [-] | | > exponential progress First you need to define what it means. What's the metric? Otherwise it's very much something you can argue about. | |
| ▲ | aoeusnth1 2 hours ago | parent | prev [-] | | We're very clearly seeing exponential progress - even above trend, on METR, whose slope keeps getting revised to a higher and higher estimate each time. Explain your perspective on the objective evidence against exponential progress? | | |
| ▲ | llmslave2 an hour ago | parent [-] | | Pretty neat how this exponential progress hasn't resulted in exponential productivity. Perhaps you could explain your perspective on that? | | |
| ▲ | mgfist 30 minutes ago | parent | next [-] | | Because that requires adoption. Devs on hackernews are already the most up to date folks in the industry and even here adoption of LLMs is incredibly slow. And a lot of the adoption that does happen is still with older tech like ChatGPT or Cursor. | |
| ▲ | HPMOR an hour ago | parent | prev | next [-] | | I think this is an open question still and very interesting. Ilya discussed this on the Dwarkesh podcast. But the capabilities of LLMs is clearly exponential and perhaps super exponential. We went from something that could string together incoherent text in 2022 to general models helping people like Terrance Tao and Scott Aaronson write new research papers. LLMs also beat IMO and the ICPC. We have entered the John Henry era for intellectual tasks... | | |
| ▲ | llmslave2 31 minutes ago | parent [-] | | > But the capabilities of LLMs is clearly exponential and perhaps super exponential By what metric? |
| |
| ▲ | viraptor an hour ago | parent | prev | next [-] | | Writing the code itself was never the main bottleneck. Designing the bigger solution, figuring out tradeoffs, taking to affected teams, etc. takes as much time as it used to. But still, there's definitely a significant improvement in code production part in many areas. | |
| ▲ | aoeusnth1 an hour ago | parent | prev [-] | | It has! CLs/engineer increased by 10% this year. LLMs from late 2024 were nearly worthless as coding agents, so given they have quadrupled in capability since then (exponential growth, btw), it's not surprising to see a modestly positive impact on SWE work. Also, I'm noticing you're not explaining yourself :) | | |
| ▲ | llmslave2 28 minutes ago | parent [-] | | Hey, I'm not the OG commentator, why do I have to explain myself! :) When Fernando Alonso (best rookie btw) goes from 0-60 in 2.4 seconds in his Aston Martin, is it reasonable to assume he will near the speed of light in 20 seconds? |
|
|
|
|
|
| ▲ | viraptor an hour ago | parent | prev | next [-] |
| Based on quite a few comments recently, it also looks like many have tried LLMs in the past, but haven't seriously revisited either the modern or more expensive models. And I get it. Not everyone wants to keep up to date every month, or burn cash on experiments. But at the same time, people seem to have opinions formed in 2024. (Especially if they talk about just hallucinations and broken code - tell the agent to search for docs and fix stuff) I'd really like to give them Opus 4.5 as an agent to refresh their views. There's lots of complain about, but the world has moved on significantly. |
| |
|
| ▲ | zvolsky 2 hours ago | parent | prev | next [-] |
| The idea of HN being dismissive of impactful technology is as old as HN. And indeed, the crowd often appears stuck in the past with hindsight. That said, HN discussions aren't homogeneous, and as demonstrated by Karpathy in his recent blogpost "Auto-grading decade-old Hacker News", at least some commenters have impressive foresight: https://karpathy.bearblog.dev/auto-grade-hn/ |
|
| ▲ | cebert 3 hours ago | parent | prev | next [-] |
| Many people feel threatened by the rapid advancements in LLMs, fearing that their skills may become obsolete, and in turn act irrationally. To navigate this change effectively, we must keep open minds, keep adaptable, and embrace continuous learning. |
| |
| ▲ | chii 3 hours ago | parent | next [-] | | > in turn act irrationally it isn't irrational to act in self-interest. If LLM threatens someone's livelihood, it matters not that it helps humanity overall one bit - they will oppose it. I don't blame them. But i also hope that they cannot succeed in opposing it. | | |
| ▲ | Davidzheng 2 hours ago | parent [-] | | It's irrational to genuinely hold false beliefs about capabilities of LLMs. But at this point I assume around half of the skeptics are emotionally motivated anyway. | | |
| ▲ | jdhsgsvsbzbd 43 minutes ago | parent [-] | | As opposed to having skin in the game for llms and are blind to their flaws??? I'd assume that around half of the optimists are emotionally motivated this way. |
|
| |
| ▲ | nickphx 3 hours ago | parent | prev [-] | | rapid advancements in what? hallucinations..? FOMO marketing? certainly nothing productive. |
|
|
| ▲ | asielen 2 hours ago | parent | prev | next [-] |
| It is an over correction because of all the empty promises of LLMs. I use Claude and chatgpt daily at work and am amazed at what they can do and how far they can come. BUT when I hear my executive team talk and see demos of "Agentforce" and every saas company becoming an AI company promising the world, I have to roll my eyes. The challenge I have with LLMs is they are great at creating first draft shiny objects and the LLMs themselves over promise. I am handed half baked work created by non technical people that now I have to clean up. And they don't realize how much work it is to take something from a 60% solution to a 100% solution because it was so easy for them to get to the 60%. Amazing, game changing tools in the right hands but also give people false confidence. Not that they are not also useful for non-technical people but I have had to spend a ton of time explaining to copywriters on the marketing team that they shouldn't paste their credentials into the chat even if it tells them to and their vibe coded app is a security nightmare. |
|
| ▲ | probably_wrong 2 hours ago | parent | prev | next [-] |
| Speaking for myself: because if the hype were to be believed we should have no relational databases when there's MongoDB, no need for dollars when there's cryptocoins, all virtual goods would be exclusively sold as NFTs, and we would be all driving self-driving cars by now. LLMs are being driven mostly by grifters trying to achieve a monopoly before they run out of cash. Under those conditions I find their promises hard to believe. I'll wait until they either go broke or stop losing money left and right, and whatever is left is probably actually useful. |
| |
| ▲ | simonw 2 hours ago | parent [-] | | The way I've been handling the deafening hype is to focus exclusively on what the models that we have right now can do. You'll note I don't mention AGI or future model releases in my annual roundup at all. The closest I get to that is expressing doubt that the METR chart will continue at the same rate. If you focus exclusively on what actually works the LLM space is a whole lot more interesting and less frustrating. | | |
|
|
| ▲ | vunderba 2 hours ago | parent | prev | next [-] |
| > I don't understand why Hacker News is so dismissive about the coming of LLMs. Eh. I wouldn’t be so quick to speak for the entirety of HN. Several articles related to LLMs easily hit the front page every single day, so clearly there are plenty of HN users upvoting them. I think you're just reading too much into what is more likely classic HN cynicism and/or fatigue. |
| |
| ▲ | ewoodrich 36 minutes ago | parent [-] | | Exactly. There was a stretch of 6 months or so right after ChatGPT was released where approximately 50% of front page posts at any given time were related to LLMs. And these days every other Show HN is some kind of agentic dev tool and Anthropic/OpenAI announcements routinely get 500+ comments in a matter of hours. |
|
|
| ▲ | Night_Thastus 2 hours ago | parent | prev | next [-] |
| LLMs hold some real utility. But that real utility is buried under a mountain of fake hype and over-promises to keep shareholder value high. LLMs have real limitations that aren't going away any time soon - not until we move to a new technology fundamentally different and separate from them - sharing almost nothing in common. There's a lot of 'progress-washing' going on where people claim that these shortfalls will magically disappear if we throw enough data and compute at it when they clearly will not. |
| |
| ▲ | Gigachad 2 hours ago | parent [-] | | Pretty much. What actually exists is very impressive. But what was promised and marketed has not been delivered. | | |
| ▲ | visarga 2 hours ago | parent | next [-] | | I think the missing ingredient is not something the LLMs lack, but something we as developers don't do - we need to constrain, channel, and guide agents by creating reactive test environments around them. Not vibes, but hard tests, they are the missing ingredient to coding agents. You can even use AI to write most of these tests but the end result depends on how well you structured your code to be testable. If you inherit 9000 tests from an existing project you can vibe code a replacement on your phone in a holiday, like Simon Willison's JustHTML port. We are moving from agents semi-randomly flailing around to constraint satisfaction. | |
| ▲ | coffeebeqn 2 hours ago | parent | prev | next [-] | | Yes and most of the investment has been kind of post-GPT4 betting that things will get exponentially more impressive | |
| ▲ | rustystump 2 hours ago | parent | prev [-] | | Markets never deliver. That isnt new, i do think llms are not far off from google in terms of impact. Search, as of today, is inferior to frontier models as a product. However, best case still misses expected returns by miles which is where the growsing comes from. Generative art/ai is still up in the air for staying power but id predict it isnt going away. |
|
|
|
| ▲ | snigsnog 3 hours ago | parent | prev | next [-] |
| The internet and smartphones were immediately useful in a million different ways for almost every person. AI is not even close to that level. Very to somewhat useful in some fields (like programming) but the average person will easily be able to go through their day without using AI. The most wide-appeal possibility is people loving 100%-AI-slop entertainment like that AI Instagram Reels product. Maybe I'm just too disconnected with normies but I don't see this taking off. Fun as a novelty like those Ring cam vids but I would never spend all day watching AI generated media. |
| |
| ▲ | raincole 2 hours ago | parent | next [-] | | The early internet and smartphones (the Japanese ones, not iPhone) were definitely not "immediately" adopted by the mass, unlike LLM. If "immediate" usefulness is the metric we measure, then the internet and smartphones are pretty insignificant inventions compared to LLM. (of course it's not a meaningful metric, as there is no clear line between a dumb phone and a smart phone, or a moderately sized language model and a LLM) | |
| ▲ | JumpCrisscross 3 hours ago | parent | prev | next [-] | | > AI is not even close to that level Kagi’s Research Assistant is pretty damn useful, particularly when I can have it poll different models. I remember when the first iPhone lacked copy-paste. This feels similar. (And I don’t think we’re heading towards AGI.) | |
| ▲ | nen-nomad 2 hours ago | parent | prev | next [-] | | ChatGPT has roughly 800 million weekly active users. Almost everyone around me uses it daily. I think you are underestimating the adoption. | |
| ▲ | SgtBastard 3 hours ago | parent | prev | next [-] | | … the internet was not immediately useful in a million different ways for almost every person. Even if you skip ARPAnet, you’re forgetting the Gopher days and even if you jump straight to WWW+email==the internet, you’re forgetting the mosaic days. The applications that became useful to the masses emerged a decade+ after the public internet and even then, it took 2+ decades to reach anything approaching saturation. Your dismissal is not likely to age well, for similar reasons. | | |
| ▲ | chii 3 hours ago | parent [-] | | the "usefulness" excuse is irrelevant, and the claim that phones/internet is "immediately useful" is just a post hoc rationalization. It's basically trying to find a reasonable reason why opposition to AI is valid, and is not in self-interest. The opposition to AI is from people who feel threatened by it, because it either threatens their livelihood (or family/friends'), and that they feel they are unable to benefit from AI in the same way as they had internet/mobile phones. |
| |
| ▲ | fragmede 2 hours ago | parent | prev | next [-] | | > The internet and smartphones were immediately useful in a million different ways for almost every person. AI is not even close to that level. Those are some very rosy glasses you've got on there. The nascent Internet took forever to catch on. It was for weird nerds at universities and it'll never catch on, but here we are. | |
| ▲ | staticassertion 3 hours ago | parent | prev | next [-] | | > Very to somewhat useful in some fields (like programming) but the average person will easily be able to go through their day without using AI. I know a lot of "normal" people who have completely replaced their search engine with AI. It's increasingly a staple for people. Smartphones were absolutely NOT immediately useful in a million different ways for almost every person, that's total revisionist history. I remember when the iPhone came out, it was AT&T only, it did almost nothing useful. Smartphones were a novelty for quite a while. | |
| ▲ | what-the-grump 2 hours ago | parent | prev [-] | | A year after the iPhone came out… it didn’t have an App Store, barely was able to play video, barely had enough power to last a day. You just don’t remember or were not around for it. A year after llms came out… are you kidding me? Two years? 10 years? Today, by adding an MCP server to wrap the same API that’s been around forever for some system, makes the users of that system prefer NLI over the gui almost immediately. |
|
|
| ▲ | 2 hours ago | parent | prev [-] |
| [deleted] |