| ▲ | gdubs 4 days ago |
| AI has been improving at a very rapid pace, which means that a lot of people have really outdated priors. I see this all the time online where people are dismissive about AI in a way that suggests it's been a while since they last checked-in on the capabilities of models. They wrote off the coding ability of ChatGPT on version 3.5, for instance, and have missed all the advancements that have happened since. Or they talk about hallucination and haven't tried Deep Research as an alternative to traditional web-search. Then there's a tendency to be so 'anti' that there's an assumption that anyone reporting that the tools are accomplishing truly impressive and useful things must be an 'AI booster' or shill. Or they assume that person must not have been a very good engineer in the first place, etc. Really is one of those examples of the quote, "In the beginner's mind there are many possibilities, but in the expert's mind there are few." It's a rapidly evolving field, and unless you actually spend some time kicking the tires on the models every so often, you're just basing your opinions on outdated experiences or what everyone else is saying about it. |
|
| ▲ | jdoliner 4 days ago | parent | next [-] |
| I feel like I see these two opposite behaviors. People who formed an opinion about AI from an older model and haven't updated it. And people who have an opinion about what AI will be able to do in the future and refuse to acknowledge that it doesn't do that in the present. And often when the two are arguing it's tricky to tell which is which, because whether or not it does something isn't totally black and white, there's some things it can sometimes do, which you can argue either way about that being in its capabilities or not. |
| |
| ▲ | forgotTheLast 4 days ago | parent | next [-] | | I.e. people who look at f(now) and assume it'll be like this forever against people who look at f'(now) and assume it'll improve like this forever | | |
| ▲ | arcastroe 4 days ago | parent [-] | | What is f''(now) looking like? | | |
| ▲ | mxkopy 3 days ago | parent [-] | | Very small. There’s very little fundamental research into AI compared to neural networks from what I can tell |
|
| |
| ▲ | nchmy 4 days ago | parent | prev [-] | | Another very significant cohort is people who formed a negative opinion without even the slightest interest in genuinely trying to learn how to use it (or even trying at all) |
|
|
| ▲ | thegrim33 4 days ago | parent | prev | next [-] |
| To play devil's advocate, how is your argument not a 'no true scottsman' argument? As in, "oh, they had a negative view of X, well that's of course because they weren't testing the new and improved X2 model which is different". Fast forward a year .. "Oh, they have a negative view on X2, well silly them, they need to be using the Y24 model, that's where it's at, the X2 model isn't good anymore". Fast forward a year .. ad infinitum. Are the models that exist today a "true scottsman" for you? |
| |
| ▲ | xwowsersx 4 days ago | parent | next [-] | | It's not a No True Scotsman. That fallacy redefines the group to dismiss counterexamples. The point here is different: when the thing itself keeps changing, evidence from older versions naturally goes stale. Criticisms of GPT-3.5 don't necessarily hold against GPT-4, just like reviews of Windows XP don't apply to Windows 11. | | |
| ▲ | cmiles74 4 days ago | parent | next [-] | | IMHO, by placing people with a negative attitude toward AI products under the guise "their priors are outdated" you effectively negate any arguments from those people. That is, because their priors are outdated their counterexamples may be dismissed. That is, indeed, the no true Scotsman! | | |
| ▲ | ludwik 4 days ago | parent [-] | | I don’t see a claim that anyone with a negative attitude toward AI shouldn’t be listened to because it automatically means that they formed their opinion on older models. The claim was simply that there’s a large cohort of people who undervalue the capabilities of language models because they formed their views while evaluating earlier versions. | | |
| ▲ | gmm1990 4 days ago | parent | next [-] | | I wouldn’t think gpt5 is any better than the previous chat gpt. I know it’s a silly example but I was trying to trip it up with the 8.6-8.11 and it got it right .49 but then it said the opposite of 8.6 - 8.12 was -.21. I just don’t see that much of a difference coding either with Claude 4 or Gemini 2.5 pro. Like they’re all fine but the difference isn’t changing anything in what I use them for. Maybe people are having more success with the agent stuff but in my mind it’s not that different than just forking a GitHub repo that already does what you’re “building” with the agent. | |
| ▲ | barrell 4 days ago | parent | prev [-] | | Yes but almost definitionally that is everyone who did not find value from LLMs. If you don’t find value from LLMs, you’re not going to use them all the time. The only people you’re excluding are the people who are forced to use it, and the random sampling of people who happened to try it recently. So it may have been accidental or indirectly, but yes, no true Scotsman would apply to your statement. |
|
| |
| ▲ | crote 4 days ago | parent | prev [-] | | > The point here is different: when the thing itself keeps changing, evidence from older versions naturally goes stale. Yes, but the claims do not. When the hypemen were shouting that GPT-3 was near-AGI, it still turned out to be absolute shit. When the hypemen were claiming that GPT-3.5 was thousands of times better than GPT-3 and beating all highschool students, it turned out to be a massive exaggeration. When the hypemen claimed that GPT-4 was a groundbreaking innovation and going to replace every single programmer, it still wasn't any good. Sure, AI is improving. Nobody is doubting that. But you can only claim to have a magical unicorn so many times before people stop believing that this time you might have something different than a horse with an ice cream cone glued to its head. I'm not going to waste a significant amount of my time evaluating Unicorn 5.0 when I already know I'll almost certainly end up disappointed. Perhaps it'll be something impressive in a decade or two, but in the meantime the fact that Big Tech keeps trying to shove it down my throat even when it clearly isn't ready yet is a pretty good indicator to me that it is still primarily just a hype bubble. | | |
| ▲ | trinsic2 4 days ago | parent [-] | | Its funny how the hype-train is not responding to any real criticisms about the false predictions and carrying on with the false narrative of AI. I agree it will probably be something in a decade, but right now, it has some interesting concepts but I do notice upon successive iterations of chat responses that its got a ways to go. It remind me of Tesla car owners buying into the self-driving terminology. Yes the drive assistant technology has improved quite a bit since cruise control, but its a far cry from self-driving. |
|
| |
| ▲ | vlovich123 4 days ago | parent | prev [-] | | How is that different than the models today are actually usable for non trivial things and more capable than yesterdays and it’s also true that tomorrow’s models will also probably be more capable than today’s? For example, I dismissed AI three years ago because it couldn’t do anything I needed it to. Today I use it for certain things and it’s not quite capable of other things. Tomorrow it might be capable of a lot more. Yes, priors have to be updated when the ground truth changes and the capabilities of AI change rapidly. This is how chess engines on supercomputers were competitive in the 90s then hybrid systems became the leading edge competitive and then machines took over for good and never looked back. | | |
| ▲ | Eggpants 4 days ago | parent [-] | | It’s not that the LLMs are better, it’s the internal tools/functions being called that do the actual work are better. They didn’t spend millions to retrain a model to statistically output the number of r’s in strawberry, but just offloaded that trivial question to a function call. So I would say the overall service provided is better than it was, thanks to functions being built based on user queries, but not the actual LLM models themselves. | | |
| ▲ | vlovich123 4 days ago | parent | next [-] | | LLMs are definitely better quality today than 3 years ago at codegen quality - there’s quantitative benchmarks as well as for me my personal qualitative experience (given the gaming that companies engage in). It is also true that the tooling and context management has gotten more sophisticated (often using models by the way). That doesn’t negate that the models themselves have gotten better at reliable tool calling so that the LLM is driving more of the show rather than purpose built coordination into the LLM and that the codegen quality is higher than it used to be. | |
| ▲ | int_19h 3 days ago | parent | prev [-] | | This is a good example of making statements that are clearly not based in fact. Anyone who works with those models knows full well what a massive gap there is between e.g. GPT 3.5 and Opus 4.1 that has nothing to do with the ability to use tools. |
|
|
|
|
| ▲ | Mars008 4 days ago | parent | prev | next [-] |
| There is another big and growing group: charlatans (influencers). People who don't know much but make bold statements, select 'proof' cases. Just to get attention. There are many of them on youtube. When you someone on thumbnail making faces this is most likely it. |
| |
| ▲ | trinsic2 4 days ago | parent | next [-] | | Here[0] is a perfect example of this. There are so many youtubers making videos about the future of AI as a dooms-day prediction. Its kind of irresponsible actually. These youtubers read a book on the the down fall of humanity because of AGI. Many of these authors seem like they are repeating the Terminator/Skynet themes. Because of all this false information, It's hard to believe anything that is being said about the future of AI on youtube now. [0]: https://www.youtube.com/watch?v=5KVDDfAkRgc | |
| ▲ | resource0x 4 days ago | parent | prev [-] | | > There are many of them on youtube. Not as many as on HN. "Influencers" have agendas and the stream of income, or other self-interest. HN always comes off as a monolith, on any subject. Counter-arguments get ignored and downvoted to oblivion. | | |
| ▲ | jkubicek 4 days ago | parent | next [-] | | I’m spending a lot of time on LinkedIn because my team is hiring and, boy oh boy, LinkedIn is terminally infested with AI influencers. It’s a hot mess. | |
| ▲ | tim333 4 days ago | parent | prev [-] | | If you type "AI" into the youtube search bar it's quite impressive. I think they win. |
|
|
|
| ▲ | barrell 4 days ago | parent | prev | next [-] |
| There are also a bunch of us who do kick the tires very often and are consistently underwhelmed. There are also those of us who have used them substantially, and seen the damage that causes to a codebase in the long run (in part due to the missing gains of having someone who understands the codebase). There are also those of us who just don’t like the interface of chatting with a robot instead of just solving the problem ourselves. There are also those of us who find each generation of model substantially worse than the previous generation, and find the utility trending downwards. There are also those of us who are concerned about the research coming out about the effects of using LLMs on your brain and cognitive load. There are also those of us who appreciate craft, and take pride in what we do, and don’t find that same enjoyment/pride in asking LLMs to do it. There are also those of us who worry about offloading our critical thinking to big corporations, and becoming dependent on a pay-to-play system, that is current being propped up by artificially lowered prices, with “RUG PULL” written all over them. There are also those of us who are really concerned about the privacy issues, and don’t trust companies hundreds of billions of dollars in debt to some of the least trust worth individuals with that data. Most of these issues don’t require much experience with the latest generation. I don’t think the intention of your comment was to stir up FUD, but I feel like it’s really easy for people to walk away with that from this sort of comment, so I just wanted to add my two cents and tell people they really don’t need to be wasting their time every 6 weeks. They’re really not missing anything. Can you do more than a few weeks ago? Sure? Maybe? But I can also do a lot more than I was able to a few weeks ago as well not using an LLM. I’ve learned and improved myself. Chances are if you’re not already using an LLM it’s because you don’t like it, or don’t want to, and that’s really ok. If AGSI comes out in a few months, all the time you would have invested now would be out of date anyways. There’s really no rush or need to be tapped in. |
| |
| ▲ | bigstrat2003 4 days ago | parent | next [-] | | > There are also a bunch of us who do kick the tires very often and are consistently underwhelmed. Yep, this is me. Every time people are like "it's improved so much" I feel like I'm taking crazy pills as a result. I try it every so often, and more often than not it still has the same exact issues it had back in the GPT-3 days. When the tool hasn't improved (in my opinion, obviously) in several years, why should I be optimistic that it'll reach the heights that advocates say it will? | | |
| ▲ | barrell 4 days ago | parent [-] | | haha I have to laugh because I’ve probably said “I feel like I’m taking crazy pills” at least 20 times this week (I spent a day using cursor with the new GPT and was thoroughly, thoroughly unimpressed). I’m open to programming with LLMs, and I’m entirely fine with people using them and I’m glad people are happy. But this insistence that progress is so crazy that you have to be tapped in at all times just irks me. LLM models are like iPhones. You can skip a couple versions it’s fine, you will have the new version at the same time with all the same functionality as everyone else buying one every year. | | |
| ▲ | energy123 4 days ago | parent [-] | | > new GPT Another sign tapping is needed. > AI is exceptional for coding! [high-compute scaffold around multiple instances / undisclosed IOI model / AlphaEvolve] > AI is awesome for coding! [Gpt-5 Pro] > AI is somewhat awesome for coding! ["gpt-5" with verbosity "high" and effort "high"] > AI is a pretty good at coding! [ChatGPT 5 Thinking through a Pro subscription with Juice of 128] > AI is mediocre at coding! [ChatGPT 5 Thinking through a Plus subscription with a Juice of 64] > AI sucks at coding! [ChatGPT 5 auto routing] | | |
| ▲ | bluefirebrand 3 days ago | parent [-] | | Yeah, frankly if you have the free time to dig through all of that to find the best models or whatever for your use cases, good on you I have code to write | | |
|
|
| |
| ▲ | libraryofbabel 4 days ago | parent | prev [-] | | There’s really three points mixed up in here. 1) LLMs are controlled by BigCorps who don’t have user’s best interests at heart. 2) I don’t like LLMs and don’t use them because they spoil my feeling of craftsmanship. 3) LLMs can’t be useful to anyone because I “kick the tires” every so often and am underwhelmed. (But what did you actually try? Do tell.) #1 is obviously true and is a problem, but it’s just capitalism. #2 is a personal choice, you do you etc., but it’s also kinda betting your career on AI failing. You may or may not have a technical niche where you’ll be fine for the next decade, but would you really in good conscience recommend a juniorish web dev take this position? #3 is a rather strong claim because it requires you to claim that a lot of smart reasonable programmers who see benefits from AI use are deluded. (Not everyone who says they get some benefit from AI is a shill or charlatan.) | | |
| ▲ | barrell 4 days ago | parent [-] | | How exactly am I betting my career on LLMs failing? The inverse is definitely true — going all in on LLMs feels like betting on the future success of LLMs. However not using LLMs to program today is not betting on anything, except maybe myself, but even that’s a stretch. After all, I can always pick up LLMs in the future. If a few weeks is long enough for all my priors to become stale, why should I have to start now? Everything I learn will be out of date in a few weeks. Things will only be easier to learn 6, 12, 18 months from now. Also no where in my post did I say that LLMs can’t be useful to anyone. In fact I said the opposite. If you like LLMs or benefit from them, then you’re probably already using them, in which case I’m not advocating anyone stop. However there are many segments of people who LLMs are not for. No tool is a panacea. I’m just trying to nip and FUD in the butt. There are so many demands for our attention in the modern world to stay looped in and up to date on everything; I’m just here saying don’t fret. Do what you enjoy. LLMs will be here in 12 months. And again in 24. And 36. You don’t need to care now. And yes I mentor several juniors (designers and engineers). I do not let them use LLMs for anything and actively discourage them from using LLMs. That is not what I’m trying to do in this post, but for those whose success I am invested in, who ask me for advice, I quite confidently advise against it. At least for now. But that is a separate matter. EDIT: My exact words from another comment in this thread prior to your comment: > I’m open to programming with LLMs, and I’m entirely fine with people using them and I’m glad people are happy. | | |
| ▲ | saltcured 4 days ago | parent [-] | | I wonder, what drives this intense FOMO ideation about AI tools as expressed further upthread? How does someone reconcile a faith that AI tooling is rapdily improving with that contradictory belief that there is some permanent early-adopter benefit? | | |
| ▲ | bluefirebrand 3 days ago | parent [-] | | I think the early adopter at all costs mentality is being driven by marketing and sales, not any rational reason to need to be ahead of the curve I agree very strongly with the poster above yours: If these tools are so good and so easy to use then I will learn them at that time Otherwise the idea that they are saving me time is likely just hype and not reality, which matches my experience |
|
|
|
|
|
| ▲ | dmead 4 days ago | parent | prev | next [-] |
| Is there anything you can tell me that will help me drop the nagging feeling that gradient descent trained models will just never be good? I understand all of what you said, but I can't get over that fact that the term AI is being used for these architectures. It seems like the industry is just trying to do a cool parlor trick in convincing the masses this is somehow AI from science fiction. Maybe I'm being overly cynical, but a lot of this stinks. |
| |
| ▲ | atleastoptimal 4 days ago | parent | next [-] | | The thing is AI is already "good" for a lot of things. It all depends on your definition of "good" and what you require of an AI model. It can do a lot of things that are generally very effective. High reliability semantic parsing from images is just one thing that modern LLM's are very reliable at. | | |
| ▲ | dmead 4 days ago | parent [-] | | You're right. I use it for api documentation and showing use cases, especially in languages i don't use often. but this other attribution people are doing- that it's going to achieve (the marketing term) AGI and everything will be awesome is clearly bullshit. |
| |
| ▲ | Zacharias030 4 days ago | parent | prev | next [-] | | Wouldn’t you say that now, finally, what people call AI combines subsymbolic systems („gradient descent“) with search and with symbolic systems (tool calls)? I had a professor in AI who was only working on symbolic systems such as SAT-solvers, Prolog etc. and the combination of things seems really promising. Oh, and what would be really nice is another level of memory or fast learning ability that goes beyond burning in knowledge through training alone. | | |
| ▲ | dmead 4 days ago | parent [-] | | I had such a professor as well, but those people used to use the more accurate term "machine learning". There was also wide understanding that those architectures were trying to imitate small bits of what we understood was happening in the brain (see marvin minsky's perceptron etc). The hope was, as I understood it that there would be some breakthrough in neuroscience that would let the computer scientists pick up the torch and simulate what we find in nature. None of that seems to be happening anymore and we're just interested in training enough to fool people. "AI" companies investing in brain science would convince me otherwise. At this point they're just trying to come up with the next money printing machine. | | |
| ▲ | app134 4 days ago | parent [-] | | You asked earlier if you were being overly cynical, and I think the answer to that is "yes" We are indeed simulating what we find in nature when we create neural networks and transformers, and AI companies are indeed investing heavily in BCI research. ChatGPT can write an original essay better than most of my students. Its also artificial. Is that not artificial intelligence? | | |
| ▲ | dmead 4 days ago | parent [-] | | It is not intelligent. Hiding the training data behind gradient descent and then making attributions to the program that responds using this model is certainly artificial though. This analogy just isn't holding water. | | |
| ▲ | tim333 4 days ago | parent [-] | | Can't you judge on the results though rather than saying AI isn't intelligent because it uses gradient descent and biology is intelligent because it uses wet neurons? | | |
| ▲ | Zacharias030 25 minutes ago | parent [-] | | I strongly believe that our concept of intelligence is like the „god of the gaps“ [0]. Intelligent is only what we haven’t yet explained. Chess computers surely must be intelligent, but then deep blue was „just search“. Go computers surely must have intelligence because it requires intuition and search is intractable, but then it’s „just CNN based pattern matching“. Writing essays surely requires intelligence, because of the creativity, but then it‘s actually just a „stochastic parrot“. We keep attributing intelligence to what is currently out of reach even as this set is rapidly shrinking before out eyes. It would be better to say that intelligence is an emergent phenomenon and that behavior that seems intelligent is intelligent. [0] https://en.m.wikipedia.org/wiki/God_of_the_gaps |
|
|
|
|
| |
| ▲ | int_19h 3 days ago | parent | prev [-] | | > It seems like the industry is just trying to do a cool parlor trick in convincing the masses this is somehow AI from science fiction. If you gave a random sci-fi writer from 1960s access to Claude, I'm fairly sure they wouldn't have any doubts over whether it is AI or not. They might argue about philosophical matters like whether it has a "soul" etc (there's plenty of that in sci-fi), but that is a separate debate. |
|
|
| ▲ | analog31 4 days ago | parent | prev | next [-] |
| There's a middle ground which is to watch and see what happens around us. Is it unholy to not have an opinion? |
|
| ▲ | scotty79 3 days ago | parent | prev | next [-] |
| > They wrote off the coding ability of ChatGPT on version 3.5, for instance, and have missed all the advancements that have happened since. I feel like I see now more dismissive comments than previously. As if people, initially confused, formed a firm belief since. And now new facts don't really change it, just entrench them in chosen belief. |
|
| ▲ | 9rx 3 days ago | parent | prev | next [-] |
| > They wrote off the coding ability of ChatGPT on version 3.5, for instance I found I had better luck with ChatGPT 3.5's coding abilities. What the newer models are really good at, though, is doing the high level "thinking" work and explaining it in plain English, leaving me to simply do the coding. |
|
| ▲ | CPLX 4 days ago | parent | prev | next [-] |
| I agree with you. I am a perpetual cynic about new technology (and a GenXer so multiply that by two) and I have deeply embraced AI in all parts of my business and basically am engaging with it all day for various tasks from helping me compare restaurant options to re-tagging a million contact records in salesforce. It’s incredibly powerful and will just clearly be useful. I don’t believe it’s going to replace intelligence or people but it’s just obviously a remarkable tool. But I think at least part of the dynamic is that the SV tech hype booster train has been so profoundly full of shit for so long that you really can’t blame people for skepticism. Crypto was and is just a giant and elaborate grift, to name one example. Also guys like Altman are clearly overstating the current trajectory. The dismissive response does come with some context attached. |
| |
| ▲ | parineum 4 days ago | parent [-] | | > But I think at least part of the dynamic is that the SV tech hype booster train has been so profoundly full of shit for so long that you really can’t blame people for skepticism. They are still full of shit about LLMs, even if it is useful. |
|
|
| ▲ | LennyHenrysNuts 2 days ago | parent | prev | next [-] |
| And that's why I keep checking back in. They're still pretty dumb if you want the to do anything (ie with MCPs) but they're not bad at writing and code. |
|
| ▲ | on_the_train 4 days ago | parent | prev | next [-] |
| But the reports are from shills. The impact of ai is almost non existent. The greatest impact it had was on role-playing. It's hardly even useful for coding. And that all wouldn't be a problem if it wasn't for the wave of bots that makes the crypto wave seem like child's play. |
| |
| ▲ | loandbehold 4 days ago | parent | next [-] | | I don't understand people who say AI isn't useful for coding. Claude Code improved my productivity 10x. I used to put solid 8 hours a day in my remote software engineering job. Now I finish everything in 2 hours and go play with my kids. And my performance is better than before. | | |
| ▲ | bigstrat2003 4 days ago | parent | next [-] | | I don't understand people who say this. My knee jerk reaction (which I rein in because it's incredibly rude) is always "wow, that person must really suck at programming then". And I try to hold to the conviction that there's another explanation. For me, the vast, vast majority of the time I try to use it, AI slows my work down, it doesn't speed it up. As a result it's incredibly difficult to understand where these supposed 10x improvements are being seen. | | |
| ▲ | loandbehold 4 days ago | parent | next [-] | | For me, most of the value comes from Claude Code's ability to 1. research codebase and answer questions about it 2. Perform adhoc testing on the code. Actually writing code is icing on the cake. I work on large code base with more than two million lines of code. Claude Code's ability to find relevant code, understand its purpose, history and interfaces is very time saving. It can answer in minutes questions that would take hours of digging through the code base. Ad hoc testing is another thing. E.g. I can just ask it to test an API endpoint. It will find correct data to use in the database, call the endpoint and verify that it returned correct data and e.g. everything was updated in db correctly. | |
| ▲ | libraryofbabel 4 days ago | parent | prev | next [-] | | Usually the "10x" improvements come from greenfield projects or at least smaller codebases. Productivity improvements on mature complex codebases are much more modest, more like 1.2x. If you really in good faith want to understand where people are coming from when they talk about huge productivity gains, then I would recommend installing Claude Code (specifically that tool) and asking it to build some kind of small project from scratch. (The one I tried was a small app to poll a public flight API for planes near my house and plot the positions, along with other metadata. I didn't give it the api schema at all. It was still able to make it work.) This will show you, at least, what these tools are capable of -- and not just on toy apps, but also at small startups doing a lot of greenfield work very quickly. Most of us aren't doing that kind of work, we work on large mature codebases. AI is much less effective there because it doesn't have all the context we have about the codebase and product. Sometimes it's useful, sometimes not. But to start making that tradeoff I do think it's worth first setting aside skepticism and seeing it at its best, and giving yourself that "wow" moment. | | |
| ▲ | mattmanser 4 days ago | parent | next [-] | | So, I'm doing that right now. You do get wow moments, but then you rapidly hit the WTF are you doing moments. One of the first three projects I tried was a spin on a to-do app. The buttons didn't even work when clicked. Yes, I keep it iterating, give it a puppeteer MCP, etc. I think you're just misunderstanding how hard it is to make a greenfield project when you have a super-charged stack overflow that AI is. Greenfield projects aren't hard, what's hard is starting them. What AI has helped me immensely with is blank page syndrome. I get it to spit out some boilerplate for a SINGLE page, then boom, I have a new greenfield project 95% my own code in a couple of days. That's the mistake I think you 10x ers are making. And you're all giddy and excited and are putting in a ton of work without realising you're the one doing the work, not the AI. And you'll eventually burn out on that. And those of us who are a bit more skeptical are realising we could have done it on our own, faster, we just wouldn't normally have bothered. I'd have gone done some gardening with that time instead. | | |
| ▲ | libraryofbabel 4 days ago | parent [-] | | I'm not a 10x-er. My job is working on a mature codebase. The results of AI in that situation are mixed, 1.2x if you're lucky. My recommendation was that it's useful to try the tools on greenfield projects, since they you can see them at their best. The productivity improvements of AI for greenfield projects are real. It's not all bullshit. It is a huge boost if you're at a small startup trying to find product market fit. If you don't believe that and think it would be faster to do it all manually I don't know what to tell you - go talk to some startup founders, maybe? | | |
| ▲ | mattmanser 3 days ago | parent [-] | | That 1.2x is suspiciously familiar to the recent study showing AI harmed productivity. 1.2x was self-reported, but when measured, developers were actually 0.85x ers using AI. |
|
| |
| ▲ | loandbehold 4 days ago | parent | prev [-] | | I was able to realize huge productivity gains working on a 20 years old codebase with 2+ million loc, as I mentioned in the sister post. So I disagree that big productivity gains are only on greenfield projects. Realizing productivity gains on mature code based requires more skill and upfront setup. You need to put some work in your claude.md and give Claude tools for accessing necessary data, logs, build process. It should be able to test your code autonomously as much as possible. In my experience, people who say they are not able to realize productivity gains don't put enough effort to understand these new tools and setup them properly for their project. | | |
| ▲ | libraryofbabel 4 days ago | parent [-] | | You should write a blog post on this! We need more discussion of how to get traction on mature codebases and less of the youtube influencers making toy greenfield apps. Of course at a high level it's all going to be "give the model the right context" (in Claude.md etc.) but the devil is in the details. |
|
| |
| ▲ | bentcorner 4 days ago | parent | prev [-] | | It depends on what kind of code you're working on and what tools you're using. There's a sliding scale of "well known language + coding patterns" combined with "useful coding tools that make it easy to leverage AI", where AI can predict what you're going to type, and also you can throw problems at the AI and it is capable of solving "bigger" problems. Personally I've found that it struggles if you're using a language that is off the beaten path. The more content on the public internet that the model could have consumed, the better it will be. |
| |
| ▲ | on_the_train 3 days ago | parent | prev [-] | | Then why don't you put in 8 hrs like before and get worldwide fame and be set for life within a year for being the best dev the world has ever seen? |
| |
| ▲ | lopatin 4 days ago | parent | prev [-] | | > They wrote off the coding ability of ChatGPT on version 3.5, for instance, and have missed all the advancements that have happened since. > It's hardly even useful for coding. I’m curious what kind of projects you’re writing where AI coding agents are barely useful. It’s the “shills” on YouTube that keep me up to date with the latest developments and best practices to make the most of these tools. To me it makes tools like CC not only useful but indispensable. Now I do not focus on writing the thing, but I focus on building agents who are capable of building the thing with a little guidance. |
|
|
| ▲ | research_pie 4 days ago | parent | prev | next [-] |
| I think one of the issue is also the sheer amount of shilling going on like crypto level I got a modest tech following and you wouldn’t believe the amount I’m offered to promote the most garbage AI company |
|
| ▲ | libraryofbabel 4 days ago | parent | prev | next [-] |
| I do see this a lot. It's hard to have a reasonable conversation about AI amidst, on the one hand, hype-mongers and boosters talking about how we'll have AGI in 2027 and all jobs are just about to be automated away, and on the other hand, a chorus of people who hate AI so much they have invested their identify in it failing and haven't really updated their priors since ChatGPT came out. Both groups repeat the same set of tired points that haven't really changed much in three years. But there are plenty of us who try and walk a middle course. A lot of us have changed our opinions over time. ("When the facts change, I change my mind.") I didn't think AI models were much use for coding a year ago. The facts changed. (Claude Code came out.) Now I do. Frankly, I'd be suspicious of anyone who hasn't changed their opinions about AI in the last year. You can believe all these things at once, and many of us do: * LLMs are extremely impressive in what they can do. (I didn't believe I'd see something like this in my lifetime.) * Used judiciously, they are a big productivity boost for software engineers and many other professions. * They are imperfect and make mistakes, often in weird ways. They hallucinate. There are some trivial problems that they mess up. * But they're not just "stochastic parrots." They can model the world and reason about it, albeit imperfectly and not like humans do. * AI will change the world in the next 20 years * But AI companies are overvalued at the present time and we're mostly likely in a bubble which will burst. * Being in a bubble doesn't mean the technology is useless. (c.f. the dotcom bubble or the railroad bubble in the 19th century.) * AGI isn't just around the corner. (There's still no way models can learn from experience.) * A lot of people making optimistic claims about AI are doing it for self-serving boosterish reasons, because they want to pump up their stock price or sell you something * AI has many potential negative consequences for society and mental health, and may be at least as nasty as social media in that respect * AI has the potential to accelerate human progress in ways that really matter, such as medical research * But anyone who claims to know the future is just guessing |
| |
| ▲ | IX-103 4 days ago | parent | next [-] | | > But they're not just "stochastic parrots." They can model the world and reason about it, albeit imperfectly and not like humans do. I've not seen anything from a model to persuade me they're not just stochastic parrots. Maybe I just have higher expectations of stochastic parrots than you do. I agree with you that AI will have a big impact. We're talking about somewhere between "invention of the internet" and "invention of language" levels of impact, but it's going to take a couple of decades for this to ripple through the economy. | | |
| ▲ | libraryofbabel 4 days ago | parent | next [-] | | What is your definition of "stochastic parrot"? Mine is something along the lines of "produces probabilistic completions of language/tokens without having any meaningful internal representation of the concepts underlying the language/tokens." Early LLMs were like that. That's not what they are now. An LLM got Gold on the Mathematical Olympiad - very difficult math problems that it hadn't seen in advance. You don't do that without some kind of working internal model of mathematics. There is just no way you can get to the right answer by spouting out plausible-sounding sentence completions without understanding what they mean. (If you don't believe me, have a look at the questions.) | | |
| ▲ | pm 4 days ago | parent [-] | | Ignoring its negative connotation, it's more likely to be a highly advanced "stochastic parrot". > "You don't do that without some kind of working internal model of mathematics." This is speculation at best. Models are black boxes, even to those who make them. We can't discern a "meaningful internal representation" in a model, anymore than a human brain. > "There is just no way you can get to the right answer by spouting out plausible-sounding sentence completions without understanding what they mean." You've just anthropomorphised a stochastic machine, and this behaviour is far more concerning, because it implies we're special, and we're not. We're just highly advanced "stochastic parrots" with a game loop. | | |
| ▲ | int_19h 3 days ago | parent [-] | | > This is speculation at best. Models are black boxes, even to those who make them. We can't discern a "meaningful internal representation" in a model, anymore than a human brain. They are not pure black boxes. They are too complex to decipher, but it doesn't mean we can't look at activations and get some very high level idea of what is going on. For world models specifically, the paper that first demonstrated that LLM has some kind of a world model corresponding to the task it is trained on came out in 2023: https://www.neelnanda.io/mechanistic-interpretability/othell.... Now you might argue that this doesn't prove anything about generic LLMs, and that is true. But I would argue that, given this result, and given what LLMs are capable of doing, assuming that they have some kind of world model (even if it's drastically simplified and even outright wrong around the edges) should be the default at this point, and people arguing that they definitely don't have anything like that should present concrete evidence ot that effect. > We're just highly advanced "stochastic parrots" with a game loop. If that is your assertion, then what's the point of even talking about "stochastic parrots" at all? By this definition, _everything_ is that, so it ceases to be a meaningful distinction. |
|
| |
| ▲ | app134 4 days ago | parent | prev | next [-] | | In-context learning is proof that LLMs are not stochastic parrots. | |
| ▲ | nuancebydefault 4 days ago | parent | prev [-] | | Stochastic parrot here (or not?). Can you tell the difference? |
| |
| ▲ | dvfjsdhgfv 4 days ago | parent | prev [-] | | > AI will change the world in the next 20 years Well, it's been changing the world for quite some time, both in good and bad ways. There is no need to add an arbitrary timestamp. |
|
|
| ▲ | kelseyfrog 4 days ago | parent | prev [-] |
| There's three important beliefs at play in the A(G)I story: 1. When(if) AGI will arrive. It's likely going to be smeared out over a couple months to years, but relative to everything else, it's a historical blip. This really is the most contention belief with the most variability. It is currently predicted to be 8 years[1]. 2. What percentage of jobs will be replaceable with AGI? Current estimates between 80-95% of professions. The remaining professions "culturally require" humans. Think live performance, artisanal goods, in-person care. 3. How quickly will AGI supplant human labor? What is the duration of replacement from inception to saturation? Replacement won't happen evenly, some professions are much easier to replace with AGI, some much more difficult. Let's estimate a 20-30 years horizon for the most stubborn to replace professions. What we have is a ticking time bomb of labor change at least an order of magnitude greater than the transition from an agricultural economy to an industrial economy or from an industrial economy to a service economy. Those happened over the course of several generations. Society: culture, education, the legal system, the economy, where able to absorb the changes over 100-200 years. Yet we're talking about a change on the same scale happening 10 times faster - within the timeline of one's professional career. And still, with previous revolutions we had incredible unrest, and social change. Taken as a whole, we'll have possibly the majority of the economy operating outside the territory of society, the legal system, and the existing economy. A kid born on the the "day" AGI arrives will become an adult in a profoundly different world as if born on a farm in 1850 and reaching adulthood in a city in 2000. 1. https://www.metaculus.com/questions/5121/date-of-artificial-... |
| |
| ▲ | semi-extrinsic 4 days ago | parent [-] | | Your only reference [1] is to a page where anybody in the world can join and vote. It literally means absolutely nothing. For [2] you have no reference whatsoever. How does AI replace a nurse, a vet, a teacher, a construction worker? | | |
| ▲ | saltcured 4 days ago | parent | next [-] | | For the AI believer who has an axiom that AGI is around the corner to take over knowledge work, isn't that just "a small matter of robotics" to either tele-operate a physical avatar or deploy a miniaturized AI in an autonomous chassis? I'm afraid it's really a matter of faith, in either direction, to predict whether an AI can take over the autonomous decision making and robotic systems can take over physical actions which are currently delegated to human professions. And, I think many robotic control problems are inherently solved if we have sufficient AI advancement. | |
| ▲ | kelseyfrog 4 days ago | parent | prev [-] | | What are you talking about? This is common knowledge. Median forecasts indicated a 50% probability of AI systems being capable of automating 90% of current human tasks in 25 years and 99% of current human tasks in 50 years[1] The scope of work replaceable by embodied AGI and the speed of AGI saturation of vastly under estimated. The bottle necks are production of a replacement workforce, not retraining human laborers. 1. https://arxiv.org/pdf/1901.08579 | | |
| ▲ | l33tbro 4 days ago | parent [-] | | Work is central to identity. It may seem like it is merely toil. You may even have a meaningless corporate job or be indentured. But work is the primary social mechanism that distributes status amongst communities. A world of 99 percent of jobs being done by AGI (which there remains no convincing grounds for how this tech would ever be achieved) feels ungrounded in the reality of human experience. Dignity, rank, purpose etc are irreducible properties of a functional society, which work currently enables. It's far more likely that we'll hit some kind of machine intelligence threshold before we see a massive social pushback. This may even be sooner than we think. | | |
| ▲ | int_19h 3 days ago | parent | next [-] | | Have you considered that perhaps tying dignity and status to work is a major flaw in our social arrangements, and AI (that would actually be good enough to replace humans) is the ultimate fix? If AI doing everything means that we'll finally have a truly egalitarian society where everyone is equal in dignity and rank, I'd say the faster we get there, the better. | |
| ▲ | kelseyfrog 4 days ago | parent | prev [-] | | Pretend I'm a farmer in 1850 and I have a belief that the current proportion of jobs in agriculture - 55% of jobs in 1850 would drop to 1.2% in 2022 due to automation and technological advances. Why would hearing "work is central to identity," and "work is the primary social mechanism that distributes status amongst communities," change my mind? | | |
| ▲ | l33tbro 4 days ago | parent [-] | | People migrated from the farms to the city. They didn't stop working. | | |
| ▲ | kelseyfrog 4 days ago | parent [-] | | My apologies if you thought I was arguing that a consequence of AGI would be a permanent reduction in the labor force. What I believe is that the baumol effect will take over non-replaceable professions. A very tiny part of our current economy will become the majority of our future economy. |
|
|
|
|
|
|