| ▲ | OtomotO 8 hours ago |
| Many people averted religion (which I can get behind with), but have never removed the dogmatic thinking that lay at its root. As so many things these days: It's a cult. I've used Claude for many months now.
Since February I see a stark decline in the work I do with it. I've also tried to use it for GPU programming where it absolutely sucks at, with Sonnet, Opus 4.5 and 4.6 But if you share that sentiment, it's always a "You're just holding it wrong" or "The next model will surely solve this" For me it's just a tool, so I shrug. |
|
| ▲ | balls187 7 hours ago | parent | next [-] |
| > I've used Claude for many months now. Since February I see a stark decline in the work I do with it. I find myself repeating the following pattern: I use an AI model to assist me with work, and after some time, I notice the quality doesn't justify the time investment. I decide to try a similar task with another provider. I try a few more tests, then decide to switch over for full time work, and it feels like it's awesome and doing a good job. A few months later, it feels like the model got worse. |
| |
| ▲ | runarberg 7 hours ago | parent | next [-] | | I wonder about this. I see two obvious possibilities (if we ignore bias): 1. The models are purposefully nerfed, before the release of the next model, similar to how Apple allegedly nerfed their older phones when the next model was out. 2. You are relying more and more on the models and are using your talent less and less. What you are observing is the ratio of your vs. the model’s work leaning more and more to the model’s. When a new model is released, it produces better quality code then before, so the work improves with it, but your talent keeps deteriorating at a constant rate. | | |
| ▲ | ehnto 7 hours ago | parent | next [-] | | I definitely find your last point is true for me. The more work I am doing with AI the more I am expecting it to do, similar to how you can expect more over time from a junior you are delegating to and training. However the model isn't learning or improving the same way, so your trust is quickly broken. As you note, the developer's input is still driving the model quite a bit so if the developer is contributing less and less as they trust more, the results would get worse. | | |
| ▲ | tonyarkles 6 hours ago | parent | next [-] | | > However the model isn't learning or improving the same way, so your trust is quickly broken. One other failure mode that I've seen in my own work while I've been learning: the things that you put into AGENTS.md/CLAUDE.md/local "memories" can improve performance or degrade performance, depending on the instructions. And unless you're actively quantitatively reviewing and considering when performance is improving or degrading, you probably won't pick up that two sentences that you added to CLAUDE.md two weeks ago are why things seem to have suddenly gotten worse. > similar to how you can expect more over time from a junior you are delegating to and training That's the really interesting bit. Both Claude and Codex have learned some of my preferences by me explicitly saying things like "Do not use emojis to indicate task completion in our plan files, stick to ASCII text only". But when you accidentally "teach" them something that has a negative impact on performance, they're not very likely to push back, unlike a junior engineer who will either ignore your dumb instruction or hopefully bring it up. > As you note, the developer's input is still driving the model quite a bit so if the developer is contributing less and less as they trust more, the results would get worse. That is definitely a thing too. There have been a few times that I have "let my guard down" so to speak and haven't deeply considered the implications of every commit. Usually this hasn't been a big deal, but there have been a few really ugly architectural decisions that have made it through the gate and had to get cleaned up later. It's largely complacency, like you point out, as well as burnout trying to keep up with reviewing and really contemplating/grokking the large volume of code output that's possible with these tools. | |
| ▲ | svnt 7 hours ago | parent | prev [-] | | Your version of the last point is a bit softer I think — parent was putting it down to “loss of talent” but yours captures the gaps vs natural human interaction patterns which seems more likely, especially on such short timescales. | | |
| ▲ | runarberg 7 hours ago | parent [-] | | I confusingly say both. First I say that the ratio of work coming from the model is increasing, and when I am clarifying I say “your talent keeps deteriorating”. You correctly point out these are distinct, and maybe this distinction is important, although I personally don‘t think so. The resulting code would be the same either way. Personally I can see the case for both interpretation to be true at the same time, and maybe that is precisely why I confused them so eagerly in my initial post. |
|
| |
| ▲ | rescbr 4 hours ago | parent | prev | next [-] | | I don’t think the providers intentionally nerf the models to make the new one look better. It’s a matter of them being stingy with infrastructure, either by choice to increase profit and/or sheer lack of resources to keep n+1 models deployed in parallel without deprecating older ones when a new one is released. I’d prefer providers to simply deprecate stuff faster, but then that would break other people’s existing workflows. | |
| ▲ | flux3125 6 hours ago | parent | prev [-] | | Point 2 is so true, I definitely find myself spending more time reading code vs writing it. LLMs can teach you a lot, but it's never the same as actually sitting down and doing it yourself. |
| |
| ▲ | e12e 7 hours ago | parent | prev [-] | | I think it might have to do with how models work, and fundamental limits with them (yes, they're stochastic parrots, yes they confabulate). Newer (past two years?) models have improved "in detail" - or as pragmatic tools - but they still don't deserve the anthropomorphism we subject them to because they appear to communicate like us (and therefore appear to think and reason, like us). But the "holes" are painted over in contemporary models - via training, system prompts and various clever (useful!) techniques. But I think this leads us to have great difficulty spotting the weak spots in a new, or slightly different model - but as we get to know each particular tool - each model - we get better at spotting the holes on that model. Maybe it's poorly chosen variable names. A tendency to write plausible looking, plausibly named, e2e tests that turns out to not quite test what they appear to test at first glance. Maybe there's missing locking of resources, use of transactions, in sequencial code that appear sound - but end up storing invalid data when one or several steps fail... In happy cases current LLMs function like well-intentioned junior coders enthusiasticly delivering features and fixing bugs. But in the other cases, they are like patholically lying sociopaths telling you anything you want to hear, just so you keep paying them money. When you catch them lying, it feels a bit like a betrayal. But the parrot is just tapping the bell, so you'll keep feeding it peanuts. |
|
|
| ▲ | taurath 7 hours ago | parent | prev | next [-] |
| I agree - the problem is it’s hard to see how people who say they’re using it effectively actually are using it, what they’re outputting, and making any sort of comparison on quality or maintainability or coherence. In the same way, it’s hard to see how people who say they’re struggling are actually using it. There’s truth somewhere in between “it’s the answer to everything” and “skill issue”. We know it’s overhyped. We know that it’s still useful to some extent, in many domains. |
| |
| ▲ | balls187 6 hours ago | parent [-] | | Well summarized. We're also seeing that the people up top are using this to cull the herd. |
|
|
| ▲ | psychoslave 7 hours ago | parent | prev | next [-] |
| What is it that is dogma free? If one goes hardcore pyrrhonism, doubting that there is anything currently doubting as this statement is processed somehow, that is perfectly sound. At some point the is a need to have faith in some stable enough ground to be able to walk onto. |
| |
|
| ▲ | ecshafer 7 hours ago | parent | prev | next [-] |
| All people think dogmatically. The only difference is what the ontological commitments and methaphysical foundations are. Take out God and people will fit politics, sports teams, tools, whatever in there. Its inescapable. |
| |
| ▲ | smallmancontrov 6 hours ago | parent | next [-] | | All people think dogmatically, but religion does not prevent people from acting dogmatically in politics, sports, etc. It just doesn't. It never did. Under normal circumstances I'd consider this a nit and decline to pick it, but the number of evangelists out there arguing the equivalent of "cure your alcohol addiction with crystal meth!" is too damn high. | |
| ▲ | bensyverson 7 hours ago | parent | prev | next [-] | | Allow me to introduce you to Buddhism | | |
| ▲ | ecshafer 6 hours ago | parent | next [-] | | Elaborate. Buddhism is going to have the same epistemological issues as anything, since its a human consciousness issue. | | | |
| ▲ | svnt 7 hours ago | parent | prev [-] | | Which one? | | |
| ▲ | bensyverson 7 hours ago | parent [-] | | Zen | | |
| ▲ | svnt 6 hours ago | parent [-] | | The Western Zen? In my experience it is downgraded from being a religion to being a system of practice which relieves it of the broader Mahayana cosmology. But I would suggest the dogma is less obvious but still there, often just somewhere else, such as in its own limitations, or in a philosophical container at a higher level such as scientism. | | |
| ▲ | bensyverson 5 hours ago | parent [-] | | All Zen is about releasing those attachments. Granted it's pretty hard, because if you succeed, you're enlightened. East, West, Religion, Practice… From a Zen perspective, you're just troubling your mind with binaries and conflict. | | |
| ▲ | svnt 4 hours ago | parent [-] | | Ah and there is the dogma -- the otherness of the enlightened. The binaries still functionally exist. I see a lot of value in reflective practices. At the same time it seems unlikely to me that the point of existing is to not trouble your mind. | | |
| ▲ | bensyverson 3 hours ago | parent [-] | | There's a saying in Zen: if you meet the buddha on the road, kill him. The point being, the very exaltation of enlightenment is an impediment. If Buddhism can be said to have a goal, it is to reduce suffering (including your own), so troubling your own mind is indeed something it can help with. The point of existence would be something interesting to meditate on. If you discover it, let us all know! |
|
|
|
|
|
| |
| ▲ | OtomotO 7 hours ago | parent | prev [-] | | Dogmatism is a spectrum and for too many people it's on the animal side of the scale. |
|
|
| ▲ | taneq 7 hours ago | parent | prev | next [-] |
| I wonder to what degree it depends on how easy you find coding in general. I find for the early steps genAI is great to get the ball rolling, but rapidly it becomes more work to explain what it did wrong and how to fix it (and repeat until it does so) than to just fix the code myself. |
|
| ▲ | redsocksfan45 6 hours ago | parent | prev [-] |
| [dead] |