| ▲ | fooker 3 hours ago |
| This might not pan out to be the glorious victory of human craft as you’re imagining it to be. Here’s a slightly different future - these AI rescue consultants are bots too, just trained for this purpose. Plausible? I have already experienced claude 4.7 handle pretty complex refactors without issues. Scale and correctness aren’t even 1% of the issue it was last year. You just have to get the high level design right, or explicitly ask it critique your design before building it. |
|
| ▲ | malfist 3 hours ago | parent | next [-] |
| > You just have to get the high level design right, or explicitly ask it critique your design before building it. Do you think people are not giving their agents specs and asking for input? |
| |
|
| ▲ | dullcrisp 2 hours ago | parent | prev | next [-] |
| And the bots training the bots are just bots that were trained to train bots? |
| |
| ▲ | fooker 2 hours ago | parent [-] | | Nothing that sexy, just thirty odd years of software engineering data from humans. Commits, design reviews, whitepapers, code reviews, test suites. And pretty concerning : chat logs and even keystrokes from employees nowadays. The way we train specialized bots now is incredibly inefficient, that part is rapidly improving. |
|
|
| ▲ | mattmanser 3 hours ago | parent | prev | next [-] |
| One AI can't vibe code out of the mess, so you'd make another AI trained on getting out of vibe coded messes? That's serious levels of circular thinking right there. |
| |
| ▲ | fooker 2 hours ago | parent [-] | | This is literally how training humans have worked for thousands of years. We train humans to do things untrained humans can not do. |
|
|
| ▲ | kilroy123 3 hours ago | parent | prev [-] |
| I think that will happen. I think several things can be true at the same time: - AI Hype - AI Psychosis - AI keeps getting better and better until it can work around big AI slop code bases |
| |
| ▲ | bluefirebrand 2 hours ago | parent [-] | | > AI keeps getting better and better until it can work around big AI slop code bases The belief in this is a form of AI psychosis, I think. Maybe in the future but certainly no evidence of this anytime soon | | |
| ▲ | fooker 2 hours ago | parent | next [-] | | > Maybe in the future but certainly no evidence of this anytime soon Here's some anecdotal evidence from me - I cleaned up multiple GPT 4.x era vibecoded projects recently with the latest claude model and integrated one of those into a fairly large open source codebase. This is something AI completely failed at last year. Maybe you should try something like this or listen to success stories before claiming 'certainly no evidence' in future? | |
| ▲ | whimsicalism 2 hours ago | parent | prev | next [-] | | No evidence? Chatgpt came out 3 years ago. You basically just need to stick a ruler up on a curve | | |
| ▲ | asveikau an hour ago | parent [-] | | I'm no expert, but the skeptic's opinion I've heard would be to ask: What evidence is there that we're not at or close to a plateau of what LLMs are capable of? How do you know the growth rate from 2023 to present will continue into 2029? eg. Is it more training data? More GPUs? What if we're kind of reaching the limits of those things already? | | |
| ▲ | whimsicalism an hour ago | parent | next [-] | | Ultimately, you are describing a fundamental problem with induction -- Hume's problem of induction to be specific. How can we know that anything that has been shown empirically in the past will continue to be true - we can't. Best to investigate mechanistically: I don't see why we would assume that we are at a plateau for RL. In many other settings, Go for instance, RL continues to scale until you reach compute limits. Some things are more easily RL'd than others, but ultimately this largely unlocks data. We are not yet compute/energy/physical world constrained. I think you would start observing clear changes in the world around you before that becomes a true bottleneck. Regardless, currently the vast majority of compute is used for inference not training so the compute overhang is large. Assuming that we plateau at {insert current moment} seems wishful and I've already had this conversation any number of times on this exact forum at every level of capability [3.5, 4, o1, o3, 4.6/5.5, mythos] from Nov 2022 onwards. | |
| ▲ | literalAardvark an hour ago | parent | prev [-] | | Since we're not experts, we treat it as a black box. What are the results? Is the quality of the results improving? Is the improvement accelerating or decelerating? And the answer appears to be that the improvement is accelerating. So how could it be stopping? https://metr.org/time-horizons/ |
|
| |
| ▲ | ashdksnndck 2 hours ago | parent | prev | next [-] | | I have personally had success telling Claude that some AI-written system is too complicated and ask it to rewrite it in a more logical way. This sometimes results in thousands of lines of code being deleted. I give an instruction like that if I see certain red flags, eg: 1) same business logic implemented in two different places, with extra code to sync between them 2) fixing apparently simple bugs results in lots of new code being written It’s a sign I need to at least temporarily dedicate more effort to overseeing work in that area. I somewhat agree with the AI psychosis framing of the OP. It takes some taste and discipline to avoid letting things dissolve into complete slop. | |
| ▲ | asveikau an hour ago | parent | prev [-] | | It's amusing to me that: * A belief that AI will keep getting better, presented without evidence, does not yield a lot of skepticism around these parts. * Your comment saying it is wrong to believe AI will keep getting better, also presented without evidence, is downvoted. |
|
|