Remix.run Logo
user34283 6 hours ago

AI works. This is evidenced by my side project which I spent some 50 hours on.

I'm not sure what your "empirical evidence and repeatable tests" is supposed to be. The AI not successfully converting a 3000 line C program to Python, in a test you probably designed to fail, doesn't strike me as particularly relevant.

Also, I suspect that AI could most likely guess that 80 lines of Python aren't correctly replicating 3000 lines of C, if you prompted it correctly.

discreteevent 5 hours ago | parent | next [-]

> AI works.

For some definition of "works". This seems to be yours:

> I'd go further and say vibe coding it up, testing the green case, and deploying it straight into the testing environment is good enough. The rest we can figure out during testing, or maybe you even have users willing to beta-test for you.

> This way, while you're still on the understanding part and reasoning over the code, your competitor already shipped ten features, most of them working.

> Ok, that was a provocative scenario. Still, nowadays I am not sure you even have to understand the code anymore. Maybe having a reasonable belief that it does work will be sufficient in some circumstances.

https://news.ycombinator.com/item?id=46315569

user34283 4 hours ago | parent [-]

Yes, as I said, it is working well in my side project. The application works and I am happy with my results so far.

It's interesting how this workflow appears to almost offend some users here.

I get it, we all don't like sloppy code that does not work well or is not maintainable.

I think some developers will need to learn to give control away rather than trying to understand every line of code in their project - depending of course on the environment and use case.

Also worth to keep in mind that even if you think you understand all the code in your project - as far as that is even possible in larger projects with multiple developers - there are still bugs anyway. And a few months later, your memory might be fuzzy in any case.

stavros 4 hours ago | parent [-]

People seem to be divided between "AI doesn't work, I told it 'convert this program' and it failed" and "AI works, I guided it through converting this program and saved myself 30 hours of work".

Given my personal experience, and how much more productive AI has made me, it seems to me that some people are just using it wrong. Either that, or I'm delusional, and it doesn't actually work for me.

SamPatt 13 minutes ago | parent | next [-]

The models are good enough now that anyone who says AI doesn't work is either not acting in good faith or is staggeringly bad at learning a new skill.

It's not hard to spend a few hours testing out models / platforms and learning how to use them. I would argue this has been true for a long time, but it's so obviously true now that I think most of those people are not acting in good faith.

PostOnce 3 hours ago | parent | prev [-]

What about "I attempted to guide it and it took 5 times longer and I ended up having to do the entire thing myself anyway"?

stavros 43 minutes ago | parent [-]

That hasn't happened to me.

blibble 2 hours ago | parent | prev [-]

why is it always accounts with 50 karma saying this?

hombre_fatal an hour ago | parent | next [-]

I have 22k karma and I think it's a trivial claim that LLMs work and that software is clearly on the cusp of being 100% solved within a couple years.

The naysaying seems to mostly come from people coping with the writing they see on the wall with their anecdote about some goalpost-moving challenge designed for the LLM to fail (which they never seem to share with us). And if their low effort attempt can't crack LLMs, then nobody can.

It reminds me of HN ten years ago where you'd still run into people claiming that Javascript is so bad that anybody who thinks they can create good software with it is wrong (trust them, they've supposedly tried). Acting like they're so preoccupied with good engineering when it's clearly something more emotional.

Meanwhile, I've barely had to touch code ever since Opus 4.5 dropped. I've started wondering if it's me or the machine that's the background agent. My job is clearly shifting into code review and project management while tabbing between many terminals.

As LLMs keep improving, there's a moment where it's literally more work to find the three files you need to change than to just instruct someone to do it, and what changes the game is when you realize it's creating output you don't even need to edit anymore.

blibble 35 minutes ago | parent [-]

wasn't sure if this was sarcasm until this point:

> with their anecdote about some goalpost-moving challenge designed for the LLM to fail (which they never seem to share with us).

literally what the boosters do on every single post!

"no no, the top model last week was complete dogshit, but this new one is world changing! no you can't see my code!"

10/10 for the best booster impression I've seen this year!

user34283 2 hours ago | parent | prev [-]

If we're going to argue on that level: Maybe it's because accounts with 12k karma spend more time posting than working on side projects and trying new tools.

blibble 2 hours ago | parent [-]

that's the great thing about non-vibe coding

faster, fewer bugs, better output

leaving more time for shitposting