Remix.run Logo
bambax 2 hours ago

I intensely agree with everything that's being said in TFA; this however could be nuanced:

> Never ask a model for confirmation; the tool agrees with everyone

If asked properly, LLMs can be used to poke holes in an existing reasoning or come up with new ideas or things to explore. So yes, never ask a model for confirmation or encouragement; but you can absolutely ask it to critique something, and that's often of value.

pkulak 2 hours ago | parent | next [-]

One of the best uses of AI I've found is code reviewing stuff I've written either entirely myself, or even code generated in a previous session.

2ndorderthought an hour ago | parent [-]

Yes or boiler plate! I usually go in and tweak it anyways because it's not good. But it does help. This agentic coding thing is madness to me.

I switched over to small local models. I do not need the vibe coder expensive models at all

pkulak 2 minutes ago | parent [-]

But those giant models get the boilerplate correct the first try! You're totally right though. My favorite thing to do these days is to hand craft the code in the middle of the app, then tell AI to make me a rest endpoint and a test. I do the fun/important part. :D

Though, that's coming from someone who can't justify thousands on personal hardware and is instead paying $20/month to Openai. Might as well use the best.

imiric an hour ago | parent | prev [-]

> never ask a model for confirmation or encouragement; but you can absolutely ask it to critique something, and that's often of value.

What's the difference? The end result is equally unreliable.

In either case, the value is determined by a human domain expert who can judge whether the output is correct or not, in the right direction or not, if it's worth iterating upon or if it's going to be a giant waste of time, and so on. And the human must remain vigilant at every step of the way, since the tool can quickly derail.

People who are using these tools entirely autonomously, and give them access to sensitive data and services, scare the shit out of me. Not because the tool can wipe their database or whatnot, but because this behavior is being popularized, normalized, and even celebrated. It's only a matter of time until some moron lets it loose on highly critical systems and infrastructure, and we read something far worse than an angry tweet.