Remix.run Logo
colechristensen 4 hours ago

This is silly. I can build products in a weekend that would take me a year by myself. I am still necessary 1% of the time for debug, design, and direction and those of not at all a shallow skill. I have some graduate algebra texts on the way my math friend is guiding me through because I have found a publishable result and need to shore up my background before writing the paper...

It's not perpetual motion, it's very real capability, you just have to be able to learn how to use it.

qsera 4 hours ago | parent | next [-]

No one is saying that it cannot do what you say now.

What I am saying is that once the high quality training data runs out, it will drop in its capabilities pretty fast. That is how I compare it to perpetual motion mechanism scams. In the case of a perpetual motion machine, it appear that it will continue to run indefinitely. That is analogous to the impression that you have now. You feel that this will go on and on for ever, and that is the scam you are falling for.

WarmWash 3 hours ago | parent | next [-]

>What I am saying is that once the high quality training data runs out, it will drop in its capabilities pretty fast.

That's more a misunderstood study that over time turned into a confidently stated fact. Yes, the model collapses if you loop the output to the input. But no, that's not how it's done.

The reality is that all the labs are already using synthetic training data, and have been for at least a year now. It basically turned out to be a non-issue if you have robust monitoring and curation in place for the generated data.

qsera 3 hours ago | parent [-]

>using synthetic training data

yea, look up how it is done.

It is exactly how a perpetual motion machine scam would project an appearance of working like using a generator to drive a motor, and the motor driving the generator..something that would obscure the fact that there is energy loss happening along the way....

WarmWash 2 hours ago | parent [-]

I'm confused with the point you are trying to make, because they are using synthetic data, and the models are getting stronger.

There is no "conservation of fallacy" law (bad data must conserve it's level of bad), so I'm struggling to connect the dots on the analogy, unless I ignore the fact that training on synthetic data works, is being used, and the models are getting better.

qsera an hour ago | parent | next [-]

If the training that did not use synthetic data failed to capture some aspect of the information contained, then using data synthesized from the original data could help to capture it, thus it could result in the models getting better.

But that is because the synthetic data helped the model capture what was already there in the training data.

But after all such information has been extracted, then it would not be possible to use synthetic data or anything that is derived from the original data to create "new" information for training....

dgb23 an hour ago | parent | prev [-]

Better by which metrics?

_aavaa_ 4 hours ago | parent | prev [-]

Why would the capabilities drop instead of stagnate?

qsera 4 hours ago | parent [-]

Because technologies, programming languages, best practices, won't stay frozen. If LLMs cannot catch up with it, I think it can be considered as a drop in capability. No?

coldtea 3 hours ago | parent [-]

Close, but no. What will happen is that "technologies, programming languages, best practices" will stay frozen because human innovation will drop, and the whole field will stagnate.

californical 2 hours ago | parent [-]

This is the biggest fear! I don’t see an easy fix.

Will the developer of a new programming language be able to reach out to model companies to give a huge amount of training data, ensuring that the models are good at that new language? I don’t think a small team can write enough code, the models already struggle in medium-popularity languages that have years of history. They hallucinate lua functionality sometimes, for example, even though I’m sure there is lots of lua code out there.

So if most people use coding agents, we’re stuck with the current most popular languages because no new language will get past the barrier of having enough code that models can write it well, meaning nobody adopts the new language, etc.

Same thing with libraries and frameworks - technical decisions are already being made based on “is this popular enough that the agents can use it well?” Rather than a newer library that meets our needs perfectly but isn’t in the training data

askafriend 4 hours ago | parent | prev | next [-]

You can see their ego trying to protect itself.

coldtea 3 hours ago | parent | prev | next [-]

>This is silly. I can build products in a weekend that would take me a year by myself

Is the world any better for them existing? The decline of coding and sw engineering skills in humans from outsourcing the practice of it to AI is it worth it and sustainable long term?

colechristensen 2 hours ago | parent [-]

>Is the world any better for them existing? The decline of coding and sw engineering skills in humans from outsourcing the practice of it to AI is it worth it and sustainable long term?

The world is going to be no worse than it was when humans transitioned from writing assembly to writing compilers for high level languages. Assembly is still necessary, but not that often. In the same way writing code is going to become less necessary as tools are going to be written in higher level language in standards and requirements documents instead of code most of the time, with more specific exact coding only occasionally.

Programmers were mostly solving the same plumbing problems over and over in secret because of "proprietary" needs to hide your code, but one million separate integrations of your billing backend with Stripe didn't really add to humanity. We're cutting out the boring middle drudgery and human effort is going to be freed up to work on the edges of human knowledge instead of tromping around in the middle.

coldtea an hour ago | parent [-]

>The world is going to be no worse than it was when humans transitioned from writing assembly to writing compilers for high level languages

When I open some Electron apps I wish we stopped right about there.

tpdly 4 hours ago | parent | prev [-]

You're fooling yourself.

People yeating a (shitty) Github clone with Claude in a week apparently can't imagine it, but if you know the shit out of Rails, start with a good a boiler plate, and have a good git library, a solo dev can also build a (shitty) Github clone in a week. And they'll be able to take it somewhere, unlike the llm ratsnest that will require increasingly expensive tokens to (frustratingly) modify.

mikkupikku 4 hours ago | parent | next [-]

You're fooling yourself. It's very easy to get demonstrably working results in an afternoon that would take weeks at least without coding agents. Demonstrably working, as in you can prove the code actually works by then putting it to use. I had a coding agent write an entire declarative GUI library for mpv userscripts, rendering all widgets with ASS subtitles, then proceeded to prove to my satisfaction that it does in fact work by using it to make a node editor for constructing ffmpeg filter graphs and an in-mpv nonlinear video editor. All of this is stuff I already knew how to do in practice, had intended to do one day for years now, but never bit the bullet because I knew it would turn into weeks of me pouring over auto-generated ASS doing things it was never intended to do to figure out why something is rendering subtly wrong. Fairly straightforward but a ton of bitch work. The LLM blasted through it like it was nothing. Fooling myself? The code works, I'm using it, you're fooling yourself.

zozbot234 3 hours ago | parent | next [-]

> Demonstrably working, as in you can prove the code actually works by then putting it to use.

That's not how you prove that code works properly and isn't going to fail due to some obscure or unforessen corner case. You need actual proof that's driven by the code's overall structure. Humans do this at least informally when they code, AI's can't do that with any reliability, especially not for non-trivial projects (for reasons that are quite structural and hard to change) so most coding agents simply work their way iteratively to get their test results to pass. That's not a robust methodology.

mikkupikku an hour ago | parent | next [-]

> That's not how you prove that code works properly

Yes it is. What do you expect, formal verification of a toy GUI library? Get real.

> and isn't going to fail due to some obscure or unforessen corner case.

That's called "a bug", they get fixed when they're found. This isn't aerospace software, failure is not only an option, it's an expected part of the process.

> You need actual proof that's driven by the code's overall structure.

I literally don't.

> Humans do this at least informally when they code, AI's can't do that with any reliability

Sounds like a borderline theological argument. Coding agents one-shot problems a lot more often than I ever did. Results are what matters, demonstrable results.

coldtea 3 hours ago | parent | prev [-]

>That's not how you prove that code works properly and isn't going to fail due to some obscure or unforessen corner case.

So? We didn't prove human code "isn't going to fail due to some obscure or unforessen corner case" either (aside the tiny niche of formal verification).

So from that aspect it's quite similar.

>so most coding agents simply work their way iteratively to get their test results to pass. That's not a robust methodology.

You seem to imply they do some sort of random iteration until the tests pass, which is not the case. Usually they can see the test failing, and describe the issue exactly in the way a human programmer would, then fix it.

zozbot234 3 hours ago | parent [-]

> describe the issue exactly in the way a human programmer would

Human programmers don't usually hallucinate things out of thin air, AIs like to do that a whole lot. So no, they aren't working the exact same way.

coldtea 3 hours ago | parent [-]

>Human programmers don't usually hallucinate things out of thin air

Oh, you wouldn't believe how much they do that too, or are unreliable in similar ways. Bullshiting, thinking they tested x when they didn't, misremembering things, confidently saying that X is the bottleneck and spending weeks refactoring without measuring (to turn out not to be), the list goes on.

>So no, they aren't working the exact same way.

However they work internally, most of the time, current agents (of say, last year and above) "describe the issue exactly in the way a human programmer would".

qsera 3 hours ago | parent [-]

That is not hallucinating...

LLM hallucinating is not an edge case. It is how they generate output 100% time. Mainstream media only calls it "hallucination" when the output is wrong, but from the point of view of a LLM, it is working exactly it is supposed to....

coldtea 2 hours ago | parent [-]

>LLM hallucinating is not an edge case. It is how they generate output 100% time

If enough of the time it matches reality -- which it does, it doesn't matter. Especially in a coding setup, where you can verify the results, have tests you wrote yourself, and the end goal is well defined.

And conversely, if a human is a bullshitter, or ignorant, or liar, or stupid, it doesn't matter if they end up with useless stuff "in a different way" than an LLM hallucinating. The end result regarding the low utility of his output is the same.

Besides, one theory of cognition (pre LLM even) is of the human brain as a prediction machine. In which case, it's not that different than an LLM in principle, even if the scope and design is better.

bachmeier 3 hours ago | parent | prev [-]

> Fairly straightforward but a ton of bitch work. The LLM blasted through it like it was nothing.

One might argue that this is a substitute for metaprogramming, not software developers.

trollbridge 2 hours ago | parent [-]

It's interesting more people haven't talked about this. A lot of so-called agentic development is really just a very roundabout way to perform metaprogramming.

At my own firm, we generally have a rule we do almost everything through metaprogramming.

cindyllm 2 hours ago | parent [-]

[dead]

colechristensen 2 hours ago | parent | prev [-]

I also did a native implementation of git so I could use an S3 compatible data store, your rails guru can't do that.

Objectively, my GitHub clone is still shitty, BUT it got several ways github is shitty out of my way and allowed me to add several features I wanted, no small one of which was GitHub not owning my data.

I don't know the shit out of Rails and I don't want to, I know the shit out of other things and I want the tools I'm using to be better and Claude is making that happen.

It's a little odd the skepticism to the level that people keep telling me I'm delusional for being satisfied that I've created something useful for myself. The opposition to AI/LLMs seems to be growing into a weird morality cult trying to convince everybody else that they're leading unhappy immoral lives. I'm exaggerating but it's looking like things are going in that direction... and in my house, so to speak, here on HN there are factions. Like programming language zealots but worse.

tpdly an hour ago | parent [-]

Hey I understand you've gotten something out of it. You hired a robot to 3d-print a mug that fits your hand. There's a place for that. You understand that it might poison you a little bit? You understand that this doesn't make ceramics irrelevant?

Hobby-project vibe coding is pretty cool (if I'm being honest, its fucking miraculous; this tech is wild) but isn't it clear that there's a problem with the linkedincels, the investors, the management that are all convinced this will remove say 50% of programming jobs? I understand these things have legitimate uses, but I'm at my wits end hearing about how deep understanding, craftsmanship, patience and hard work aren't "results oriented".

There's definitely zealotry developing against AI, but I suspect it is a proportional (if unhelpful) response to the hype machine. Is it really zealotry to insist on the value of your mind and your competence? These people saying you should never "hand write" your code-- how the fuck did the discourse move so much that this isn't a laughably stupid thing to say? "I'm a CEO, and if you aren't using consultants to make your decisions you've already lost"

colechristensen 29 minutes ago | parent [-]

>isn't it clear that there's a problem with the linkedincels, the investors, the management that are all convinced this will remove say 50% of programming jobs

These people have always been doing this. Starting in the 90s it was outsourcing programming jobs, they were right then, they got more work for less money and you could have less expertise on staff farming out work somewhere else that was cheaper. You also got worse results sometimes. So it goes.

LLMs are making people more powerful and sucking a lot of income off to the people who provide them. Yup. It makes idiot shysters more powerful just the same as it makes experts more powerful.

People are acting like the software engineering industry is full of fine artistry building the finest bespoke tools instead of duct taping rocks to sticks. I'm sorry but there is a tremendous amount of crap out there by people who barely know what they're doing.

Yes new technology empowers idiots, but it also empowers smart people and if you use it well it'll lead to more quality. Yes you're going to have the same problems you had before of someone doing something cheaply competing with someone trying to be careful to build something well. There also will continue to be idiots spouting off about it.

Nothing changed but the tools got more powerful and people are whining complaining about this change this time ruining everything. Like they always have forever.