Remix.run Logo
epiccoleman 7 days ago

I love this article just for the spirit of fun and experimentation on display. Setting up a VPS where Claude is just asked to go nuts - to the point where you're building a little script to keep Claude humming away - is a really fun idea.

This sort of thing is a great demonstration of why I remain excited about AI in spite of all the hype and anti-hype. It's just fun to mess with these tools, to let them get friction out of your way. It's a revival of the feelings I had when I first started coding: "wow, I really can do anything if I can just figure out how."

Great article, thanks for sharing!

bkettle 6 days ago | parent | next [-]

> “wow, I really can do _anything_ if I can just figure out how

Except this time it’s “if I can just figure out how and pay for the Claude API usage”.

This is one of the sadder things about AI usage getting more standard that I haven’t seen discussed much—-the barrier to entry is now monetary rather than just knowledge-based, which will make it _much_ harder for young people with no money to pick up.

Yes, they can still write code the manual way, but if the norm is to use AI I suspect that beginner’s guides, tutorials, etc. will become less common.

infecto 6 days ago | parent | next [-]

There has generally always been some barrier. Computer access, internet access, books etc. If AI coding stays around, which looks like it will, it will just be the current generations barrier.

I don’t think it is sad at all. There are barriers to all aspects of life, life is not fair and at least in our lifetimes will never be. The best anyone can do is to help those around them and not get caught up the slog of the bad things happening in the world.

rurp 6 days ago | parent | next [-]

Having been a poor person learning how to code I'd say there's a huge difference between just needing a computer vs needing that plus a dozens per month subscription.

I don't know that there's much we can do about that potentially becoming the new normal in the future, but it bums me out.

reidjs 6 days ago | parent | next [-]

There are free and offline options, like Llama.cpp, but you will have to pay by giving up your privacy to Meta (or similar large companies)

Revisional_Sin 10 hours ago | parent | next [-]

How does using an offline model give up your privacy?

Also, running models locally requires good hardware to get acceptable performance. It's still a large barrier to entry.

rurp 6 days ago | parent | prev [-]

Sure, for now, and maybe in the future. But it's possible that paid models will end up greatly outpacing free ones, and at some point the companies controlling them will stop burning billions of dollars per month and jack up prices.

infecto 5 days ago | parent | prev [-]

No it’s not much different. I grew up poor. It was a struggle to have internet access.

xnorswap 6 days ago | parent | prev | next [-]

But traditional barriers have been able to be knocked down more easily with charity, because it's easier to raise charity money for capex than opex.

It was common to have charity drives to get computers into schools, for example, but it's much harder to see people donating money for tokens for poor people.

Previous-generation equipment can be donated, and can still spark an interest in computing and programming. Whereas you literally now can't even use ChatGPT-4.

conradev 6 days ago | parent | next [-]

Small models and processors are going to continue improving to the point that you’ll be able to vibe code locally on your phone at some point.

When the iPhone came out, not everyone had a smartphone. Now 90% of the US has a smartphone, and many of these smartphones run generative local models.

pc86 6 days ago | parent | prev | next [-]

"It's harder to convince other people to pay for this for me" is an insane criticism. Not every AI model needs a premium account, you can even run many excellent models locally if you don't want to pay for an internet connection.

At some point you just have to accept that yes things are easier if you have a little bit of spending money for things. That's not "sad" it's a basic fact of life.

xnorswap 6 days ago | parent [-]

You have been mean with your interpretation of my statement.

I am not saying, "It's harder to convince other people to pay for this for me".

I am saying, "It is harder for me to pay for this for someone else".

pc86 6 days ago | parent [-]

It probably is harder to convince others to pay for opex than capex - maybe that's a good thing, maybe not.

But it's certainly not any harder for you to donate this money to someone because they want to spend that money on tokens instead of a computer, for example.

Zambyte 6 days ago | parent [-]

I think you're still misinterpreting them. It's harder to donate tokens than it is to donate books or a computer. Not all donations are cash.

Though, in a sense, I am donating tokens. I run an LLM box at my apartment, and I let some people in my community have a free account through my hosted Open WebUI.

infecto 5 days ago | parent | prev [-]

This feels like picking a random thing against LLMs to complain about. These tools are not even necessary today I am not sure why they would be necessary tomorrow beyond efficiency. If that day does come though, you would have to assume open source models would also be coming a long way.

michaelrpeskin 6 days ago | parent | prev [-]

I attribute my barriers to entry as things that forced me to really learn. All my family could afford was a 386 with 16MB of ram when 486s where pretty common. I had to really hack to make things work. Working under constraints meant I was exploring limits and understanding efficiency.

I still carry that in my day job which I _think_ helps me write better code - even in unconstrained systems, thinking in terms of memory and instruction efficiency can still help make better/faster code.

mickael-kerjean 6 days ago | parent | prev | next [-]

Yep, I used to spend a lot of time learning PHP on a web server which was part of my internet subscription. Without it being free, I would never have learn how to create websites and would have never got in programming, the trigger was that free web hosting with PHP that was part of the internet connection my parents were already paying for

alwillis 6 days ago | parent | next [-]

There are plenty of free models available; many that rival their paid counterparts.

A kid interested in trying stuff can use Qwen Coder for free [1].

If the kid's school has Apple Silicon Macs (or iPads), this fall, each one of them will have Apple's 3 billion parameter Foundation Models available to them for free [2].

Swift Playground [3] is a free download; Apple has an entire curriculum for schools. I would expect an upgrade to incorporate access to the on-board LLM

[1]: https://openrouter.ai/qwen/qwen3-coder:free

[2]: https://developer.apple.com/videos/play/wwdc2025/286

[3]: https://developer.apple.com/swift-playground/

yoz-y 6 days ago | parent | next [-]

I guess hardware being able to run a local model will eventually get cheap enough, but for a lot of people even buying an Apple device or something with a good enough GPU is prohibitive.

PeterStuer 6 days ago | parent | next [-]

True, it will get cheap to run today's frontier models. But, by that time, how much more advanced will the frontier models of that time be.

It is a real question. It all depends on whether the AI future is linear or exponential.

hadlock 6 days ago | parent | prev [-]

I think we are already there. You can run a pretty ok LLM on a 4gb raspberry pi that will write most any simple 20-150 line bash script today, or toy application in python/rust. Old laptops pulled out of the trash are probably capable of running smaller LLMs and can explain how functions work. They're no claude code but you probably want a rough-around-the-edges LLM that can't do everything for you, if you're planning on using it to learn to code.

andai 5 days ago | parent | prev | next [-]

Speaking of free models on OpenRouter, DeepSeek R1 0528 is also available for free.

https://openrouter.ai/deepseek/deepseek-r1-0528:free

socalgal2 6 days ago | parent | prev [-]

Swift and swift playground might be a good introduction to programming, but it feels likely not to lead to as many opportunities as a more popular system. And I don’t just mean job opportunities.

pc86 6 days ago | parent | prev [-]

"Already being paid for by someone else" is very different than "free."

rurp 6 days ago | parent | prev | next [-]

Very true. One of the greatest aspects of the field is how accessible it is, and that is certainly going to get worse with LLM usage.

I'd probably be toiling away in a less productive industry if I hadn't been able to easily download Python and start learning it for free.

nostrademons 6 days ago | parent | prev | next [-]

They're not that expensive for anyone that has the tech skills to actually make good use out of them. I've been paying around with Claude Code, using API credits rather than the monthly fee. It costs about $5 per one-hour session. If you're going to be doing this professionally it's worth springing for the $100/month membership to avoid hitting credit limits, but if you just want to try it out, you can do so without breaking the bank.

A bigger question for me is "Does this actually increase my productivity?" The jury is still out on that - I've found that you really need to babysit the algorithm and apply your CS knowledge, and you also have to be very clear about what you're going to tell it later, don't let it make bad assumptions, and in many cases spell out the algorithm in detail. But it seems to be very good at looking up API details, writing the actual code, and debugging (if you guide it properly), all things that take a non-trivial amount of tedium in everyday programming.

eloisius 6 days ago | parent | next [-]

12-year-old me wasn’t putting my tech skills to good use enough to pay $5 every time I sat down at the computer. I was making things though, and the internet was full of tutorials, chat rooms, and other people you could learn from. I think it would be sad if the same curious kid today was told “just pay $5 and ask Claude” when pestering someone in IRC about how to write a guestbook in Perl.

nostrademons 6 days ago | parent | next [-]

12-year-old me wasn't either, but he was noodling around on a computer that cost $2500 (more like $5500 in today's dollars). I think our parents loved us very much and must have had some means to afford the capital cost of a computer back then.

I don't see my 7-year-old paying $5 for each hour he wants to program (and no way in hell would I give him my credit card), but I could easily envision paying $20/month for a Claude subscription and letting him use it. We pay more than that for Netflix & Disney+.

kelnos 6 days ago | parent | next [-]

> noodling around on a computer that cost $2500 (more like $5500 in today's dollars)

Wow! 12-year-old me was noodling around on a computer that my dad brought home from work because it would have otherwise ended up in landfill. We had very little money for computers back then, and I was thrilled when my parents gave me a budget to buy parts to build my own from scratch when I was about to go off to college (I'd saved up a bit myself, but not nearly enough).

I think your experience is pretty privileged, and not at all common.

nostrademons 5 days ago | parent | next [-]

There's always more to the story than the Internet assumes.

We were quite possibly less privileged than you were, if your dad brought a computer home from work. I grew up with a teacher and a househusband for parents; single-income, and that income made about 1/3 of what engineers or other computer professionals made. My kid had more passport stamps at age 2 than I did at age 18.

It was $2500 because it was a Mac LC, and it was a Mac LC because that could take an Apple 2E card and run both Mac and Apple software, and that was important because my mom was a teacher and had a large library of educational software at school that she could take home. Recall that in those days, software was sold in retail stores (no Internet), and cost $50-100 for kiddie stuff, and like $400 for productivity and compilers. 25 titles and the cost of the computer paid for itself in free software. I think we used about that.

It's a matter of priorities. My parents always prioritized my education: they bought a computer, and whatever software didn't come from my mom's workplace, and any books I wanted, and paid for my college education in full. We didn't have a whole lot other than that: we didn't take a lot of vacations or plane trips (and a single plane trip would cost more than that $2500 in those days), ran our cars into the ground (my mom owned 2 cars over my entire childhood), wore hand-me-downs.

matwood 6 days ago | parent | prev | next [-]

Everyone has some level of privilege. I didn't get my first PC until I was a freshman in college. I had to spend part of my college loan buying one (~$3k IIRC). Up to that point I had only played with the Apple IIc's and the few Macs they had at my high school..

Information on programming also wasn't as readily available as it is now. I used to go the book stores and use pencil and paper and copy out solutions since $50+ for a book was way more money than I could spend.

Everything today is crazy inexpensive for the value.

pc86 6 days ago | parent | prev [-]

So what?

That sounds dismissive, and maybe it is, but I'm being serious here. What is the point of coming here and saying "when I was 12, my parents had less money than yours did when you were 12?" Privilege is relative, "common" is relative, and constantly being dragged into oppression Olympics of who has or had things slightly worse is exhausting and not conducive to conversation.

Better keep in mind that someone here almost certainly had it even worse than you when they were in elementary school, lest you go a few seconds without acknowledging your privileged upbringing, for some reason, in a conversation where it bears absolutely no relation to anything.

jack_pp 6 days ago | parent | prev | next [-]

12-year-old me was mostly procrastinating but sometimes fired up Pascal which required me to insert a floppy disk in my 486 compaq machine for it to work. the machine was a donation from my aunt, could only run DOS.

However chatgpt or gemini free tier is more than enough for a kid to figure out how python works and build some simple software. While I have the Gemini subscription I only got it because my family drive storage was nearly full. I could've probably got by with ChatGPT free to just stop using stackoverflow.

piva00 6 days ago | parent | prev | next [-]

10-year-old me was programming on a salvaged 386 that my dad got from a friend after the company they worked at went bankrupt, and left the machine as trash. Instead of Christmas gifts I asked for programming books and pitched in some of the birthday money my grandparents would give me (about US$ 2 every birthday).

Not everyone was privileged, some of us were just lucky.

davidwritesbugs 6 days ago | parent [-]

[flagged]

dingnuts 6 days ago | parent | prev [-]

I'm extremely privileged and I had a quarter of what you did growing up. Your experience and your kids' is not typical.

sgarland 6 days ago | parent | prev [-]

12-year-old me had (or rather, my family had) a Celeron 333 MHz and a Pentium III 550 MHz, both from Gateway, because that was the sole awesome perk my dad got from working there: literally free computers, with a required number of years of employment to pay them off. In 2000, the P3 was still pretty hot shit. I dual-booted them with every Linux distro under the sun. Since we had dial-up, the only way I had those distros was from 4-H [0], which at the time in Nebraska had a partnership with University of Nebraska to do tech instruction; once a quarter, we’d drive down to a campus (usually UNL) and spend a weekend learning something (LAMP stack, hardware troubleshooting, etc.), and having a LAN party at night. Also we had free access to their (at the time) screamingly fast internet, so I would download distros and packages to try out later.

My online upbringing was very much of the RTFM variety, and I am convinced that was and is a good method to learn. It’s not like the grumpy graybeards were cruel, they just didn’t want to waste their time answering the same “how do I…” questions from noobs. If you explained what you were experiencing, what you had read, and what you had tried, they were more than happy to help out. I don’t think that’s an unreasonable approach.

[0]: https://4-h.org/

hdjrudni 6 days ago | parent | prev | next [-]

I think you said it. $100/mo and you're not even sure if it'll increase your productivity. Why on earth would I pay that? Do I want to flush $100 down the toilet and waste several days of my life to find out?

nostrademons 6 days ago | parent | next [-]

You don't have to pay $100 to find out, you can do that for ~$5-20 by directly buying API credits.

I don't know for sure whether it's worth it yet. Further experimentation is needed, as well as giving it an honest shot and trying to learn the nuances of the tool. But the way I look at it - if this actually is a future career path, the net present value of its payoff is measured in the millions of dollars. It's worth spending ~$20 and a few nights of my time to figure that out, because the odds can be pretty damn low and still have the expected value pencil out. It's sorta like spending $200 on 1/4 of a Bitcoin in 2013 because I was curious about the technology - I fully expected it to be throwing money down the toilet, but it ended up being quite worth it. (I wish I'd had the same mindset when I could've bought into the Ethereum ICO at a penny or so an ETH.)

handfuloflight 6 days ago | parent | prev [-]

Are you spending every $100 absolutely efficiently? How do you know?

jrflowers 6 days ago | parent [-]

Can I have a hundred dollars? I assure you giving it to me is efficient

kyleee 6 days ago | parent [-]

And it’s even more efficient to give me $90

barrell 6 days ago | parent | prev [-]

I have the tech skills to use them. In my 30s and I could not spend $5 on a one hour coding session even if it 10xed my productivity. 1-2 hours would literally break the bank for me

mark_l_watson 6 days ago | parent | prev | next [-]

yes indeed, who will pay? I run a lot through open models locally using LM Studio and Ollama, and it is nice to only be spending a tiny amount of extra money for electricity.

I am retired and not wanting to spend a ton of money getting locked long term into using an expensive tool like Claude Code is a real thing. It is also more fun to sample different services. Don’t laugh but I am paying Ollama $20/month just to run gpt-oss-120b very fast on their (probably leased) hardware with good web search tooling. Is it worth $20/month? Perhaps not but I enjoy it.

I also like cheap APIs: Gemini 2.5-flash, pro when needed, Kimi K2, open models on Groq, etc.

The AI, meaning LLM, infrastructure picture is very blurred because of so many companies running at a loss - which I think should be illegal because long term I think it is misleading consumers.

piva00 6 days ago | parent [-]

> The AI, meaning LLM, infrastructure picture is very blurred because of so many companies running at a loss - which I think should be illegal because long term I think it is misleading consumers.

In a sense it is illegal, even though the whole tech scene has been doing it for decades, price dumping is an illegal practice and I still don't understand why it has never been considered as such with tech.

Most startups with VC investors work only through price dumping, most unicorns came to be from this bullshit practice...

nl 6 days ago | parent [-]

"Price dumping" isn't an economic term in common use.

"Dumping" in international trade is somewhat similar but the reasons that is illegal are very different: https://en.m.wikipedia.org/wiki/Dumping_(pricing_policy)

Pricing at a loss by VC funded companies is great for consumers. It rarely is at a loss though - they look at the lifetime value.

Pricing at a loss by big tech could be viewed as anticompetitive. Personally I like that Gemini keeps OpenAI prices lower but one could argue it has stopped OpenAIs growth.

piva00 6 days ago | parent | next [-]

> Pricing at a loss by VC funded companies is great for consumers. It rarely is at a loss though - they look at the lifetime value.

It's great for consumers only in the short term, the strategy to drive out competition that are not as well-funded has only one goal: to remove competition in the long-term to drive up prices at your will since most competitors won't have the chance to exist.

Edit: yes, technically dumping is a specific type of predatory pricing, so swap "price dumping" on my first comment to "predatory pricing" instead.

nl 6 days ago | parent [-]

It doesn't have one goal.

In fact driving out competition is rarely the goal at all.

Instead the goal is usually to reduce the barrier to people trying the thing - especially when it is a developer API which you hope developers will incorporate into their product.

piva00 6 days ago | parent | next [-]

> In fact driving out competition is rarely the goal at all.

Driving out competition is definitely a goal, the further you can snowball that makes your company a much more attractive investment since your competition will be bleeding money, attrition is definitely used as a tactic by VCs when a startup gets traction. Hell, it's one of the arguments they use to run further rounds of investments to others "this startup is very well capitalised and the competition has 1/10th of their funds, investing elsewhere is a losing proposition".

> Instead the goal is usually to reduce the barrier to people trying the thing - especially when it is a developer API which you hope developers will incorporate into their product.

I thought we were talking about unicorns such as Uber, AirBnb, etc., not some dev startup packaging APIs to serve other startups which is a whole other incestuous industry.

guappa 5 days ago | parent | prev [-]

AI company founder and CTO defends those practices… yawn.

Filligree 6 days ago | parent | prev [-]

I wouldn’t assume Gemini is being run at a loss, though. At least not that, if it weren’t, that would help OpenAI much.

Google uses Google hardware, which costs them 1/10 what nvidia hardware costs everyone else.

coldtea 6 days ago | parent | prev | next [-]

>the barrier to entry is now monetary rather than just knowledge-based, which will make it _much_ harder for young people with no money to pick up.

Considering opportunity cost, a young person paying $20 or $100 per month to Claude API access is way cheaper than a young person spending a couple of years to learn to code, and some months coding something the AI can spit in 10 minutes.

AI coding will still create generations that even programming graduates know fuck all about how to code, and are also bad at reasoning about the AI produced code they depend on or thinking systematically (and that wont be getting any singularity to bail them out), but that's beside the point.

sdenton4 6 days ago | parent | next [-]

Applying opportunity cost to students is a bit strange...

People need to take time to get good at /something/. It's probably best to work with the systems we have and find the edge where things get hard, and then explore from there. It's partly about building knowledge, but also about gumption and getting some familiarity with how things work.

typewithrhythm 6 days ago | parent | prev | next [-]

But all the other students are doing the same, so the expectation will quickly become use of tools for potentially years.

My introduction to programming was through my dad's outdated PC and an Arduino, and that put me on par with the best funded.

6 days ago | parent | prev [-]
[deleted]
palata 6 days ago | parent | prev | next [-]

> This is one of the sadder things about AI usage getting more standard that I haven’t seen discussed much—-the barrier to entry is now monetary

Agreed. And on the one hand you have those who pay an AI to produce a lot of code, and on the other hand you have those who have to review that code. I already regularly review code that has "strange" issues, and when I say "why does it do this?" the answer is "the AI did it".

Of course, one can pay for the AI and then review and refactor the code to make it good, but my experience is that most don't.

guappa 5 days ago | parent [-]

At my work I'm just never reviewing code of certain team members and I let the team leader do it. Their hire, their problem.

If the code becomes a mess and fixing anything becomes slower… well who cares I'm paid per hour not per task done.

noelwelsh 6 days ago | parent | prev | next [-]

I agree that access is a problem now, but I think it is one that hardware improvements will solve very quickly. We are a few generations of Strix Halo type hardware away from effortlessly running very good LLMs locally. (It's already possible, but the hardware is about $2000 and the LLMs you can run are good but not very good.) AFAIK AMD have not released the roadmap for Medusa Halo, but the rumours [1] are increased CPU and GPU performance, and increased bandwidth. Another iteration or two of this will make Strix Halo hardware more affordable, and the top-of-the-line models will be beasts for local LLMs.

[1]: https://www.notebookcheck.net/Powerful-Zen-6-Medusa-Halo-iGP...

block_dagger 6 days ago | parent | prev | next [-]

LLMs are quickly becoming cheaper. Soon they will be “cheap as free,” to quote Homestar Runner. Then programming will be solved, no need for meatbags. Enjoy the 2-5 years we have left in this profession.

barrell 6 days ago | parent | next [-]

You say that, but subscription prices keep going up. Token price goes down but token count goes up. Companies are burning billions to bring you the existing prices, and multiple hundreds per month is not enough to clear the bar to use these tools.

I’m personally hoping for a future with free local LLMs, and I do hope the prices go down. I also recognize I can do things a little cheaper each year with the API.

However it is far from a guaranteed which direction we’re heading in, and I don’t think we’re on track to get close to removing the monetary barrier anytime soon.

frognumber 6 days ago | parent [-]

My bill for LLMs is going up over time. The more capable, higher-context models dramatically increase my productivity.

The spend prices most of the developing world out -- an programmer earning $10k per year can't pay for a $200/month Claude Max subscription..

And it does better than $6k-$10k programmers in Africa, India, and Asia.

It's the mainframe era all over again, where access to computing is gated by $$$.

achenet 6 days ago | parent [-]

> The spend prices most of the developing world out -- an programmer earning $10k per year can't pay for a $200/month Claude Max subscription..

No, but a computer earning $10k per year can probably afford a $200 used ThinkPad, install Linux on it, build code that helps someone, rent a cheap server from a good cloud provider, advertise their new SaaS on HN, and have it start pulling in enough revenue to pay for a $200 Claude Max subscription.

> It's the mainframe era all over again, where access to computing is gated by $$$.

It's still the internet era, where access to $$$ is gated by computing skill :)

achenet 6 days ago | parent | prev [-]

Did you read the original article?

LLM code still needs to be reviewed by actual thinking humans.

miohtama 6 days ago | parent | prev | next [-]

One can create a free Google account and use Gemini for free.

Or think it this way: It's easy to get base level free LLM (Toyota) but one should not expect free top of the shelf (Porsche).

ubercow13 6 days ago | parent [-]

Previously most Porsche development tools were available to everyone though, such as GCC.

ijidak 6 days ago | parent [-]

Software development costs hundreds of dollars in the 90s. My parents bought VB 6 for $600.

Only in tech are we shocked when things cost money. I don't know that any other industry expects such a reality.

mosselman 6 days ago | parent | prev | next [-]

Does anyone have a good recommendation of a claude code like tool that uses locally hosted models?

heap_perms 2 days ago | parent | next [-]

Yes! https://github.com/musistudio/claude-code-router

Claude Code wrapper to run any models. Even local ones.

jsnider3 6 days ago | parent | prev [-]

I believe gemini-cli can do this. I'm not sure though.

sbarre 6 days ago | parent | prev | next [-]

Eh back in the day computers were expensive and not everyone could afford one (and I don't mean a library computer that you can work on, one you can code and hack on). The ubiquity of computing is not something that's been around forever.

There have always been costs and barriers for the cutting edge.

horacemorace 6 days ago | parent [-]

The problem isn’t cost, it’s reproducibility and understanding. If rely on a service you can’t fully understand to get something done, you’re beholden to the whims of its provider.

sbarre 6 days ago | parent [-]

Sure but that's not what the person I was replying to was talking about, nor what I was talking about.

Cost of access is absolutely a problem in tech.

The problem can certainly be multi-faceted though.

dirkc 6 days ago | parent | prev | next [-]

Maybe local models can address this, but for me the issue is that relying on LLMs for coding introduces gatekeepers.

> Uh oh. We're getting blocked again and I've heard Anthropic has a reputation for shutting down even paid accounts with very few or no warnings.

I'm in the slack community where the author shared their experiment with the autonomous startup and what stuck out to me is that they stopped the experiment out of fear of being suspended.

Something that is fun should not go hand-in-hand with fear of being cut off!

Arisaka1 6 days ago | parent | prev | next [-]

You made me realize exactly why I love skill-based video games, and shun the gacha games (especially those with PvP). You swiped to gain power over players who don't. Yay?

The knowledge check will also slowly transfer towards the borders of fast iteration and not necessarily knowledge depth. The end goal is to make a commodity out of the myth of the 10x dev, and take more leverage away from the devs.

noobermin 6 days ago | parent | prev [-]

This is a pro for a lot of the people whom AI people are targeting: idiots with money.

icemelt8 6 days ago | parent [-]

be careful maybe the idiots will be the only one left with money, and the smart people like you could be homeless.

kubb 6 days ago | parent [-]

If the trend of the last four decades continues, this is the outcome.

georgeburdell 7 days ago | parent | prev | next [-]

For me, I can’t get into using AI tools like Claude Code. As far as I go is chat style where I’m mostly in control. I enjoy the actual process of crafting code myself. For similar reasons, I could never be a manager.

Agents are a boon for extraverts and neurotypical people. If it gets to the point where the industry switches to agents, I’ll probably just find a new career

fsloth 7 days ago | parent | next [-]

I strongly disagree agents are for extroverts.

I do agree it’s definetly a tool category with a unique set of features and am not surprised it’s offputting to some. But it’s appeal is definetly clear to me as an introvert.

For me LLM:s are just a computer interface you can program using natural language.

I think I’m slightly ADD. I love coding _interesting_ things but boring tasks cause extreme discomfort.

Now - I can offload the most boring task to LLM and spend my mental energy on the interesting stuff!

It’s a great time to be a software engineer!

pron 7 days ago | parent | next [-]

> For me LLM:s are just a computer interface you can program using natural language.

I wish they were, but they're not that yet because LLMs aren't very good at logical reasonsing. So it's more like an attempt to program using natural language. Sometimes it does what you ask, sometimes not.

I think "programming" implies that the machine will always do what you tell it, whatever the language, or reliably fail and say it can't be done because the "program" is contradictory, lacks sufficient detail, or doesn't have the necessary permissions/technical capabilities. If it only sometimes does what you ask, then it's not quite programming yet.

> Now - I can offload the most boring task to LLM and spend my mental energy on the interesting stuff!

I wish that, too, were true, and maybe it will be someday soon. But if I need to manually review the agent's output, then it doesn't feel like offloading much aside from the typing. All the same concentration and thought are still required, even for the boring things. If I could at least trust the agent to tell me if it did a good job or is unsure that would have been helpful, but we're not even there yet.

That's not to say the tools aren't useful, but they're not yet "programming in a natural language" and not yet able to "offload" stuff to.

keeda 6 days ago | parent | next [-]

> ... LLMs aren't very good at logical reasonsing.

I'm curious about what experiences led you to that conclusion. IME, LLMs are very good at the type of logical reasoning required for most programming tasks. E.g. I only have to say something like "find the entries with the lowest X and highest Y that have a common Z from these N lists / maps / tables / files / etc." and it spits out mostly correct code instantly. I then review it and for any involved logic, rely on tests (also AI-generated) for correctness, where I find myself reviewing and tweaking the test cases much more than the business logic.

But then I do all that for all code anyway, including my own. So just starting off with a fully-fleshed out chunk of code, which typically looks like what I'd pictured in my head, is a huge load off my cognitive shoulders.

pron 6 days ago | parent | next [-]

The experience was that I once asked an LLM to write a simple function and it produced something very wrong that nothing with good reasoning abilities should ever do. Of course, a drunk or very tired human could have done the same mistake, but they would have at least told me that they were impaired and unsure of their work.

I agree that most of the time it does most simple tasks mostly right, but that's not good enough to truly "offload" my mental effort. Again, I'm not saying it's not useful, but more than working with a junior developer it's like working with a junior developer who may or may not be drunk or tired and doesn't tell you.

But mostly my point is that LLMs seem to do logical reasoning worse than other things they do better, such as generating prose or summarising a document. Of course, even then you can't trust them yet.

> But then I do all that for all code anyway, including my own

I don't, at least not constantly. I review other people's code only towards the very end of a project, and in between I trust that they tell me about any pertinent challenge or insight, precisely so that I can focus on other things unless they draw my attention to something I need to think about.

I still think that working with a coding assistant is interesting and even exciting, but the experience of not being able to trust anything, for me at least, is unlike working with another person or with a tool and doesn't yet allow me to focus on other things. Maybe with more practice I could learn to work with something I can't trust at all.

darkerside 6 days ago | parent | next [-]

> working with a junior developer who may or may not be drunk or tired and doesn't tell you.

Bad news, friend.

Overall though, I think you're right. It's a lot like working with people. The things you might be missing are that you can get better at this with practice, and that once you are multiplexing multiple Claudes, you can become hyper efficient. These are things I'm looking into now.

Do I know these for a fact? Not yet. But, like any tool, I'm sure that the investment won't pay off right away.

kenjackson 6 days ago | parent | prev [-]

What was the simple function?

throwaway31131 6 days ago | parent | next [-]

I’m not sure what their simple function was but I tried to use Claude to recreate C++ code to implement the algorithms in this paper as practice for me in LLM use and it didn’t go well. But I’ll be the first to admit that I’m probably holding it wrong.

https://users.cs.duke.edu/~reif/paper/chen/graph/graph.pdf

pron 6 days ago | parent | prev [-]

Can't remember, but it was something very basic - a 10/15-line routine that a first-year student would write in 3 minutes if they knew the relevant API. The reason I asked the model in the first place is because I didn't know the API. If memory serves, the model inverted an if or a loop condition.

p1esk 6 days ago | parent | next [-]

Did you use one of the latest frontier reasoning models? If not, how is your experience relevant?

totallykvothe 6 days ago | parent | next [-]

In what world is this an appropriate thing to say to someone?

p1esk 6 days ago | parent | next [-]

In the world where you do not claim that LLMs suck today based on your attempt to use some shitty model three years ago.

guappa 6 days ago | parent | prev [-]

In the creed of "AI is perfect, if you claim otherwise you're broken" that so many here embrace.

6 days ago | parent | prev [-]
[deleted]
jama211 6 days ago | parent | prev [-]

So you tried it once and then gave up?

pron 6 days ago | parent [-]

I didn't give up, I just know that I can only use a model when I have the patience to work with something I can't trust at all on anything. So that's what I do.

jama211 5 days ago | parent [-]

Sounds like the spirit of my question remains intact

foobarbecue 6 days ago | parent | prev | next [-]

In your example, you didn't ask the LLM to do any logic. You asked it to translate your logic into code.

Asking an LLM to do logic would be saying something like: "I have a row of a million light switches. They all start off. I start at the beginning and flip on every fourth one. Then I flip on every eighth one, then sixteen, and all the powers of two until I'm over a million. Now I do the same for the powers of three, then four, then five, and so on. How many light switches are on at the end? Do not use any external coding tools for this; use your own reasoning."

Note that the prompt itself is intentionally ambiguous -- a human getting this question should say "I don't understand why you started with every fourth instead of every second. Are you skipping the first integer of every power series or just when the exponent is two?"

When I asked GPT5 to do it, it didn't care about that; instead it complimented me on my "crisp statement of the problem," roughly described a similar problem, and gave a belivable but incorrect answer 270,961 .

I then asked it to write python code to simulate my question. It got the code correct, and said "If you run this, you’ll see it matches the 270,961 result I gave earlier." except, that was a hallucination.

Running the code actually produced 252711.

I guess it went with 270,961 because that was a lexically similar answer to some lexically similar problems in the training data.

cma 6 days ago | parent | next [-]

I would posit that most randomly selected AGIs (people) you ask this of with no tools allowed (pencil and paper) won't get closer on average. This doesn't prove people don't use logic

And allowing python shells for both I think the randomly selected human would do worse on average.

And I also think e.g. Google's IMO model would be able to pass but I have no way to verify that.

pron 6 days ago | parent | next [-]

At work, you can trust people to either get the right answer or tell you they may not have the right answer. If someone is not trustworthy, you don't work with them again. The experience of trying to work with something that is completely not trustworthy on all fronts is novel and entirely dissimilar to working with either people or tools.

keeda 5 days ago | parent [-]

People themselves don't know when they are wrong, and that is why high-functioning organizations have all sorts of guardrails in place. Trivial example, code reviews. Now, code reviews are multi-purpose, and their primary benefit is not just catching bugs, but they do catch bugs pretty often (there are actual studies about this.)

So my experience in working with AI is actually much more similar to working with people, except I have to correct the AI much less frequently.

I always say, AI is technology that behaves like people, and so the trick to working with it effectively is to approach it like working with a colleague, with all their specific quirks and skillsets, rather than a tool.

foobarbecue 6 days ago | parent | prev [-]

That's true about people with pencils and paper. But this is a computer, and my point is that with LLMs, you have a computer program that doesn't compute or do logic -- and for some reason people act like it does.

I mean, this is a super basic python task that I could write in 2 minutes. The naiive implementation is a couple of basic for loops. I wouldn't hire a junior dev that couldn't write it.

My point here is that for someone who can program, it's easier and better to write this in python than to write it in english and have an LLM translate it to python (and LLMs themselves are incapable of logic or reasoning, so the only option is for them to translate it to real code).

cma 6 days ago | parent [-]

I added this line to the end of your prompt so it knew it could clarify and didn't need an immediate answer:

"You can ask any clarifying questions first before starting on the problem."

Here's what it asked (Claude Opus, web search off so it didn't find your post):

> Looking at this problem, I want to make sure I understand the process correctly before diving into the solution. Let me clarify a few things:

> When you say "flip on every fourth one" for the powers of two, do you mean:

> We flip switches at positions 4, 8, 12, 16, 20, 24, ... (every 4th position) Then flip switches at positions 8, 16, 24, 32, 40, 48, ... (every 8th position) And so on?

> When you move to "powers of three", do you mean:

> Start with every 3rd switch (3, 6, 9, 12, ...) Then every 9th switch (9, 18, 27, 36, ...) Then every 27th switch, etc.?

> Or does "powers of three" mean we start with 3² = 9 as the first increment, similar to how we started with 2² = 4 for powers of two?

> When you say "then four, then five, and so on" - do you mean we continue this process for all integers k ≥ 2, using powers of k as our increments?

> The final state of each switch depends on whether it gets flipped an odd or even number of times total, so I want to make sure I understand the flipping pattern correctly before calculating the answer.

keeda 5 days ago | parent | prev [-]

That may be the case, but how often do you have to provide these kinds of puzzles as programming tasks? And even then, by your own example, it did product correct code, so it did "reason" through the problem statement!

foobarbecue 5 days ago | parent [-]

The point I'm making is that translating from english to python does not involve logical reasoning or any form of problem solving. You essentially replace english phrases with their python equivalents; it's a fuzzy version of looking things up in a dictionary.

It can translate, but it cannot reason or calculate. The python can calculate.

keeda 4 days ago | parent [-]

I've seen countless examples where it does do a lot of logical reasoning from a very high-level requirement. I only gave that example because that is the more common form of programming logic required, but there is a huge range of logical and coding tasks it handles seamlessly. It figures out the problem, the high-level approach, the appropriate algorithms, any intermediate book-keeping data structures required, any edge cases, what functions and classes from the contextual code matters, and how it should integrate with the rest of the codebase.

Here is one recent exchange (some proprietary details redacted) with Claude Sonnet 3.7 via the Cody VSCode Extension (RIP):

---

Me: At this point, "segments" represents a list of bounding boxes that can overlap. Write a function to coalesce only those boxes that are substantially (i.e. within some tolerance) horizontally aligned along one or both edges or their centers, either: 1) overlapping completely or partially; or 2) very closely spaced like the words of a sentence should be (you can assume the threshold for spacing could be a reasonable fraction of the height of the box).

Claude: I'll create a function to coalesce horizontally aligned boxes based on your requirements. This function will identify boxes that are either overlapping or closely spaced and horizontally aligned (by top edge, bottom edge, or vertical center).

```javascript function coalesceHorizontallyAlignedSegments(segments, options = {}) { // Default options const { verticalAlignmentTolerance = 0.1, // 10% tolerance for vertical alignment horizontalSpacingThreshold = 0.5, // Max horizontal spacing as a fraction of box height overlapThreshold = 0.2 // Minimum overlap required (20% of the smaller box) } = options;

  if (!segments || segments.length <= 1) {
    return segments;
  }

  // Sort segments by x position (left to right)
  const sortedSegments = [...segments].sort((a, b) => a.x - b.x);

  // Track which segments have been merged
  const merged = new Array(sortedSegments.length).fill(false);
  const mergedSegments = [];

  for (let i = 0; i < sortedSegments.length; i++) {
    // Skip if this segment has already been merged
    if (merged[i]) continue;

    const currentSegment = sortedSegments[i];
    let mergedSegment = { ...currentSegment };
    merged[i] = true;

    // Check for potential merges with subsequent segments
    let didMerge = true;

    while (didMerge) {
      didMerge = false;

      for (let j = 0; j < sortedSegments.length; j++) {
        if (merged[j]) continue;

        const nextSegment = sortedSegments[j];

        // Check if horizontally aligned (top, bottom, or center)
        const topAligned = Math.abs(mergedSegment.y - nextSegment.y) <=
                           verticalAlignmentTolerance * Math.min(mergedSegment.height, nextSegment.height);

        const bottomAligned = Math.abs((mergedSegment.y + mergedSegment.height) -
                              (nextSegment.y + nextSegment.height)) <=
                              verticalAlignmentTolerance * Math.min(mergedSegment.height, nextSegment.height);

        const centerAligned = Math.abs((mergedSegment.y + mergedSegment.height/2) -
                              (nextSegment.y + nextSegment.height/2)) <=
                              verticalAlignmentTolerance * Math.min(mergedSegment.height, nextSegment.height);

        const isVerticallyAligned = topAligned || bottomAligned || centerAligned;

        if (!isVerticallyAligned) continue;

        // Check for overlap or close spacing
        const mergedRight = mergedSegment.x + mergedSegment.width;
        const nextLeft = nextSegment.x;

        // Calculate horizontal spacing
        const horizontalSpacing = nextLeft - mergedRight;
        const maxAllowedSpacing = horizontalSpacingThreshold * Math.min(mergedSegment.height, nextSegment.height);

        // Check for overlap
        const isOverlapping = mergedRight >= nextLeft;

        // Calculate overlap percentage if overlapping
        let hasSignificantOverlap = false;
        if (isOverlapping) {
          const overlapWidth = Math.min(mergedRight, nextSegment.x + nextSegment.width) -
                              Math.max(mergedSegment.x, nextSegment.x);
          const smallerWidth = Math.min(mergedSegment.width, nextSegment.width);
          hasSignificantOverlap = overlapWidth / smallerWidth >= overlapThreshold;
        }

        // Merge if overlapping significantly or closely spaced
        if (hasSignificantOverlap || (horizontalSpacing >= 0 && horizontalSpacing <= maxAllowedSpacing)) {
          // Create a new bounding box that encompasses both segments
          const newX = Math.min(mergedSegment.x, nextSegment.x);
          const newY = Math.min(mergedSegment.y, nextSegment.y);
          const newRight = Math.max(mergedSegment.x + mergedSegment.width,
                                   nextSegment.x + nextSegment.width);
          const newBottom = Math.max(mergedSegment.y + mergedSegment.height,
                                    nextSegment.y + nextSegment.height);

          mergedSegment = {
            x: newX,
            y: newY,
            width: newRight - newX,
            height: newBottom - newY,
            // Preserve the virtual path from the leftmost segment
            imagePath: mergedSegment.x <= nextSegment.x ? mergedSegment.imagePath : nextSegment.imagePath
          };

          merged[j] = true;
          didMerge = true;
          break;
        }
      }
    }

    // Add the topLeft and bottomRight properties for consistency
    mergedSegment.topLeft = { x: mergedSegment.x, y: mergedSegment.y };
    mergedSegment.bottomRight = {
      x: mergedSegment.x + mergedSegment.width,
      y: mergedSegment.y + mergedSegment.height
    };

    mergedSegments.push(mergedSegment);
  }

  return mergedSegments;
} ```

This function works as follows:

1. It sorts segments from left to right to process them in reading order 2. For each segment, it looks for potential merges with other segments based on: - Vertical alignment (top edges, bottom edges, or centers within tolerance) - Either significant overlap (controlled by overlapThreshold) or - Close horizontal spacing (controlled by horizontalSpacingThreshold as a fraction of box height) 3. When segments are merged, it creates a new bounding box that encompasses both 4. It continues merging until no more merges are possible

You can integrate this function into your segmentation pipeline by adding it after the segment extraction:

```javascript function segmentImage(ctx, imageData, options = {}) { // ... existing code ...

  const segments = boundingRects.map(toSegment);

  // Coalesce horizontally aligned segments
  const coalescedSegments = coalesceHorizontallyAlignedSegments(segments, options.coalesceOptions);

  return { segments: coalescedSegments, contourTrees };
} ```

Me: Do not write any code yet, but would it help if the segments were first grouped into approximate "rows" and then coalesced, rather than simply sorted by the x-coordinate?

Claude: Yes, grouping segments into approximate "rows" before coalescing would be significantly more effective than simply sorting by x-coordinate. Here's why:

1. *Better handling of multi-column layouts*: When a page has multiple columns of text or elements, sorting purely by x-coordinate would attempt to merge segments across different columns if they happen to be horizontally aligned.

2. *Respecting natural reading flow*: Text typically flows in rows, and grouping by rows first would respect this natural organization of content.

3. *More accurate merging decisions*: By first establishing which segments belong to the same "row" or line of content, we can make more informed decisions about which segments should be considered for merging.

4. *Handling of varying vertical positions*: Text lines on a webpage aren't always perfectly aligned. Pre-grouping into rows would allow us to handle slight variations in vertical positioning within what should be considered the same line.

5. *Efficiency*: By limiting the coalescing operations to segments within the same row, we reduce the number of comparisons needed.

The approach would work like this: 1. Group segments into rows based on vertical overlap or proximity 2. Within each row, sort segments by x-coordinate 3. Apply the coalescing logic only to segments within the same row

This approach would better preserve the semantic structure of the content while still merging segments that represent parts of the same logical unit (like words in a sentence).

Looking at the codebase, this approach would align well with how visual content is typically structured and would produce more meaningful segment groupings for downstream processing.

---

And then it proceeded to update the code as discussed. Sure this is not a very novel problem and it required a bit of back and forth, but look at the ratio of prompt to code. This exchange took a couple of minutes; I'd estimate it would have taken me an hour to get that code down with all edge cases handled. Look at the exposition, the quality of code, the choice to use optional parameters for tolerances, and edge-case handling.

It's very, very hard for me to not see this as reasoning. I mean, how is this not mind-blowing?

__MatrixMan__ 6 days ago | parent | prev | next [-]

I'm not who you're replying to but I had a scenario where I needed to notice that a command had completed (exit code received) but keep listening for any output that was still buffered and only stop processing tokens after it had been quiet for a little bit.

Trying to get Claude to do this without introducing a deadlock and without exiting too early and leaving valuable output in the pipe was hellish.

It's very good at some kinds of reasoning and very bad at others. There's not much it's mediocre at.

keeda 5 days ago | parent [-]

> It's very good at some kinds of reasoning and very bad at others. There's not much it's mediocre at.

Yes, one of the terms for this phenomenon is "jagged frontier" from here:

https://www.oneusefulthing.org/p/centaurs-and-cyborgs-on-the...

This is also the main reason people have such diverging experiences and the discussion around AI is so bimodal in nature.

jpfromlondon 5 days ago | parent | prev [-]

https://arstechnica.com/ai/2025/08/researchers-find-llms-are...

yomismoaqui 7 days ago | parent | prev | next [-]

You can view Claude Code as a non-deterministic compiler where you input english and get functioning code on the other end.

The non-determinism is not as much as a problem because you are reading over the results and validating that what it is created matches what you tell it to do.

I'm not talking about vibe-coding here, I'm grabbing the steering wheel with both hands because this car allows me to go faster than if I was driving myself, but sometimes you have to steer or brake. And the analogy favors Claude Code here because you don't have to react in milliseconds while programming.

TL;DR: if you do the commit you are responsible for the code it contains.

pron 7 days ago | parent [-]

Sure, and that may be valuable, but it's neither "programming" nor "offloading mental effort" (at least not much).

Some have compared it to working with a very junior programmer. I haven't done that in a long while, but when I did, it didn't really feel like I was "offloading" much, and I could still trust even the most junior programmer to tell me whether the job was done well or not (and of any difficulties they encountered or insight they've learnt) much more than I can an agent, at least today.

Trust is something we have, for the most part, when we work with either other people or with tools. Working without (or with little) trust is something quite novel. Personally, I don't mind that an agent can't accomplish many tasks; I mind a great deal that I can't trust it to tell me whether it was able to do what I asked or not.

fsloth 6 days ago | parent | next [-]

”it's neither "programming"

Sure it is.

Modern ecosystem is sadly full of API:s like WPF on Windows that are both verbose and configuration heavy. Now, some people may be able to internalize xaml with little effort but not all us - and then you basically move forward iteratively, looking for code example, trying this or that … basically random walking towards something usable.

Or you use an agentic LLM and it does this peeking and poking for you, and with decades old APIs like WPF likely has enough context to do the thing you asked it to do far more competently than you could train yourself to program WPF in a few days.

Of course in the context of this example WPF was your main duty, you _would_ learn the ins and outs of it.

In quite many jobs a task like this may not infact be even among your top ten duties but you need to do it.

In these sort of situations a LLM is really nice. The worst it provides is a good first guess how to do something.

If your contex is something like radiology treatment then no, don’t use LLM!

But there are thoushands of miserable non-critical but necessary components in production for which LLM is just fine.

gf000 6 days ago | parent | prev [-]

There is probably a significant factor of what domain you work in, that might be part of the reason why you have different experiences.

I don't know your current domain, but stuff like Loom requires very complex reasoning capabilities, most of which lives outside the actual codebase itself. Business code on the other hand often has a more direct mapping from idea to code, so LLMs might show more promise there.

nostrademons 6 days ago | parent | prev [-]

So here's a transcript of a recent Claude Code interaction I had, trying to build an MP3 organizer that relies on heuristics and fuzzy matching to make sense out of all the random MP3s I downloaded in the 90s. This all happened in-between prompts: I gave it a correct example and an incorrect example and asked it why it was giving the incorrect example.

● Let me trace through what happens with the "Jordan Rudess" example:

  1. Directory: Jordan Rudess/
  2. file_rel_path would be Path("Jordan Rudess") (after removing MP3 Disk prefix)
  3. If Jordan Rudess matches in path component 0, scope_path would be Path("Jordan Rudess")
  4. _path_is_in_scope checks if "Jordan Rudess".startswith("Jordan Rudess") → True

  This looks correct. Let me check the inference logic:
● Read(mp3_organizer_v2.py) ⎿ Read 15 lines (ctrl+r to expand)

● The inference priority looks correct - path_component_match is indeed the highest priority. But I think I see the issue now. Let me look at the actual logic flow in the main organize method:

● Read(mp3_organizer_v2.py) ⎿ Read 10 lines (ctrl+r to expand)

● The flow looks right. Let me examine the actual problem more carefully. Looking back at the _apply_path_component_matching method, I think there might be an issue with when the matching happens. Let me check the logic:

● Update Todos ⎿ Check _path_is_in_scope logic for correctness Verify artist inference priority logic works with path_component_match Test and fix any remaining issues with artist propagation

● I think I see the issue now. The problem is that the current code only looks for matches in files that already have file_id3 tags, but it needs to propagate those matches to all files in the directory. However, the current implementation is still only applying matches to files that have matching candidates, not ALL files. Let me check the exact logic:

● Read(mp3_organizer_v2.py) ⎿ Read 8 lines (ctrl+r to expand)

● Actually, looking at this more carefully, the logic should work. The matched_artists dictionary contains the matches found from ANY file that had a file_id3 tag matching a path component, and then it applies that match to ALL files in the file_infos list via the loop at line 737.

  Wait, let me understand the scoping issue better. Let me check what file_rel_path and scope_path would actually be in your example:
● Read(mp3_organizer_v2.py) ⎿ Read 5 lines (ctrl+r to expand)

● Ah! I think I see a potential issue. The file_rel_path is calculated as the directory path, not the individual file path. So for all files in the same directory, file_rel_path will be the same. This should actually work correctly.

  But let me double-check the scope matching logic. If the directory is "Jordan Rudess" and the scope_path is also "Jordan Rudess", then _path_is_in_scope should return
  True.

  Let me debug this by adding some debug output to see what's actually happening. But first, let me look at a potential issue: maybe the match isn't being found in the first place?
It's doing a pretty complex chain of step-by-step reasoning here, looking in detail at the code. It ended up by printing out more debug info and having me re-run it again, then paste in the debug info, then add even more debug info and run it again. But it did eventually get the bug, which was non-trivial to identify (it was an aliasing problem where Claude was mutating state on a list and that list was shared with all the other files in the directory).
Applejinx 6 days ago | parent [-]

Huh. Alternate explanation: there's a layer of indirection, drawing upon the unthinkable size of the source data, so rather than 'issue forth tokens as if there is a person answering a question', you've got 'issue forth tokens as if there is a person being challenged to talk about their process', something that's also in the training data but in different contexts.

I'm not sure statements of 'aha, I see it now!' are meaningful in this context. Surely this is just the em-dash of 'issue tokens to have the user react like you're thinking'?

nostrademons 5 days ago | parent [-]

I wonder if something else is going on, and perhaps Claude is using the LLM to identify the likely culprits within the codebase, sending the code around them to execute with an actual Python interpreter on their servers, feeding both the code and the result as the context window to another LLM query with a system prompt something like "What is this code doing, when it runs on this input and this output?", feeding the result of that back to the user, and then repeating as long as the overall bug remains unsolved. I've found that feedback is a very effective technique with LLMs, asking them to extract some data, testing that data through out-of-band mechanisms, then feeding the test results and the original context back into the LLM to explain its reasoning and why it got the result. The attention mechanisms in the transformer model function very well when they're prompted with specifics and asked to explain their reasoning.

Only an Anthropic engineer would know for sure. I'm pretty sure that it was making multiple queries on my behalf during the chat transcript - each "Read ... mp3organizer_v2.py" is a separate network round-trip.

thewebguyd 7 days ago | parent | prev | next [-]

> For me LLM:s are just a computer interface you can program using natural language.

Sort of. You still can't get a reliable output for the same input. For example, I was toying with using ChatGPT with some Siri shortcuts on my iPhone. I do photography on the side, and finding a good time for lighting for photoshoots is a usecase I use a lot so I made a shortcut which sends my location to the API along with a prompt to get the sunset time for today, total amount of daylight, and golden hour times.

Sometimes it works, sometimes it says "I don't have specific golden hour times, but you can find those on the web" or a useless generic "Golden hour is typically 1 hour before sunset but can vary with location and season"

Doesn't feel like programming to me, as I can't get reproducible output.

I could just use the LLM to write some API calling script from some service that has that data, but then why bother with that middle man step.

I like LLMs, I think they are useful, I use them everyday but what I want is a way to get consistent, reproducible output for any given input/prompt.

stirfish 6 days ago | parent [-]

For things where I don't want creativity, I tell it to write a script.

For example, "write a comprehensive spec for a script that takes in the date and a location and computes when golden hour is." | "Implement this spec"

That variability is nice when you want some creativity, e.g. "write a beautiful, interactive boids simulation as a single file in html, css, and JavaScript."

Words like "beautiful" and interactive" are open to interpretation, and I've been happy with the different ways they are interpreted.

klipklop 7 days ago | parent | prev | next [-]

>I think I’m slightly ADD. I love coding _interesting_ things but boring tasks cause extreme discomfort. >Now - I can offload the most boring task to LLM and spend my mental energy on the interesting stuff!

I agree and I feel that having LLM's do boilerplate type stuff is fantastic for ADD people. The dopamine hit you get making tremendous progress before you get utterly bored is nice. The thing that ADD/ADHD people are the WORST at is finishing projects. LLM will help them once the thrill of prototyping a green-field project is over.

80hd 7 days ago | parent | next [-]

Seconding this. My work has had the same problem - by the time I've got things all hooked up, figured out the complicated stuff - my brain (and body) clock out and I have to drag myself through hell to get to 100%. Even with ADHD stimulant medication. It didn't make it emotionally easier, just _possible_ lol.

LLMs, particularly Claude 4 and now GPT-5 are fantastic at working through these todo lists of tiny details. Perfectionism + ADHD not a fun combo, but it's way more bearable. It will only get better.

We have a huge moat in front of us of ever-more interesting tasks as LLMs race to pick up the pieces. I've never been more excited about the future of tech

r_lee 6 days ago | parent [-]

Same here, especially for making bash scripts or lots of if this if that with logging type stuff, error handling etc..

Oh and also, from what I know, ADHD and perfectionism is a very common combination, I'm not sure if everyone has that but I've heard it's the case for many with ADD. Same with "standards" being extremely high for everything

whartung 6 days ago | parent | prev | next [-]

I'm kind of in this cohort. While in the groove, yea, things fly but, inevitably, my interest wanes. Either something too tedious, something too hard (or just a lot of work). Or, just something shinier shows up.

Bunch of 80% projects with, as you mentioned, the interesting parts finished (sorta -- you see the line at the end of the tunnel, it's bright, just don't bother finishing the journey).

However, at the same time, there's conflict.

Consider (one of) my current projects, I did the whole back end. I had ChatGPT help me stand up a web front end for it. I am not a "web person". GUIs and what not are a REAL struggle for me because on the one hand, I don't care how things look, but, on the other, "boy that sure looks better". But getting from "functional" to "looks better" is a bottomless chasm of yak shaving, bike shedding improvements. I'm even bad at copying styles.

My initial UI was time invested getting my UI to work, ugly as it was, with guidance from ChatGPT. Which means it gave me ways to do things, but mostly I coded up the actual work -- even if it was blindly typing it in vs just raw cut and paste. I understood how things were working, what it was doing, etc.

But then, I just got tired of it, and "this needs to be Better". So, I grabbed Claude and let it have its way.

And, its better! it certainly looks better, more features. It's head and shoulders better.

Claude wrote 2-3000 lines of javascript. In, like, 45m. It was very fast, very responsive. One thing Claude knows is boiler plate JS Web stuff. And the code looks OK to me. Imperfect, but absolutely functional.

But, I have zero investment in the code. No "ownership", certainly no pride. You know that little hit you get when you get Something Right, and it Works? None of that. Its amazing, its useful, its just not mine. And that's really weird.

I've been striving to finish projects, and, yea, for me, that's really hard. There is just SO MUCH necessary to ship. AI may be able to help polish stuff up, we'll see as I move forward. If nothing else it may help gathering up lists of stuff I miss to do.

brailsafe 6 days ago | parent | prev | next [-]

Ironically, I find greenfield projects the least stimulating and the most rote, aside from thinking about system design.

I've always much preferred figuring out how to improve or build on existing messy systems and codebases, which is certainly aided by LLMs for big refactoring type stuff, but to be successful at it requires thinking about how some component of a system is already used and the complexity of that. Lots of edge cases and nuances, people problems, relative conservativeness.

skc 6 days ago | parent | prev [-]

Looks like the definition of boilerplate will continue to shift up the chain

lsaferite 6 days ago | parent | prev | next [-]

I find Claude great at all of the boilerplate needed to get testing in place. It's also pretty good at divining test cases to lock in the current behavior, even if it's buggy. I use Claude as a first pass on tests, then I run through each test case myself to make sure it's a meaningful test. I've let it loose on the code coverage loop as well, so it can drill in and get those uncommon lines covered. I still don't have a good process for path coverage, but I'm not sure how easy that is in go as I haven't checked into it much yet.

I'm with you 100% on the boring stuff. It's generally good at the boring stuff *because* it's boring and well-trod.

amelius 6 days ago | parent | prev | next [-]

Last week there was this post about flow state, and pretty much explains the issue:

https://news.ycombinator.com/item?id=44811457

vlan0 6 days ago | parent [-]

What is described in that post is a skill issue. The attention issues spoke about can be worked through. Because attention is a skill like anything else. And if one does not learn to practice. One becomes victim to it

WhyOhWhyQ 7 days ago | parent | prev | next [-]

It's interesting that every task in the world is boring to somebody, which means nothing left in the world will be done by those interested in it, because somebody will gladly shotgun it with an AI tool.

sleepybrett 6 days ago | parent | prev | next [-]

Programming implies that it's going to do what i say. I wish it did.

kiitos 6 days ago | parent | prev [-]

> For me LLM:s are just a computer interface you can program using natural language. ... boring tasks cause extreme discomfort ... Now - I can offload the most boring task to LLM and spend my mental energy on the interesting stuff!

The problem with this perspective, is that when you try to offload exactly the same boring task(s), to exactly the same LLM, the results you get back are never even close to being the same. This work you're offloading via natural language prompting is not programming in any meaningful sense.

Many people don't care about this non-determinism. Some, because they don't have enough knowledge to identify, much less evaluate, the consequent problems. Others, because they're happy to deal with those problems, under the belief that they are a cost that's worth the net benefit provided by the LLM.

And there are also many people who do care about this non-determinism, and aren't willing to accept the consequent problems.

Bluntly, I don't think that anyone in group (1) can call themselves a software engineer.

filoleg 7 days ago | parent | prev | next [-]

> Agents are a boon for extraverts and neurotypical people.

This sounds like a wild generalization.

I am in neither of those two groups, and I’ve been finding tools like Claude Code becoming increasingly more useful over time.

Made me much more optimistic about the direction of AI development in general too. Because with each iteration and new version it isn’t getting anywhere closer to replacing me or my colleagues, but it is becoming more and more useful and helpful to my workflow.

And I am not one of those people who are into “prompt engineering” or typing novels into the AI chatbox. My entire interaction is typically short 2-3 sentences “do this and that, make sure that XYZ is ABC”, attach the files that are relevant, let it do its thing, and then manual checks/adjustments. Saves me a boatload of work tbh, as I enjoy the debugging/fixing/“getting the nuanced details right” aspect of writing code (and am pretty decent at it, I think), but absolutely dread starting from a brand new empty file.

Terretta 7 days ago | parent | prev | next [-]

> I can’t get into using AI tools like Claude Code. As far as I go is chat style where I’m mostly in control.

Try aider.chat (it's in the name), but specifically start with "ask" mode then dip a toe into "architect" mode, not "code" which is where Claude Code and the "vibe" nonsense is.

Let aider.chat use Opus 4.1 or GPT-5 for thinking, with no limit on reasoning tokens and --reasoning-effort high.

> agents are a boon for extraverts and neurotypical people.

On the contrary, I think the non-vibe tools are force multipliers for those with an ability to communicate so precisely they find “extraverts and neurotypical people” confounding when attempting to specify engineering work.

I'd put both aider.chat and Claude Code in the non-vibe class if you use them Socratically.

dionian 7 days ago | parent [-]

thanks for this, going to try it out - i need to use paid api and not my claude max or gpt pro subn, right?

victorbjorklund 6 days ago | parent [-]

Aider actually has a mode that is called "copy-paste" where it basically gives you a context to paste in an LLM chat and then you copy-paste back the reply to aider

mock-possum 6 days ago | parent [-]

Sounds like extra steps

wredcoll 6 days ago | parent | prev | next [-]

> Agents are a boon for extraverts and neurotypical people.

Please stop with this kind of thing. It isn't true, it doesn't make sense and it doesn't help anyone.

taftster 6 days ago | parent | prev | next [-]

For me (an introvert), I have found great value in these tools. Normally, I kind of talk to myself about a problem / algorithm / code segment as I'm fleshing it out. I'm not telling myself complete sentences, but there's some sort of logical dialog I am having with myself.

So I just have to convert that conversation into an AI prompt, basically. It just kind of does the typing for the construct already in my head. The trick is to just get the words out of my head as prompt input.

That's honestly not much different than an author writing a book, for example. The story line is in their head, they just have to get it on paper. And that's really the tricky part of writing a novel as much as writing code.

I therefore don't believe this is an introvert/extrovert thing. There are plenty of book authors which are both. The tools available as AI code agents are really just an advanced form of dictation.

MrDarcy 6 days ago | parent | prev | next [-]

For what it’s worth I’m neurodivergent, introverted and have avoided management up to the staff+level. Claude Code is great I use it all day every day now.

kevinsync 7 days ago | parent | prev | next [-]

I kind of think we will see some industry attrition as a result of LLM coding and agent usage, simply because the ~vIbEs~ I'm witnessing boil down to quite a lot of resistance (for multiple reasons: stubbornness, ethics, exhaustion from the hype cycle, sticking with what you know, etc)

The thing is, they're just tools. You can choose to learn them, or not. They aren't going to make or break your career. People will do fine with and without them.

I do think it's worth learning new tools though, even if you're just a casual observer / conscientious objector -- the world is changing fast, for better or worse, and you'll be better prepared to do anything with a wider breadth of tech skill and experience than with less. And I'm not just talking about writing software for a living, you could go full Uncle Ted and be a farmer or a carpenter or a barista in the middle of nowhere, and you're going to be way better equipped to deal with logistical issues that WILL arise from the very nature of the planet hurtling towards 100% computerization. Inventory management, crop planning, point of sale, marketing, monitoring sensors on your brewery vats, whatever.

Another thought I had was that introverts often blame their deficits in sales, marketing and customer service on their introversion, but what if you could deploy an agent to either guide, perform, or prompt (the human) with some of those activities? I'd argue that it would be worth the time to kick the tires and see what's possible there.

It feels like early times still with some of these pie in the sky ideas, but just because it's not turn-key YET doesn't mean it won't be in the near future. Just food for thought!

HardCodedBias 7 days ago | parent [-]

"ethics"

I agree with all of your reasons but this one sticks out. Is this a big issue? Are many people refusing to use LLMs due to (I'm guessing here): perceived copyright issues, or power usage, or maybe that they think that automation is unjust?

kevinsync 6 days ago | parent | next [-]

I can't tell how widespread any of is, to be honest.. mostly because it's anecdata, and impossible to determine if what I'm seeing is just ragebait, or shallow dunks by reply-guys in comment sections, or particularly-loud voices on social media that aren't representative of the majority opinion, etc

That said, the amount of sort-of-thoughtless, I'm-just-repeating-something-I-heard-but-don't-really-understand outrage towards AI that I'm seeing appears to be increasing -- "how many bottles of water did that slop image waste??", "Clanker"-adjacent memes and commentary (include self-driving + robots in this category), people ranting about broligarchs stealing art, music, movies, books to train their models (oddly often while also performatively parroting party lines about how Spotify rips artists off), all the way to refusing to interact with people on dating apps if they have anything AI in their profiles hahaha (file "AI" alongside men holding fish in their pics, and "crypto" lol)

It's all chronically-online nonsense that may well just be perception that's artificially amplified by "the algorithm".

Me, I have no fundamental issue with any of it -- LLMs, like anything else, aren't categorically good or bad. They can be used positively and negatively. Everything we use and consume has hidden downsides and unsavory circumstances.

codingdave 6 days ago | parent | prev [-]

Yes, people are refusing for those reasons. I don't know how many, but I'd say about half of the the people I know who do not work in tech are rejecting AI, with ethics being the primary reason. That is all just anecdata, but I suspect the tech bubble around AI is making people in tech underestimate how many people in the world simply are not interested in it being part of their lives.

burnte 6 days ago | parent | prev | next [-]

> Agents are a boon for extraverts and neurotypical people.

As an extrovert the chances I'll use an AI agent in the next year is zero. Not even a billion to one but a straight zero. I understand very well how AI works, and as such I have absolutely no trust in it for anything that isn't easy/simple/solved, which means I have virtually no use for generative AI. Search, reference, data transformation, sure. Coding? Not without verification or being able to understand the code.

I can't even trust Google Maps to give me a reliable route anymore, why would I actually believe some AI model can code? AI tools are helpers, not workers.

ragequittah 6 days ago | parent [-]

>no trust in it for anything that isn't easy/simple/solved

I'm not sure what part of programming isn't generally solved thousands of times over for most languages out there. I'm only using it for lowly web development but I can tell you that it can definitely do it at a level that surprises me. It's not just "auto-complete" it's actually able to 'think' over code I've broken or code that I want improved and give me not just one but multiple paths to make it better.

burnte 6 days ago | parent [-]

In the case of programming is not quite as problematic with unsolved problems as much as others, like completeness. In the case of programming, it's context and understanding. It's great for small chunks of code but people think you can vibe code entire interactive applications with no programming knowledge, but LLMs simply don't understand, so they can't keep a cohesive idea of what the end goal is in mind. The larger the codebase it needs to work on the more likely it is to make catastrophic errors, create massive security flaws, or just generate nonfunctional code.

Programming LLMs will become awesome when we create more narrowly targeted LLMs rather than these "train on everything" models.

sixo 6 days ago | parent | prev | next [-]

At one point in my life I liked crafting code. I took a break, came back, and I no longer liked it--my thoughts ranged further, and the fine-grained details of implementations were a nuisance rather than ~pleasurable to deal with.

Whatever you like is probably what you should be doing right now. Nothing wrong with that.

block_dagger 6 days ago | parent | prev | next [-]

I bet your code sucks in quality and quantity compared to the senior+ engineer who uses the modern tools. My code certainly did even after 20 years of experience, much of that as senior/staff level at well paying companies.

joshred 7 days ago | parent | prev | next [-]

I think they're fantastic at generating the sort of thing I don't like writing out. For example, a dictionary mapping state names to their abbreviations, or extracting a data dictionary from a pdf so that I can include it with my documentation.

mock-possum 6 days ago | parent | prev | next [-]

It is effin nutzo that you would try to relate chatting with AI and agentic LLM codegen workflows to the intra/extra vert dichotomy or to neuro a/typicality - you so casually lean way into this absolute spectrum that I don’t even think associates the way you think it does, and it’s honestly kind of unsettling, like - what do you think you know about me, and about My People, that apparently I don’t know??

If it doesn’t work for you that’s fine, but turning it into some tribalised over-generalization is just… why, why would you do that, who is that kind of thing useful for??

wahnfrieden 7 days ago | parent | prev | next [-]

You are leaving a lot of productivity on the table by not parallelizing agents for any of your work. Seemingly for psychological comfort quirks rather than earnestly seeking results.

Automation productivity doesn’t remove your own agency. It frees more time for you to apply your desire for control more discerningly.

dexterlagan 6 days ago | parent [-]

I can imagine there are plenty of use cases, but I could not find one for myself. Can you give an example?

wahnfrieden 6 days ago | parent [-]

I personally use it for a lot of SwiftUI work. I parallelize it across at least 3 projects at once. I use only the largest models on highest thinking modes. I give instruction on implementation and provide reference implementations.

I also use it for adding features/feature polish that address user pain points but that I can't prioritize for my own manual work just yet. There are a variety of user requests which sometimes LLMs are able to get done very suddenly when I give it a shot quickly. For these tasks, it's ok to abandon & defer them if the LLM spins its wheels.

klipklop 7 days ago | parent | prev | next [-]

>Agents are a boon for extraverts and neurotypical people.

I completely disagree. Juggling several agents (and hopping from feature-to-feature) at once, is perfect for somebody with ADHD. Being an agent wrangler is great for introverts instead of having to talk to actual people.

cpldcpu 7 days ago | parent | prev | next [-]

I think you misunderstand what this does. It is not only a coding agent. It is an abstraction layer between you and the computer.

starfallg 7 days ago | parent | prev | next [-]

Pretty sure we can make LLM agents to transform declarative inputs to agentic action.

garciasn 7 days ago | parent | prev | next [-]

Agents are boon for introverts who fucking hate dealing with other people (read: me). I can iterate rapidly with another 'entity' in a technical fashion and not have to spend hours explaining in relatable language what to do next.

I feel as if you need to work with these things more, as you would prefer to work, and see just how good they are.

bastawhiz 7 days ago | parent | prev [-]

> Agents are a boon for extraverts and neurotypical people

As a neurodivergent introvert, please don't speak for the rest of us.

antonvs 6 days ago | parent [-]

That stuck out to me as well. People will make up all sorts of stories to justify their resistance to change.

bastawhiz 6 days ago | parent [-]

It's the same as saying that writing good commit messages is a boon for extroverts and neurotypicals. It's a computer. You're giving it instructions, and the only difference to traditional coding is that the input is English text.

joks 5 days ago | parent | prev | next [-]

> It's just fun to mess with these tools

I think this is the main sentiment I can't wrap my head around. Using Claude Code or Cursor has been entirely a mind-numbingly tedious experience to me (even when it's been useful.) It's often faster, but 80% of the time is spent just sitting there waiting for it to finish working, and I'm not proud of the result because I didn't do anything except come up with the idea and figure out how to describe it well. It just ends up feeling like the coding equivalent of...like...copying down answers to cheat on a test. Not in the sense that it feels gross and wrong and immoral, but in the sense that it's unsatisfying and unfulfilling and I don't feel any pride in the work I've done.

For things where I just want something that does something I need as quickly as possible, sure, I wasn't going to care either way, but personal projects are where I find myself least wanting to vibe code anything. It feels like hiring someone else to do my hobbies for me.

pyrale 7 days ago | parent | prev | next [-]

On one hand, I agree with you that there is some fun in experimenting with silly stuff. On the other hand...

> Claude was trying to promote the startup on Hackernews without my sign off. [...] Then I posted its stuff to Hacker News and Reddit.

...I have the feeling that this kind of fun experiments is just setting up an automated firehose of shit to spray places where fellow humans congregate. And I have the feeling that it has stopped being fun a while ago for the fellow humans being sprayed.

the__alchemist 7 days ago | parent | next [-]

This is an excellent point that will immediately go off-topic for this thread. We are, I believe, committed, into a mire of CG content enveloping the internet. I believe we will go through a period where internet communications (like HN, Reddit, and pages indexed by search engines) in unviable. Life will go on; we will just be offline more. Then, the defense systems will be up to snuff, and we will find a stable balance.

mettamage 7 days ago | parent | next [-]

I hope you're right. I don't think you will be, AI will be too good at impersonating humans.

lukan 7 days ago | parent [-]

"we will just be offline more"

I think it will be quite some time into the future, before AI can impersonate humans in real life. Neither hardware, nor software is there, maybe something to fool humans for a first glance maybe, but nothing that would be convincing for a real interaction.

theshrike79 7 days ago | parent | prev | next [-]

My theory (and hope) is the rise of a web of trust system.

Implemented so that if a person in your web vouches for a specific url (“this is made by a human”) you can see it in your browser.

Analemma_ 7 days ago | parent | next [-]

If your solution to this problem is the web of trust, to be blunt, you don't have a solution. I am techie whose social circle is mostly other techies, and I know precisely zero people who have ever used PGP keys or any other WoT-based system, despite 30 years of evangelism. It's just not a thing anybody wants.

theshrike79 6 days ago | parent [-]

It's 99.99% an UI issue.

If Google wouldn't have let perfect be the enemy of good and had added PGP support to Gmail early on (even just the shittiest signatures that are automatically applied and verified), the world would be a completely different place. Scams just wouldn't exist at this scale when signing mails with a known key would be the standard.

The tech is there, now we have Matrix and XMPP and PubSub and god knows how many protocols to share keys. Even Keybase.io still kind of exists.

What is lacking is a browser ecosystem for people to use their known identities to vouch for a specific url (with smart hashing so that changing the content would invalidate the trust).

We have the technology. Someone(tm) "just" needs to build it :)

gowld 7 days ago | parent | prev [-]

"Web of Trust" has been the proposed answer for, what, 30 years now? But everyone is too lazy to implement and abide by it.

vineyardmike 7 days ago | parent [-]

Don’t worry, it’s coming for real this time. The governments have been proposing a requirement that web companies connect accounts to government IDs.

If that isn’t exciting enough, Sam Altman (yea the one who popularized this LLM slop) will gladly sell you his WorldCoin to store your biometric data on the blockchain!

johnecheck 7 days ago | parent | prev [-]

Indeed. I worry though. We need those defense systems ASAP. The misinformation and garbage engulfing the internet does real damage. We can't just tune it out and wait for it to get better.

epiccoleman 7 days ago | parent | prev | next [-]

I definitely understand the concern - I don't think I'd have hung out on HN for so long if LLM generated postings were common. I definitely recognize this is something you don't want to see happening at scale.

But I still can't help but grin at the thought that the bot knows that the thing to do when you've got a startup is to go put it on HN. It's almost... cute? If you give AI a VPS, of course it will eventually want to post its work on HN.

It's like when you catch your kid listening to Pink Floyd or something, and you have that little moment of triumph - "yes, he's learned something from me!"

7 days ago | parent [-]
[deleted]
sixhobbits 7 days ago | parent | prev | next [-]

(author here) I did feel kinda bad about it as I've always been a 'good' HNer until that point but honestly it didn't feel that spammy to me compared to some human generated slop I see posted here, and as expected it wasn't high quality enough to get any attention so 99% of people would never have seen it.

I think the processes etc that HN have in place to deal with human-generated slop are more than adequate to deal with an influx of AI generated slop, and if something gets through then maybe it means it was good enough and it doesn't matter?

felixgallo 7 days ago | parent | next [-]

That kind of attitude is exactly why we're all about to get overwhelmed by the worst slop any of us could ever have imagined.

The bar is not 'oh well, it's not as bad as some, and I think maybe it's fine.'

taude 7 days ago | parent [-]

well, he was arguing that it's not worse than 99% of the human slop that gets posted, so where do you draw the line?

* well crafted, human only? * Well crafted, whether human or AI? * Poorly crafted, human * well crafted, AI only * Poorly crafted, AI only * Just junk?

etc.

I think people will intuitively get a feel for when content is only AI generated. If people spend time writing a prompt that doesn't make it so wordy, and has personality, and it OK, then fine.

Also, big opportunity going to be out there for AI detected content, whether in forums, coming in inmail inboxes, on your corp file share, etc...

AtlasBarfed 7 days ago | parent | prev [-]

Did you?

Spoiler: no he didn't.

But the article is interesting...

It really highlights to me the pickle we are in with AI: because we are at a historical maximum already of "worse is better" with Javascript, and the last two decades have put out a LOT of javascript, AI will work best with....

Javascript.

Now MAYBE better AI models will be able to equivalently translate Javascript to "better" languages, and MAYBE AI coding will migrate "good" libraries in obscure languages to other "better" languages...

But I don't think so. It's going to be soooo much Javascript slop for the next ten years.

I HOPE that large language models, being language models, will figure out language translation/equivalency and enable porting and movement of good concepts between programming models... but that is clearly not what is being invested in.

What's being invested in is slop generation, because the prototype sells the product.

DrSiemer 7 days ago | parent | prev | next [-]

I'm not a fan of this option, but it seems to me the only way forward for online interaction is very strong identification on any place where you can post anything.

postexitus 7 days ago | parent | next [-]

Back in FidoNet days, some BBSs required identification papers for registering and only allowed real names to be used. Though not known for their level headed discussions, it definitely added a certain level of care in online interactions. I remember the shock seeing the anonymity Internet provided later, both positive and negative. I wouldn't be surprised if we revert to some central authentication mechanism which has some basic level of checks combined with some anonymity guarantees. For example, a government owned ID service, which creates a new user ID per website, so the website doesn't know you, but once they blacklist that one-off ID, you cannot get a new one.

sleepybrett 6 days ago | parent | next [-]

Smaller communities too.

I grew up in... slightly rural america in the 80s-90s, we had probably a couple of dozen local BBSes the community was small enough that after a bit I just knew who everyone was OR could find out very easily.

When the internet came along in the early 90s and I started mudding and hanging out in newsgroups I liked them small where I could get to know most of the userbase, or at least most of the posing userbase. Once mega 'somewhat-anonymous' (i.e. posts tied to a username, not like 4chan madness) communities like slashdot, huge forums, etc started popping up and now with even more mega-communities like twitter and reddit. We lost something, you can now throw bombs without consequence.

I now spend most of my online time in a custom built forum with ~200 people in it that we started building in an invite only way. It's 'internally public' information who invited who. It's much easier to have a civil conversation there, though we still do get the occasional flame-out. Having a stable identity even if it's not tied to a government name is valuable for a thriving and healthy community.

DrSiemer 6 days ago | parent [-]

Sounds good!

A German forum I'm on allows members limited invites based on participation. The catch is, you are responsible for the people you invite. If they get in trouble, you will share a part of the punishment.

benterix 7 days ago | parent | prev | next [-]

Honestly, having seen how it can be used against you, retroactively, I would never ever engage in a discussion under my real name.

(The fact that someone could correlate posts[0] based on writing style, as previously demonstrated on HN and used to doxx some people, makes things even more convoluted - you should think twice what you write and where.)

[0] https://news.ycombinator.com/item?id=33755016

postexitus 6 days ago | parent [-]

This is a subset of "I don't have anything to hide" argument - if we use our real names, I think we'll have more responsibility about what we say. Of course, that's assuming our seemingly democratic governments don't turn authoritarian all of a sudden, as a Turkish citizen, I know that's not a given.

andoando 7 days ago | parent | prev [-]

id.me?

Not government owned, but even irs.gov uses it

xnorswap 7 days ago | parent | prev [-]

That can be automated away too.

People will be more than willing to say, "Claude, impersonate me and act on my behalf".

withinboredom 6 days ago | parent | next [-]

I do this every time I find myself typing something I could get written up over or even fired for.

1. I'm usually too emotional to write out why I feel that way instead of saying what I feel.

2. I really don't like the person (or their idea) but I don't want to get fired over it.

Claude is really great at this: "Other person said X, I think it is stupid and they're a moron for suggesting this. Explain to them why this is a terrible idea or tell me I'm being an idiot."

Sometimes it tells me I'm being an idiot, sometimes it gives me nearly copy-pasta text that I can use and agree with.

pyrale 7 days ago | parent | prev | next [-]

> People will be more than willing to say, "Claude, impersonate me and act on my behalf".

I'm now imagining a future where actual people's identities are blacklisted just like some IP addresses are dead to email, and a market develops for people to sell their identity to spammers.

simonw 7 days ago | parent [-]

That's always been the biggest flaw in the Worldcoin idea in my opinion: if you have a billion+ humans get their eyeball scanned in exchange for some kind of cryptographic identity, you can guarantee that a VERY sizable portion of those billion people will happily sell that cryptographic identity (which they don't understand the value of) to anyone who offers them some money.

As far as I can tell the owner of the original iris can later invalidate an ID that they've sold, but if you buy an ID from someone who isn't strongly technically literate you can probably extract a bunch of value from it anyway.

zoeysmithe 7 days ago | parent | prev | next [-]

I mean, that's fine I guess as long as its respectable and respects the forum.

"Claude write a summary of the word doc I wrote about x and post it as a reply comment," is fine. I dont see why it wouldnt be. Its a good faith effort to post.

"Claude, post every 10 seconds to reddit to spam people to believe my politics is correct," isn't but that's not the case. Its not a good faith effort.

The moderation rules for 'human slop' will apply to AI too. Try spamming a well moderated reddit and see how far you get, human or AI.

antonvs 6 days ago | parent [-]

The problem is speed and quantity. Humans weren't able to fight off the original email spam, it took automated systems. Forums will have to institute much stronger rate limiting and other such measures.

gowld 7 days ago | parent | prev [-]

That's fine, because once someone is banned, the impersonations are also banned.

bookofjoe 7 days ago | parent | prev | next [-]

See also: https://news.ycombinator.com/item?id=44860174 (posted 12 hours ago)

zoeysmithe 7 days ago | parent | prev | next [-]

I mean I can spam HN right now with a script.

Forums like HN, reddit, etc will need to do a better job detecting this stuff, moderator staffing will need to be upped, AI resistant captchas need to be developed, etc.

Spam will always be here in some form, and its always an arms race. That doesnt really change anything. Its always been this way.

kbar13 7 days ago | parent | prev | next [-]

it's annoying but it'll be corrected by proper moderation on these forums

as an aside i've made it clear that just posting AI-written emoji slop PR review descriptions and letting claude code directly commit without self reviewing is unacceptable at work

bongodongobob 7 days ago | parent | prev [-]

The Internet is already 99% shit and always has been. This doesn't change anything.

zanellato19 7 days ago | parent [-]

It's gotten much worse. Before it was shit from people, now it's corporate shit. Corporate shit is so much worse.

Lerc 6 days ago | parent | prev | next [-]

This is the kind of thing people should be doing with AI. Weird and interesting stuff that has a "Let's find out!" Attitude.

Often there's as much to be learned from why it doesn't work.

I see the AI hype to be limited to a few domains.

People choosing to spend lots of money on things speculatively hoping to get a slice of whatever is cooking, even if they don't really know if it's a pie or not.

Forward looking imagining of what would change if these things get massively better.

Hyperbolic media coverage of the above two.

There are companies taking about adding AI for no other reason than they feel like that's what they should be doing, I think that counts as a weak driver of hype, but only because cumulatively, lots of companies are doing it. If anything I would consider this an outcome of hype.

Of these the only one that really affects me is AI being shoehorned into places it shouldn't

The media coverage stokes fires for and against, but I think it only changes the tone of annoyance I have to endure. They would do the same on another topic in the absence of AI. It used to be crypto,

I'm ok with people spending money that is not mine on high risk, high potential reward. It's not for me to judge how they calculate the potential risk or potential reward. It's their opinion, let them have it.

The weird thing I find is the complaints about AI hype dominating. I have read so many pieces where the main thrust of their argument is about the dominance of fringe viewpoints that I very rarely encounter. Frequently they take the stance that anyone imagining how the world might change from any particular form of AI as a claim that that form is inevitable and usually imminent. I don't see people making those claims.

I see people talking about what they tried, what they can do, and what they can't do. Everything they can't do is then held up by others as if it were a trophy and proof of some catestrophic weakness.

Just try stuff, have fun, if that doesn't interest you, go do something else. Tell us about what you are doing. You don't need to tell us that you aren't doing this particular thing, and why. If you find something interesting tell us about that, maybe we will too.

dizlexic 6 days ago | parent | prev | next [-]

every vibe coded thing I've built is trash, but it's amazingly fun to do.

I've tried to explain it to other devs that it's like dumping out a 10000 piece jigsaw puzzle and trying to put it together again.

it's just fun.

mmcconnell1618 6 days ago | parent | next [-]

There was a time when everyone hand-coded HTML. Then came Macromedia Dreamweaver and Microsoft FrontPage which promised a WYSIWYG experience. No one would ever need to "learn HTML and CSS" because the tool could write it for them. Those tools could crank out a website in minutes.

When those tools created some awful, complex and slow output, only the people who knew HTML could understand why it wasn't working and fix things.

Vibe coding is in a similar place. It demos really well. It can be powerful and allows for quick iteration on ideas. It works, most of the time. Vibe coding can produce some really terrible code that is not well architected and difficult to maintain. It can introduce basic logic errors that are not easily corrected through multiple prompts back to the system.

I don't know if they will ever be capable of creating production quality systems on par with what senior engineers produce or if they will only get incrementally better and remain best for prototypes and testing ideas.

cesarvarela 6 days ago | parent | prev | next [-]

It is addicting

throwaway31131 6 days ago | parent | prev [-]

> it’s just fun

For some definitions of fun… :)

indigodaddy 7 days ago | parent | prev | next [-]

Not sure if I'd want Claude doing whatever on a production vps/node, but I like the idea of a way to use Claude Code on the go/wherever you are. I'm going to setup KASM workspaces on my free OCI server and see how it works there.

https://hub.docker.com/r/linuxserver/kasm

prashantsengar 7 days ago | parent [-]

Thanks for sharing this! I have been trying on and off to run RooCode on a VPS to use it on the go. I tried Code Server but it does not share "sessions". KASM seems interesting for this. Do share if you write a blog post on setting it up

indigodaddy 6 days ago | parent [-]

It’s pretty straightforward through the Linuxserver docker image deployment. I have some notes here re: configuration and package persistence strategy via brew:

https://gist.githubusercontent.com/jgbrwn/28645fcf4ac5a4176f...

j45 6 days ago | parent | prev [-]

Maintaining scheduled playing with what's changed/new/different is mandatory with the tools one already uses, let alone any new ones.