| ▲ | htx80nerd 17 hours ago |
| You have to hold AI hand to do even simple vanilla JS correctly. Or do framework code which is well documented all over the net. I love AI and use it for programming a lot, but the limitations are real. |
|
| ▲ | xtracto 9 hours ago | parent | next [-] |
| The other day I (well, the AI) just wrote a Rust app to merge two (huge, GB of data) tables by discovering columns with data in common based on text distance (levenshtein and Dice) . It worked beautifully An i have NEVER made one line of Rust. I dont understand nay-sayers, to me the state of gen.AI is like the simpsons quote "worst day so far". Look were we are within 5 years of the first real GPT/LLM. The next 5 years are going to be crazy exciting. The "programmer" position will become a "builder". When we've got LLMs that generate Opus quality text at 100x speed (think, ASIC based models) , things will get crazy. |
| |
| ▲ | dannersy 3 hours ago | parent | next [-] | | Because if you don't know the language or problem space, there are footguns in there that you can't find, you won't know what to look for to find them. Only until you try to actually use this in a production environment will the issues become evident. At that point, you'll have to either know how to read and diagnose the code, or keep prompting till you fix it, which may introduce another footgun that you didn't know that you didn't know. This is what gets me. The tools can be powerful, but my job has become a thankless effort in pointing out people's ignorance. Time and again, people prompt something in a language or problem space they don't understand, it "works" and then it hits a snag because the AI just muddled over a very important detail, and then we're back to the drawing board because that snag turned out to be an architectural blunder that didn't scale past "it worked in my very controlled, perfect circumstances, test run." It is getting really frustrating seeing this happen on repeat and instead of people realizing they need to get their hands dirty, they just keep prompting more and more slop, making my job more tedious. I am basically at the point where I'm looking for new avenues for work. I say let the industry just run rampant with these tools. I suspect I'll be getting a lot of job offers a few years from now as everything falls apart and their $10k a day prompting fixed one bug to cause multiple regressions elsewhere. I hope you're all keeping your skills sharp for the energy crisis. | | |
| ▲ | psyklic 3 hours ago | parent [-] | | Before LLMs, I've watched in horror as colleagues immediately copy-paste-ran Stack Overflow solutions in terminal, without even reading them. LLM agents are basically the same, except now everyone is doing it. They copy-paste-run lots of code without meaningfully reviewing it. My fear is that some colleagues are getting more skilled at prompting but less skilled at coding and writing. And the prompting skills may not generalize much outside of certain LLMs. |
| |
| ▲ | npinsker 9 hours ago | parent | prev | next [-] | | Human minds are built to find patterns, and you should be careful not to assume the rate of improvement will continue forever based on nothing but a pattern. | | |
| ▲ | throwawaytea 8 hours ago | parent | next [-] | | Just the fact that even retail quality hardware is still improving at local LLM significantly is still a great sign.
If AI quality remained the same, and the cost for local hardware dropped to $1000, it would still be the greatest thing since the internet IMO.
So even if the worst happens and all progress stops, I'm still very happy with what we got. | | |
| ▲ | leptons 7 hours ago | parent [-] | | >I'm still very happy with what we got "One person's slop is another person's treasure" I'm not all that impressed with "AI". I often "race" the AI by giving it a task to do, and then I start coding my own solution in parallel. I often beat the AI, or deliver a better result. Artificial Intelligence is like artificial flavoring. It's cheap and tastes passable to most people, but real flavors are far better in every way even if it costs more. | | |
| ▲ | LadyCailin an hour ago | parent [-] | | At their current stage, this feels like the wrong way to use them. I use them fully supervised, (despite the fact that feels like I’m fighting the tools), which is kind of the best of both worlds. I review every line of code before I allow the edit, and if something is wrong, I tell it to fix it. It learns over time, especially as I set rules in memories, and so the process has sped up, to the point that this goes way faster than if I would have done that myself. Not all tasks are appropriate for LLMs at all, but when they are, this supervised mode is quite fast, and I don’t believe the output to be slop, but anyways I feel like I own every line of code still. |
|
| |
| ▲ | fauchletenerum 8 hours ago | parent | prev [-] | | The overall trend in AI performance will still be up and to the right like everything else in computing over the past 50 years, improvement doesn't have to be linear | | |
| |
| ▲ | fastforwardius 5 hours ago | parent | prev | next [-] | | I seem to remember doing it in SQL (EDIT_DISTANCE) 20ish years ago. While I wouldn't say it worked beautifully, I also didn't need to make a single line of Rust :) also no more than 2 line s of SQL were needed. | | |
| ▲ | jqbd 18 minutes ago | parent [-] | | And how many years of experience you needed to know what to write, and what if you can replace that time with how long prompting takes? |
| |
| ▲ | 8 hours ago | parent | prev | next [-] | | [deleted] | |
| ▲ | sjeiuhvdiidi 8 hours ago | parent | prev [-] | | Let me explain the naysayers, they know "programmer" has always meant "builder" and just because search is better and you can copy and paste faster doesn't mean you've built anything.First thing people need to realize is no proprietary code is in those databases, and using Ai will ultimately just get you regurgitated things people don't really care about. Use it all you want, you won't be able to do anything interesting, they aren't giving you valuable things for free. Anything of value will still take time and knowledge. The marketing hype is to reduce wages and prevent competition. Go for it. |
|
|
| ▲ | sp00chy 17 hours ago | parent | prev | next [-] |
| Exactly that is also my experience also with Claude Code. It can create a lot of stuff impressively but with LOTS of more code than necessary. It’s not really effective in the end. I have more than 35 years of coding experience and always dig into the newest stuff. Quality wise it’s still not more than junior dev stuff even with latest models, sorry. And I know how to talk to these machines. |
| |
| ▲ | TuxSH 17 hours ago | parent | next [-] | | I don't have as many years of professional experience as you do, but IMO code pissing is one of the areas LLMs and "agentic tools" shine the least. In both personal projects and $dayjob tasks, the highest time-saving AI tasks were: - "review this feature branch" (containing hand-written commits) - "trace how this repo and repo located at ~/foobar use {stuff} and how they interact with each other, make a Mermaid diagram" - "reverse engineer the attached 50MiB+ unstripped ELF program, trace all calls to filesystem functions; make a table with filepath, caller function, overview of what caller does" (the table is then copy-pasted to Confluence) - basic YAML CRUD Also while Anthropic has more market share in B2B, their model seems optimized for frontend, design, and literary work rather than rigorous work; I find it to be the opposite with their main competitor. Claude writes code rife with safety issues/vulns all the time, or at least more than other models. | |
| ▲ | iamflimflam1 5 hours ago | parent | prev [-] | | Try the new /simplify command. |
|
|
| ▲ | jcranmer 16 hours ago | parent | prev | next [-] |
| I must say, I do love how this comment has provoked such varying responses. My own observations about using AI to write code is that it changes my position from that of an author to a reviewer. And I find code review to be a much more exhausting task than writing code in the first place, especially when you have to work out how and why the AI-generated code is structured the way it is. |
| |
| ▲ | thegrim33 11 hours ago | parent | next [-] | | There's a very wide range of programming tasks of differing difficulty that people are using / trying to use it for, and a very wide range of intelligence amongst the people that are using / trying to use it, and who are evaluating its results. Hence, different people have very different takes. | |
| ▲ | seanmcdirmid 9 hours ago | parent | prev | next [-] | | > especially when you have to work out how and why the AI-generated code is structured the way it is. You could just ask it? Or you don’t trust the AI to answer you honestly? | | |
| ▲ | chmod775 7 hours ago | parent [-] | | You're anthropomorphizing. LLMs can't lie nor can they tell the truth. These concepts just don't apply to them. They also cannot tell you what they were "thinking" when they wrote a piece of code. If you "ask" them what they were thinking, you just get a plausible response, not the "intention" that may or may not have existed in some abstract form in some layer when the system selected tokens*. That information is gone at that point and the LLM has no means to turn that information into something a human could understand anyways. They simply do not have what in a human might be called metacognition. For now. There's lots of ongoing experimental research in this direction though. Chances are that when you ask an LLM about their output, you'll get the response of either someone who now recognized an issue with their work, or the likeness of someone who believes they did great work and is now defending it. Obviously this is based on the work itself being fed back through the context window, which will inform the response, and thus it may not be entirely useless, but... this is all very far removed from what a conscious being might explain about their thoughts. The closest you can currently get to this is reading the "reasoning" tokens, though even those are just some selected system output that is then fed back to inform later output. There's nothing stopping the system from "reasoning" that it should say A, but then outputting B. Example: https://i.imgur.com/e8PX84Z.png * One might say that the LLM itself always considers every possible token and assigns weights to them, so there wouldn't even be a single chain of thought in the first place. More like... every possible "thought" at the same time at varying intensities. | | |
| ▲ | pipes 6 hours ago | parent [-] | | Or ask another model to tell you what the changes do. | | |
| ▲ | cyclopeanutopia 2 hours ago | parent | next [-] | | And you could first read the thing to which you are replying. Don't tell me it was too long. | |
| ▲ | chmod775 6 hours ago | parent | prev | next [-] | | Did you mean to reply to some other comment? I'm having trouble contextualizing your response - pardon the pun. | |
| ▲ | 6 hours ago | parent | prev [-] | | [deleted] |
|
|
| |
| ▲ | tayo42 8 hours ago | parent | prev [-] | | Your always reviewing code though. Either a team mates pr or maybe your own code in 3 months, or some legacy thing. | | |
| ▲ | christophilus 16 minutes ago | parent [-] | | Human code is still easier to review. Also, I program 80% of the time and review PRs 20% of the time. With AI, that becomes: I review 80% of the time, and write markdown and wait 20% of the time. |
|
|
|
| ▲ | wek 17 hours ago | parent | prev | next [-] |
| This is not my experience either. If you put the work in upfront to plan the feature, write the test cases, and then loop until they pass... you can build a lot of high quality software quickly. The difference between a junior engineer using it and a great architect using it is significant. I think of it as an amplifier. |
| |
| ▲ | andrekandre 5 hours ago | parent | next [-] | | > If you put the work in upfront to plan the feature, write the test cases, and then loop until they pass...
it can be exhausting and time consuming front-loading things so deeply though; sometimes i feel like i would have been faster cutting all that out and doing it myself because in the doing you discover a lot of missing context (in the spec) anyways... | |
| ▲ | bluefirebrand 8 hours ago | parent | prev | next [-] | | This honestly reads to me like "if you spend a lot of time doing tedious monotonous shit you can save a lot of time on the interesting stuff" I have no interest being a "great architect" if architects don't actually build anything | |
| ▲ | Mars008 8 hours ago | parent | prev [-] | | > The difference between a junior engineer using it and a great architect using it is significant Yes, juniors are trying to use AI with the minimum input. This alone tells a lot.. |
|
|
| ▲ | seanmcdirmid 17 hours ago | parent | prev | next [-] |
| Not in my experience. But then again, lots of programmers are limited in how they use AI to write code. Those limitations are definitely real. |
|
| ▲ | keeganpoppen 17 hours ago | parent | prev | next [-] |
| that's just not even remotely my experience. and i am ~20k hours into my programming career. ai makes most things so much faster that it is hard to justify ever doing large classes of things yourself (as much as this hurts my aesthetic sensibilities, it simply is what it is). |
| |
| ▲ | lumost 17 hours ago | parent | next [-] | | Part of this depends on if you care that the AI wrote the code "your way." I've been in shops with rather exotic and specific style guides and standards which the AI would not or will not conform to. | | |
| ▲ | igor47 7 hours ago | parent | next [-] | | Yeah, I also highly value consistency in my projects, which forces me to keep an eye on the LLM and steer it often. This limits my overall velocity especially on larger features. But I'm still much faster with the agent. Recent example, https://github.com/igor47/csheet/pull/68 -- this took me a couple of hours pairing with Claude code, which is insane give the size of the work here. Though this PR creates a bunch of tables, routes, services -- it's not just greenfield CRUD work. We're figuring out how to model a complicated domain, integrating with existing code, thinking through complex integrations including with LLMs at run time. Claude is writing almost all the code, I'm just steering | |
| ▲ | localhost 15 hours ago | parent | prev [-] | | then have ai write a deterministic transformation tool that turns it into the specific style and standard that is needed |
| |
| ▲ | leptons 6 hours ago | parent | prev [-] | | I've never seen a human estimate their "programming career" in kilohours. Is that supposed to look more impressive than years? So, you've been programming only about 7 years? I guess I'm at about "170 kilohours". | | |
| ▲ | ralferoo 3 hours ago | parent | next [-] | | As well as the peer comment about Gladwell (10k hours is considered the point you've mastered a skill), it's also a far more honest metric about how much time you've spent actually programming. Maybe you were writing code, make design choices and debugging 8 hours a day.
Maybe you were primarily doing something else and only writing code for an hour a day.
Who would be the better programmer? The first guy with one year of experience or the second guy with 7 years? I personally would only measure my experience in years, because it's approaching 3 decades full-time in industry (plus an additional decade of cutting my teeth during school and university), but I can certainly see that earlier on in a career it's a useful metric in comparison to the 10,000 hours. | |
| ▲ | kennywinker 5 hours ago | parent | prev [-] | | I think it’s probabky because of the malcom gladwell “ten thousand hours” idea. |
|
|
|
| ▲ | moezd 17 hours ago | parent | prev | next [-] |
| AI assisted code can't even stick to the API documentation, especially if the data structures are not consistent and have evolved over time. You would see Claude literally pulling function after function from thin air, desperately trying to fulfill your complicated business logic and even when it's complete, it doesn't look neat at all. Yes, it will have test coverage, but one more feature request will probably break the back of the camel. And if you raise that PR to the rest of your team, good luck trying to summarise it all to your colleagues. However if you just have an easy project, or a greenfield project, or don't care about who's going to maintain that stuff in 6 months, sure, go all in with AI. |
| |
| ▲ | ccosky 12 hours ago | parent | next [-] | | I definitely wonder if the people going all-in on AI harnessing are working on greenfield projects, because it seems overwhelming to try to get that set up on a brownfield codebase where the patterns aren't consistent and the code quality is mixed. | |
| ▲ | tayo42 8 hours ago | parent | prev [-] | | So just iterate on it? Your complaint is that the model isn't one shotting the problem and reading your mind about style. It's like any coding workflow, make it work, then make it nice. | | |
| ▲ | moezd 7 hours ago | parent [-] | | No, I never expect AI to one-shot (if I see such a miracle, it's usually because I needed a one-liner or something really simple and well documented, which I can also write on the whiteboard from memory). Try iterating over well known APIs where the response payloads are already gigantic JSONs, there are multiple ways to get certain data and they are all inconsistent and Claude spits out function after function, laying waste to your codebase. I found no amount of style guideline documents to resolve this issue. I'd rather read the documentation myself and write the code by hand rather than reviewing for the umpteenth time when Claude splits these new functions between e.g. __init__.py and main.py and god knows where, mixing business logic with plumbing and transport layers as an art form. God it was atrocious during the first few months of FastMCP. |
|
|
|
| ▲ | 17 hours ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | GalaxyNova 17 hours ago | parent | prev | next [-] |
| Not what I've experienced |
|
| ▲ | neversupervised 8 hours ago | parent | prev | next [-] |
| It’s crazy how some people feel the ai and others don’t. But one group is wrong. It’s a matter of time before everyone feels the AI. |
|
| ▲ | fudfomo 15 hours ago | parent | prev [-] |
| Most of this thread is debating whether models are good or bad at writing code... however, I think a more important question is what we feed the AI with because that dramatically determines the quality of the output. When your agent explores your codebase trying to understand what to build, it read schema files, existing routes, UI components etc... easily 50-100k tokens of implementation detail. It's basically reverse-engineering intent from code. With that level of ambiguous input, no wonder the results feel like junior work. When you hand it a structured spec instead including data model, API contracts, architecture constraints etc., the agent gets 3-5x less context at much higher signal density. Instead of guessing from what was built it knows exactly what to build. Code quality improves significantly. I've measured this across ~47 features in a production codebase with amedian ratio: 4x less context with specs vs. random agent code exploration. For UI-heavy features it's 8-25x. The agent reads 2-3 focused markdown files instead of grepping through hundreds of KB of components. To pick up @wek's point about planning from above: devs who get great results from agentic development aren't better prompt engineers... they're better architects. They write the spec before the code, which is what good engineering always was... AI just made the payoff for that discipline 10x more visible. |