| ▲ | gordonhart 13 hours ago |
| Whenever I get worried about this I comb through our ticket tracker and see that ~0% of them can be implemented by AI as it exists today. Once somebody cracks the memory problem and ships an agent that progressively understands the business and the codebase, then I'll start worrying. But context limitation is fundamental to the technology in its current form and the value of SWEs is to turn the bigger picture into a functioning product. |
|
| ▲ | nemo1618 10 hours ago | parent | next [-] |
| "The steamroller is still many inches away. I'll make a plan once it actually starts crushing my toes." You are in danger. Unless you estimate the odds of a breakthrough at <5%, or you already have enough money to retire, or you expect that AI will usher in enough prosperity that your job will be irrelevant, it is straight-up irresponsible to forgo making a contingency plan. |
| |
| ▲ | overgard 3 hours ago | parent | next [-] | | What contingency plan is there exactly? At best you're just going from an automated-already job to a soon-to-be-automated job. Yay? I'm baffled that so many people think that only developers are going to be hit and that we especially deserve it. If AI gets so good that you don't need people to understand code anymore, I don't know why you'd need a project manager anymore either, or a CFO, or a graphic designer, etc etc. Even the people that seem to think they're irreplaceable because they have some soft power probably aren't. Like, do VC funds really need humans making decisions in that context..? Anyway, the practical reason why I'm not screaming in terror right now is because I think the hype machine is entirely off the rails and these things can't be trusted with real jobs. And honestly, I'm starting to wonder how much of tech and social media is just being spammed by bots and sock puppets at this point, because otherwise I don't understand why people are so excited about this hypothetical future. Yay, bots are going to do your job for you while a small handful of business owners profit. And I guess you can use moltbot to manage your not-particularly-busy life of unemployment. Well, until you stop being able to afford the frontier models anyway, which is probably going to dash your dream of vibe coding a startup. Maybe there's a handful of winners, until there's not, because nobody can afford to buy services on a wage of zero dollars. And anyone claiming that the abundance will go to everyone needs to get their head checked. | |
| ▲ | Gigachad 10 hours ago | parent | prev | next [-] | | My contingency plan is that if AI leaves me unable to get a job, we are all fucked and society as a whole will have to fix the situation and if it doesn’t, there is nothing I could have done about it anyway. | | |
| ▲ | chadcmulligan 8 hours ago | parent | next [-] | | As a fellow chad I concur. Though I am improving my poker skills - games of chance will still be around | | |
| ▲ | selylindi 4 hours ago | parent [-] | | You likely already know, but the "Pluribus" poker bot was beating humans back in 2019. Games of chance will be around if people are around, but you'll have to be careful to ensure you're playing against people, unassisted people. https://en.wikipedia.org/wiki/Pluribus_(poker_bot) | | |
| ▲ | chadcmulligan 3 hours ago | parent [-] | | Yeah, thanks, I only play live games. I'm in australia so online poker is illegal here. I was thinking of getting a vpn and having a play online, then I saw this recently https://www.reddit.com/r/Damnthatsinteresting/comments/1qi69... | | |
| ▲ | Gigachad 3 hours ago | parent [-] | | So much of these degenerate online gambling / "investment" platforms are illegal here for good reason. If you are just a normal person playing fairly, you are being scammed. Same for things like Polymarket, the only winners are the people with insider knowledge. | | |
| ▲ | chadcmulligan 3 hours ago | parent [-] | | Even horse racing, it's a solved problem, and if you start winning they'll just cancel your a/c (happened to a friend of mine) |
|
|
|
| |
| ▲ | tired-turtle 9 hours ago | parent | prev | next [-] | | This is a sensible plan, given your username. | | |
| ▲ | nikkwong 7 hours ago | parent [-] | | Yeah seriously. Don't people understand the fact that society is not good at mopping up messes like this—there has been a K shaped economy for several decades now and most Americans have something like $400 in their bank accounts. The bottom had already fallen out for them, and help still hasn't arrived. I think it's more likely that what really happens is that white collar workers, especially the ones on the margin, join this pool—and there is a lot of suffering for a long time. Personally, rather devolving into nihilism, I'd rather try to hedge against suffering that fate. Now is the time to invest and save money. (or yesterday) | | |
| ▲ | sarchertech 6 hours ago | parent [-] | | If white collar workers as a whole suffer severe economic setback over a short term timespan, your savings and investments won’t help you. Unless you’re investing in guns, ammo, food, and a bunker. We’re talking worse unemployment than depression era Germany. And structurally more significant unemployment because the people losing their jobs were formally very high earners. | | |
| ▲ | nikkwong 4 hours ago | parent [-] | | That’s the cataclysmic outcome, though. Although I deemed that that’s certainly possible and I would put a double digit percentage probability on it, another very likely outcome is a very severe recession, or a recession, wear a lot of, but not all, white collar work is wiped out. Maybe there’s a significant restructuring in the economy I think in a scenario like that, which also seems to be in the realm of possibility, I think having resources still matters. Speech to text, sorry for the poor grammar. | | |
| ▲ | sarchertech 3 hours ago | parent [-] | | It’s definitely possible that there’s an impact that is bad but not cataclysmic. I figure in thst case though my regular savings is enough to switch to something else. I could retire now if I was willing to move somewhere cheap and live on $60k a year. There’s a lot of things that could cause that level of recession though without the need for AI. I do also think the mid level bad outcome isn’t super likely because of AI is good enough to replace a lot of white collar jobs, I think it could replace almost all of them. |
|
|
|
| |
| ▲ | everettde 7 hours ago | parent | prev [-] | | this has been me ever since my philosophy undergrad. |
| |
| ▲ | nitwit005 9 hours ago | parent | prev | next [-] | | > You are in danger. Unless you estimate the odds of a breakthrough at <5% It's not the odds of the breakthrough, but the timeline. A factory worker could have correctly seen that one day automation would replace him, and yet worked his entire career in that role. There have been a ton of predictions about software engineers, radiologists, and some other roles getting replaced in months. Those predictions have clearly been not so great. At this point the greater risk to my career seems to be the economy tanking, as that seems to be happening and ongoing. Unfortunately, switching careers can't save you from that. | | |
| ▲ | energy123 29 minutes ago | parent | next [-] | | We are the French artisans being replaced by English factories. OpenAI and its employees are the factory. | |
| ▲ | 6 hours ago | parent | prev [-] | | [deleted] |
| |
| ▲ | adamkittelson 9 hours ago | parent | prev | next [-] | | I'm not worried about the danger of losing my job to an AI capable of performing it. I'm worried about the danger of losing my job because an executive wanted to be able to claim that AI has enhanced productivity to such a degree that they were able to eliminate redundancies with no regard for whether there was any truth to that statement or not. | |
| ▲ | jopsen 7 hours ago | parent | prev | next [-] | | > it is straight-up irresponsible to forgo making a contingency plan. What contingencies can you really make? Start training a physical trade, maybe. If this the end of SWE jobs, you better ride the wave. Odds are you're estimate on when AI takes over are off by half a career, anyways. | | |
| ▲ | sarchertech 6 hours ago | parent [-] | | Working in the trades won’t help you at 40-50% unemployment. Who’s going to pay for your services. And even the meager work remains would be fought over by the hundred million unemployed who are all suddenly fighting tooth and nail for any work they can get. |
| |
| ▲ | zozbot234 9 hours ago | parent | prev | next [-] | | So AI is going to steamroll all feasible jobs, all at once, with no alternatives developing over time? That's just a fantasy. | | |
| ▲ | hirvi74 9 hours ago | parent [-] | | It'd probably be cold day in Hell before AI replaces veterinary services, for example. Perhaps for mild conditions, but I cannot imagine an AI robot trying to restrain an animal. | | |
| ▲ | laichzeit0 2 hours ago | parent | next [-] | | All these so-called safe jobs still depend on someone being able to afford those services. If I don't have a job, I can't go see the vet, the fact that no one else can do the vets job is irrelevant at such a point. I would like to know if there's some kind of inflection point, like the so-called Laffer curve for taxes, where once an economy has X% unemployment, it effectively collapses. I'd imagine it goes: recession -> depression -> systemic crisis and appears to be somewhere between 30-40% unemployment based on history. | |
| ▲ | ares623 7 hours ago | parent | prev [-] | | Every job deemed "safe" will be flooded by desparate applicants from unsafe jobs. |
|
| |
| ▲ | themafia 9 hours ago | parent | prev [-] | | > Unless you estimate the odds of a breakthrough at <5% I do. Show me any evidence that it is imminent. > or you expect that AI will usher in enough prosperity that your job will be irrelevant Not in my lifetime. > it is straight-up irresponsible to forgo making a contingency plan. No, I'm actually measuring the risk, you're acting as if the sky is falling. What's your contingency plan? Buy a subscription to the revolution? | | |
| ▲ | adriand 9 hours ago | parent [-] | | > What's your contingency plan? Buy a subscription to the revolution? I’ve been working on my contingency plan for a year-and-a-half now. I won’t get into what it is (nothing earth shattering) but if you haven’t been preparing, I think you’re either not paying enough attention or you’re seriously misreading where this is all going. | | |
| ▲ | small_model 9 hours ago | parent [-] | | This ^ been a SWE for 20 years the market is the worst I have seen it, many good devs been looking for 1-2 years and not even getting a response, whereas 3-4 years ago they would have had multiple offers. Im still working but am secure in terms of money so will be ok not working (financially at least). But I expect a tsunami of layoffs this and next year, then you are competing with 1000x other devs and Indians who will works for 30% of your salary. | | |
| ▲ | realusername 8 hours ago | parent | next [-] | | That's called an economic crisis, it has nothing to do with AI, my friends also have trouble to find 100% manual jobs which were easily available 2 years ago. Yes I said the word that none of these company want to say in their press conference. | | |
| ▲ | small_model 7 hours ago | parent [-] | | Thats because there are more tech/service workers competing for the manual jobs now. | | |
| ▲ | realusername 7 hours ago | parent [-] | | Tech workers aren't numerous enough to have that effect. Besides that, why aren't we seeing any metrics change on Github? With a supposedly increase of productivity so large a good chunk of the workforce is fired, we would see it somewhere. |
|
| |
| ▲ | Seattle3503 7 hours ago | parent | prev [-] | | A lot of non-AI things have happened though. |
|
|
|
|
|
| ▲ | rockbruno 13 hours ago | parent | prev | next [-] |
| While true, my personal fear is that the higher-ups will overlook this fact and just assume that AI can do everything because of some cherry-pick simple examples, leading to one of those situations where a bunch of people get fired for no reason and then re-hired again after some time. |
| |
| ▲ | palmotea 12 hours ago | parent | next [-] | | > leading to one of those situations where a bunch of people get fired for no reason and then re-hired again after some time. More likely they get fired for no reason, never rehired, and the people left get burned out trying to hold it all together. | | |
| ▲ | easymodex 9 hours ago | parent [-] | | Exactly, now which one do you wanna be? The burned out ones but still working in SWE or the fired ones which in the long run converge to manual labor which AI can't do.
Not to mention in SWE case the salaries would be pushed down to match cost of AI doing it. |
| |
| ▲ | themafia 9 hours ago | parent | prev [-] | | As if "higher-ups" is an assigned position. If you fail as a "higher up" you're no longer higher up. Then someone else can take your place. To the extent this does not naturally happen is evidence of petty or major corruptions within the system. | | |
| ▲ | Seattle3503 7 hours ago | parent [-] | | In competitive industries, bad firms will fail. Some industries are not competitive though. I have a friend that went a little crazy working as a PM at a large health insurance firm. |
|
|
|
| ▲ | sensanaty 9 hours ago | parent | prev | next [-] |
| I look through the backlog for my team consisting of 9 trillion ill-defined (if defined at all) tickets that tells you basically nothing. The large, overwhelming majority of my team's time is spent on combing through these tickets and making sense of them. Once we know what the ticket is even trying to say, we're usually out with the solution in a few days at most, so implementation isn't the bottleneck, nowhere near. This scenario has been the same everywhere I've ever worked, at large, old institutions as well as fresh startups. The day I'll start worrying is when the AI is capable of following the web of people involved to translate what the vaguely phrased ticket that's been backlogged for God knows how long actually means |
|
| ▲ | e_i_pi_2 13 hours ago | parent | prev | next [-] |
| A lot of this can be provided or built up by better documentation in the codebase, or functional requirements that can also be created, reviewed, and then used for additional context. In our current codebase it's definitely an issue to get an AI "onboarded", but I've seen a lot less hand-holding needed in projects where you have the AI building from the beginning and leaving notes for itself to read later |
| |
| ▲ | gordonhart 13 hours ago | parent | next [-] | | Curious to hear if you've seen this work with 100k+ LoC codebases (i.e. what you could expect at a job). I've had some good experiences with high autonomy agents in smaller codebases and simpler systems but the coherency starts to fizzle out when the system gets complicated enough that thinking it through is the hard part as opposed to hammering out the code. | | |
| ▲ | sensanaty 9 hours ago | parent | next [-] | | I'd estimate we're near a million LoC (will double check tomorrow, but wouldn't be surprised if it was over that to be honest). Huge monorepo, ~1500 engineers, all sorts of bespoke/custom tooling integrated, fullstack (including embedded code), a mix of languages (predominantly Java & JS/TS though). In my case the AI is actively detrimental unless I hand hold it with every single file it should look into, lest it dive into weird ancient parts of the codebase that bear no relevance to the task at hand. Letting the latest and "greatest" agents loose is just a recipe for frustration and disaster despite lots of smart people trying their hardest to make these infernal tools be of any use at all. The best I've gotten out of it was some light Vue refactoring, but even then despite AGENTS.md, RULES.md and all the other voodoo people say you should do it's a crapshoot. | | |
| ▲ | zozbot234 9 hours ago | parent [-] | | Ask the AI to figure out your code base (or self-contained portions of it, as applicable) and document its findings. Then correct and repeat. Over time, you end up with a scaffold in the form of internal documentation that will guide both humans and AIs in making more productive edits. |
| |
| ▲ | wenc 11 hours ago | parent | prev | next [-] | | If you vector index your code base, agents can explore it without loading it into context. This is what Cursor and Roo and Kiro and probably others do. Claude Code uses string searches. What helps is also getting it to generate a docs of your code so that it has map. This is actually how humans understand a large code base too. We don’t hold a large code base in memory — we navigate it through docs and sampling bits of code. | |
| ▲ | servercobra 10 hours ago | parent | prev | next [-] | | cloc says ours is ~350k LoC and agents are able to implement whole features from well designed requirement docs. But we've been investing in making our code more AI friendly, and things like Devin creating and using DeepWiki helps a lot too. | | |
| ▲ | sarchertech 8 hours ago | parent [-] | | If you have agents that can implement entire features, why is it only 350k loc? Each engineer should be cranking out at least 1 feature a week. If each feature is 1500-2000 lines times 10 engineers that’s 20k lines a week. If the answer is that the AI cranks out code faster than the team can digest and review it and faster than you can spec out the features, what’s the point? I can see completely shifting your workflow, letting skills atrophy, adopting new dependencies, and paying new vendors if it’s boosting your final output 5 or 10x. But if it’s a 20% speed up is it worth it? | | |
| ▲ | wtetzner 5 hours ago | parent [-] | | Since when do we measure productivity by lines of code? | | |
| ▲ | sarchertech 3 hours ago | parent [-] | | It’s not a measure of productivity, but some number of new lines is generally necessary for new functionality. And in my experience AI tends to produce more lines of code than a decent human for similar functionality. So I’d be very shocked if an agent completing a feature didn’t crank out 1500 lines or more. |
|
|
| |
| ▲ | enraged_camel 10 hours ago | parent | prev | next [-] | | Around 250k here. The AI does an excellent job finding its way around, fixing complex bugs (and doing it correctly), doing intensive refactors and implementing new features using existing patterns. | |
| ▲ | christkv 11 hours ago | parent | prev [-] | | Our codebase is well over 250k and we have a hierarchy of notes for the modules so we read as much as we need for the job with a base memory that explains how the notes work |
| |
| ▲ | tharkun__ 13 hours ago | parent | prev [-] | | We have this in some of our projects too but I always wonder how long it's going to take until it just fails. Nobody reads all those memory files for accuracy. And knowing what kind of BS the AI spews regularly in day to day use I bet this simply doesn't scale. |
|
|
| ▲ | UncleOxidant 8 hours ago | parent | prev | next [-] |
| The memory problem is already being addressed in various ways - antigravity seems to keep a series of status/progress files describing what's been done, what needs doing, etc. A bit clunky, but it seems to work - I can open it up on a repo that I was working in a few days back and it seems to pick up this context such that I don't have to completely bring it up to speed every time like I used to have to do. I've heard that claude code has similar mechanisms. I've been doing stuff with recent models (gemini 3, claude 4.5/6, even smaller, open models like GLM5 and Qwen3-coder-next) that was just unthinkable a few months back. Compiler stuff, including implementing optimizations, generating code to target a new, custom processor, etc. I can ask for a significant new optimization feature in our compiler before going to lunch and come back to find it implemented and tested. This is a compiler that targets a custom processor so there is also verilog code involved. We're having the AI make improvements on both the hardware and software sides - this is deep-in-the-weeds complex stuff and AI is starting to handle it with ease. There are getting to be fewer and fewer things in the ticket tracker that AI can't implement. A few months ago I would've completely agreed with you, but the game is changing very rapidly now. |
| |
| ▲ | taysco 8 hours ago | parent [-] | | this works fine for like 2-3 small instruction sets. once you start getting to scale of a real enterprise system, the AI falls down and can't handle that amount of context. It will start ignoring critical pieces or not remember them. And without constant review AI will start priotizing things that are not your business priority. I don't agree they have solved this problem, at all, or really in any way that's actually usable. | | |
| ▲ | UncleOxidant 7 hours ago | parent [-] | | What I'm saying is, don't get to thinking that the memory problem is some kind of insurmountable, permanent barrier that's going to keep us safe. It's already being addressed, maybe crudely at first, but the situation is already much better than it was - I no longer have to bring the model up to speed completely every time I start a new session. Part of this is much larger context windows (1M tokens now). New architectures are also being proposed to deal with the issue, as well. |
|
|
|
| ▲ | matt_heimer 13 hours ago | parent | prev | next [-] |
| It's not binary. Jobs will be lost because management will expect the fewer developers to accomplish more by leveraging AI. |
| |
| ▲ | louiereederson 12 hours ago | parent | next [-] | | Big tech might ahead of the rest of the economy in this experiment. Microsoft grew headcount by ~3% from June 2022 to June 2025 while revenue grew by >40%. This is admittedly weak anecdata but my subjective experience is their products seem to be crumbling (GitHub problems around the Azure migration for instance), and worse than they even were before. We'll see how they handle hiring over the next few years and if that reveals anything. | | |
| ▲ | JetSpiegel 11 hours ago | parent [-] | | Well, Google just raised prices by 30% on the GSuite "due to AI value delivered", but you can't even opt out, so even revenue is a bullshit metric. |
| |
| ▲ | datsci_est_2015 10 hours ago | parent | prev [-] | | Already built in. We haven’t hired recently and our developers are engaged in a Cold War to set the new standard of productivity. |
|
|
| ▲ | deet 10 hours ago | parent | prev | next [-] |
| Just keep in mind that there are many highly motivated people directly working on this problem. It's hard to predict how quickly it will be solved and by whom first, but this appears to be a software engineering problem solvable through effort and resources and time, not a fundamental physical law that must be circumvented like a physical sciences problem. Betting it won't be solved enough to have an impact on the work of today relatively quickly is betting against substantial resources and investment. |
| |
| ▲ | slopinthebag 10 hours ago | parent | next [-] | | Why do you think it's not a physical sciences problem? It could be the case that current technologies simply cannot scale due to fundamental physical issues. It could even be a fundamental rule of intelligent life, that one cannot create intelligence that surpasses its own. Plenty of things get substantial resources and investment and go nowhere. Of course I could be totally wrong and it's solved in the next couple years, it's almost impossible to make these predictions either way. But I get the feeling people are underestimating what it takes to be truly intelligent, especially when efficiency is important. | | |
| ▲ | jatari 10 hours ago | parent [-] | | >It could even be a fundamental rule of intelligent life, that one cannot create intelligence that surpasses its own. Well that is easily disproved by the fact that people have children with higher IQ's than their own. | | |
| ▲ | slopinthebag 10 hours ago | parent [-] | | That's not what I mean, rather than humans cannot create a type of intelligence that supersedes what is roughly capable from human intelligence, because doing so would require us to be smarter basically. Not to say we can't create machines that far surpass our abilities on a single or small set of axis. | | |
| ▲ | mitthrowaway2 8 hours ago | parent | next [-] | | Think hard about this. Does that seem to you like it's likely to be a physical law? First of all, it's not necessary for one person to build that super-intelligence all by themselves, or to understand it fully. It can be developed by a team, each of whom understands only a small part of the whole. Secondly, it doesn't necessarily even require anybody to understand it. The way AI models are built today is by pressing "go" on a giant optimizer. We understand the inputs (data) and the optimizer machine (very expensive linear algebra) and the connective structure of the solution (transformer) but nobody fully understands the loss-minimizing solution that emerges from this process. We study these solutions empirically and are surprised by how they succeed and fail. We may find we can keep improving the optimization machine, and tweaking the architecture, and eventually hit something with the capacity to grow beyond our own intelligence, and it's not a requirement that anyone understands how the resulting model works. We also have many instances in nature and history of processes that follow this pattern, where one might expect to find a similar "law". Mammals can give birth to children that grow bigger than their parents. We can make metals puter than the crucible we melted them in. We can make machines more precise than the machines that made those parts. Evolution itself created human intelligence from the repeated application of very simple rules. | | |
| ▲ | slopinthebag 6 hours ago | parent [-] | | > Think hard about this. Does that seem to you like it's likely to be a physical law? Yes, it seems likely to me. It seems like the ultimate in hubris to assume we are capable of creating something we are not capable of ourselves. | | |
| ▲ | selylindi 4 hours ago | parent [-] | | On the contrary, nearly every machine we've created is capable of things that we are not capable of ourselves. Cars travel more than twice as fast as the swiftest human. Airplanes fly. Calculators do math in an instant that would take a human months. Lightbulbs emit light. Cranes lift many tons. And so on and so forth. So to create something that exceeds our capabilities is not a matter of hubris (as if physical laws cared about hubris anyway), it's an unambiguously ordinary occurrence. | | |
| ▲ | slopinthebag 3 hours ago | parent [-] | | > Not to say we can't create machines that far surpass our abilities on a single or small set of axis. |
|
|
| |
| ▲ | small_model 9 hours ago | parent | prev | next [-] | | Given SOTA models are Phd level in just about every subject this is clearly provably wrong. | | |
| ▲ | zozbot234 8 hours ago | parent | next [-] | | I'll believe that claim when a SOTA model can autonomously create content that matches the quality and length of any average PhD dissertation. As of right now, we're nowhere near that and don't know how we could possibly get there. SOTA models are superhuman in a narrow sense, in that they have solid background knowledge of pretty much any subject they've been trained on. That's great. But no, it doesn't turn your AI datacenter into "a country of geniuses". | |
| ▲ | slopinthebag 6 hours ago | parent | prev [-] | | Are humans just Phd students in a vat? Can a SOTA model walk? Humans in general find that task, along with a trillion other tasks that SOTA models cannot do, to be absolutely trivial. |
| |
| ▲ | ordersofmag 8 hours ago | parent | prev [-] | | Seems like if evolution managed to create intelligence from slime I wouldn't bet on there being some fundamental limit that prevents us from making something smarter than us. |
|
|
| |
| ▲ | ThrowawayR2 6 hours ago | parent | prev | next [-] | | Many highly motivated people with substantial resources and investment have worked on a lot of things and then failed at them with nothing to show for it. | |
| ▲ | datsci_est_2015 10 hours ago | parent | prev [-] | | The implication of your assertion is pretty much a digital singularity. You’re implying that there will be no need for humans to interact with the digital world at all, because any work in the digital world will be achievable by AI. Wonder what that means for meatspace. Edit: Would also disagree this isn’t a physics problem. Pretty sure power required scales according to problem complexity. At a certain level of problem complexity we’re pretty much required to put enough carbon in the atmosphere to cook everyone to a crisp. Edit 2: illustrative example, an Epic in Jira: “Design fusion reactor” |
|
|
| ▲ | audience_mem 38 minutes ago | parent | prev | next [-] |
| 0%? This is as wrong as people who say it can do 100% of tasks. |
|
| ▲ | krackers 5 hours ago | parent | prev | next [-] |
| >progressively understands the business This is no different than onboarding a new member of the team, and I think openAI was working on that "frontier" >We started by looking at how enterprises already scale people. They create onboarding processes. They teach institutional knowledge and internal language. They allow learning through experience and improve performance through feedback. They grant access to the right systems and set boundaries. AI coworkers need the same things. And tribal knowledge will not be a moat once execs realize that all they need to do is prioritize documentation instead of "code velocity" as a metric (sure any metric gets goodhearted, but LLMs are great at sifting through garbage to find the high perplexity tokens). >But context limitation is fundamental to the technology in its current form This may not be the case, large enough context-windows plus external scratchpads would mostly obviate the need for true in context learning. The main issue today is that "agent harnesses" suck. The fact that claude code is considered good is more an indication of how bad everything else is. Tool traces read like a drunken newb brute-forcing his way through tasks. LLMs can mostly "one-shot" individual functions, but orchestrating everything is the blocker. (Yes there's progress in metr or whatever but I don't trust any of that, else we'd actually see the results in real-world open source projects). LLMs don't really know how to interact with subagents. They're generally sort of myopic even with tool calls. They'll spend 20 minutes trying to fix build issues going down a rabbit hole without stepping back to think. I think some sort of self-play might end up solving all of these things, they need to develop a "theory of mind" in the same way that humans do, to understand how to delegate and interact with the subagents they spawn. (Today a failure case is agents often don't realize subagents don't share the same context.) Some of this is certainly in the base model and pretraining, but it needs to be brought out in the same way RL was needed for tool use. |
|
| ▲ | malyk 13 hours ago | parent | prev | next [-] |
| Can you give an example to help us understand? I look at my ticket tracker and I see basically 100% of it that can be done by AI. Some with assistance because business logic is more complex/not well factored than it should be, but most of the work that is done AI is perfectly capable of doing with a well defined prompt. |
| |
| ▲ | gordonhart 13 hours ago | parent | next [-] | | Here's an example ticket that I'll probably work on next week: Live stream validation results as they come in
The body doesn't give much other than the high-level motivation from the person who filed the ticket. In order to implement this, you need to have a lot of context, some of which can be discovered by grepping through the code base and some of which can't:- What is the validation system and how does it work today? - What sort of UX do we want? What are the specific deficiencies in the current UX that we're trying to fix? - What prior art exists on the backend and frontend, and how much of that can/should be reused? - Are there any scaling or load considerations that need to be accounted for? I'll probably implement this as 2-3 PRs in a chain touching different parts of the codebase. GPT via Codex will write 80% of the code, and I'll cover the last 20% of polish. Throughout the process I'll prompt it in the right direction when it runs up against questions it can't answer, and check its assumptions about the right way to push this out. I'll make sure that the tests cover what we need them to and that the resultant UX feels good. I'll own the responsibility for covering load considerations and be on the line if anything falls over. Does it look like software engineering from 3 years ago? Absolutely not. But it's software engineering all the same even if I'm not writing most of the code anymore. | | |
| ▲ | Rodeoclash 12 hours ago | parent | next [-] | | This right here is my view on the future as well. Will the AI write the entire feature in one go? No. Will the AI be involved in writing a large proportion of the code that will be carefully studied and adjusted by a human before being used? Absolutely yes. This cyborg process is exactly how we're using AI in our organisation as well. The human in the loop understands the full context of what the feature is and what we're trying to achieve. | |
| ▲ | codegangsta 12 hours ago | parent | prev | next [-] | | But planning like this is absolutely something AI can do. In fact, this is exactly the kind of thing we start with on our team when it comes to using AI agents. We have a ticket with just a simple title that somebody threw in there, and we asked the AI to spin up a bunch of research agents to understand and plan and ask itself those questions. Funny enough, all the questions that you posed are things that come up right away that the agent asks itself, and then goes and tries to understand and validate an answer, sometimes with input from the user. But I think this planning mechanism is really critical to being able to have an AI generate an understanding, then have it be validated by a human before beginning implementation. And by planning I don't necessarily mean plan mode in your agent harness of choice. We use a custom /plan skill in Claude Code that orchestrates all of this using multiple agents, validation loops, and specific prompts to weed out ambiguities by asking clarifying questions using the ask user question tool. This results in taking really fuzzy requirements and making them clear, and we automate all of this through linear but you could use your ticket tracker of choice. | | |
| ▲ | adriand 9 hours ago | parent [-] | | Absolutely. Eventually the AI will just talk to the CEO / the board to get general direction, and everything will just fall out of that. The level of abstraction the agents can handle is on a steady upward trajectory. | | |
| ▲ | sarchertech 6 hours ago | parent [-] | | If AIs can do that, they won’t be talking to a CEO or the board of a software company. There won’t be a CEO or a board because software companies won’t exist. They’ll talk to the customers and build one off solutions for each of them. There will be 3 “software” companies left. And shortly after that society will collapse because of AI can do that it can do any white collar job. |
|
| |
| ▲ | fragmede 12 hours ago | parent | prev [-] | | I mean, what is the validation system? Either it exists in code, and thus can be discovered if you point the AI at repo, or... what, it doesn't exist? For the UX, have it explore your existing repos and copy prior art from there and industry standards to come up with something workable. Web scale issues can be inferred by the rest of the codebase. If your terraform repo has one RDS server, vs a fleet of them, multi-region, then the AI, just as well as a human, can figure out if it needs Google Spanner level engineering or not. (probably not) Bigger picture though, what's the process of a human logs an under specified ticket and someone else picks it up and has no clue what to do with it? They're gonna go ask the person who logged the bug for their thoughts and some details beyond "hurr Durr something something validation". If we're at the point where AI is able to make a public blog post shaming the open source developer for not accepting a patch, throwing questions back to you in JIRA about the details of the streaming validation system is well within its capabilities, given the right set of tools. | | |
| ▲ | gordonhart 12 hours ago | parent [-] | | Honestly curious, have you seen agents succeed at this sort of long-trajectory wide breadth task, or is it theoretical? Because I haven't seen them come close (and not for lack of trying) | | |
| ▲ | codegangsta 12 hours ago | parent | next [-] | | Yeah I absolutely see it every day. I think it’s useful to separate the research/planning phase from the building/validadation/review phase. Ticket trackers are perfect for this. Just start with asking AI to take this unclear, ambiguous ticket and come up with a real plan for how to accomplish it. Review the plan, update your ticket system with the plan, have coworkers review it if you want. Then when ready, kick off a session for that first phase, first PR, or the whole thing if you want. | |
| ▲ | kolinko 10 hours ago | parent | prev | next [-] | | In my expedience, Claude Code with opus 4.5 is the first one to tackle such issues well. | |
| ▲ | fragmede 10 hours ago | parent | prev [-] | | Opus 4.6, with all of the random tweaks I've picked up off of here, and twitter, is in the middle of rewriting my golang cli program for programmers into a swiftui Mac app that people can use, and it's totally managing to do it. Claude swarm mode with beads is OP. |
|
|
| |
| ▲ | lbrito 13 hours ago | parent | prev | next [-] | | Then why isn't it? Just offload it to the clankers and go enjoy a margarita at the beach or something. | | |
| ▲ | Gud 12 hours ago | parent [-] | | There are plenty of people who are enjoying margarita by the beach while you, the laborer, are working for them. | | |
| ▲ | lbrito 12 hours ago | parent [-] | | Preach. That's always been the case though, AI just makes it slightly worse. |
|
| |
| ▲ | contagiousflow 13 hours ago | parent | prev | next [-] | | Why do you have a backlog then? If a current AI can do 100% of it then just run it over the weekend and close everything | | |
| ▲ | fishpham 13 hours ago | parent [-] | | As always, the limit is human bandwidth. But that's basically what AI-forward companies are doing now. I would be curious which tasks OP commenter has that couldn't be done by an agent (assuming they're a SWE) | | |
| ▲ | Analemma_ 13 hours ago | parent [-] | | This sounds bogus to me: if AI really could close 100% of your backlog with just a couple more humans in the loop, you’d hire a bunch of temps/contractors to do that, then declare the product done and lay off everybody. How come that isn’t happening? | | |
| ▲ | fishpham 12 hours ago | parent [-] | | Because there's an unlimited amount of work to do. This is the same reason you are not fired once completing a feature :-) The point of hiring a FTE is to continue to create work that provides business value. For your analogy, FTEs often do that by hiring temp, and you can think of the agent as the new temp in this case - the human drives an infinite amount of them | | |
| ▲ | catmanjan 16 minutes ago | parent | next [-] | | Sounds more like busy work rather than something that makes money | |
| ▲ | sarchertech 5 hours ago | parent | prev [-] | | Why hasn’t any of the software I use started shipping features at a breakneck speed then? The only thing any of them have added is barely working AI features. Why aren’t there 10x the number of games on steam? Why aren’t people releasing new integrated programming language/OS/dev environments? Why does our backlog look exactly the same as when I left for posterity leave 4 months ago? | | |
| ▲ | fishpham 4 hours ago | parent [-] | | Questions posed in bad faith can only be answered by the author. | | |
| ▲ | sarchertech 4 hours ago | parent [-] | | Someone asked why the backlog doesn’t get finished. You answered that it does but the backlog just refills. So I asked where is the backlog evidence that the original backlog was completed. I’m still waiting for the evidence. I still haven’t seen externally verifiable evidence that AI is a net productivity boost for the ability to ship software. That doesn’t mean that it isn’t. It does mean that it isn’t big enough to be obvious. I’m very closet watching every external metric I can find. Nothing yet. Just saw the steam metrics for January. Fewer titles than January last year. |
|
|
|
|
|
| |
| ▲ | rockbruno 13 hours ago | parent | prev | next [-] | | I think the "well defined prompt" is precisely what the person you responded to is alluring to. They are saying they don't get worried because AI doesn't get the job done without someone behind it that knows exactly what to prompt. | |
| ▲ | dwa3592 13 hours ago | parent | prev [-] | | >>I look at my ticket tracker and I see basically 100% of it that can be done by AI. That's a sign that you have spurious problems under those tickets or you have a PM problem. Also, a job is a not a task- if your company has jobs which is a single task then those jobs would definitely be gone. |
|
|
| ▲ | vincent_s 4 hours ago | parent | prev | next [-] |
| Ha, this triggered me. I'm building exactly this. It's a coding agent that takes a ticket from your tracker, does the work asynchronously, and replies with a pull request. It does progressively understand the codebase. There's a pre-warming step so it's already useful on the first ticket, but it gets better with each one it completes. The agent itself is done and working well. Right now I'm building out the infrastructure to offer it as a SaaS. If anyone wants to try it, hit me up. Email is in my profile. Website isn't live yet, but I'm putting together a waitlist. |
|
| ▲ | yodsanklai 9 hours ago | parent | prev | next [-] |
| > ~0% of them can be implemented by AI as it exists today I think it's more nuanced than that. I'd say that
- 0% can't be implemented by AI
- but a lot of them can be implemented much faster thanks to AI
- a lot of them can be implemented slower when using AI (because author has to fix hallucinations, revert changes that caused bugs) As we learn to use these tools, even in their current state, they will increase productivity by some factor and reduce needs for programmers. |
| |
| ▲ | sarchertech 6 hours ago | parent [-] | | What factor of increased productivity will lead to reduced need for programmers? I have seen numerous 25-50% productivity boosts over my career. Not a single one of them reduced the overall need for programmers. I can’t even think of one that reduced the absolute number of programmers in a specific field. |
|
|
| ▲ | danesparza 13 hours ago | parent | prev | next [-] |
| Apparently you haven't seen ChatGPT enterprise and codex. I have bad news for you ... |
| |
| ▲ | gordonhart 13 hours ago | parent [-] | | Codex with their flagship model (currently GPT-5.3-Codex) is my daily driver. I still end up doing a lot of steering! |
|
|
| ▲ | pupppet 12 hours ago | parent | prev | next [-] |
| We're all slowly but surely lowering our standards as AI bombards us with low-quality slop. AI doesn't need to get better, we all just need to keep collectively lowering our expectations until they finally meet what AI can currently do, and then pink-slips away. |
| |
| ▲ | tines 4 hours ago | parent [-] | | Exactly. This happens in every aspect of life. Something convenient comes along and people will accommodate it despite it being worse, because people don’t care. |
|
|
| ▲ | ninetyninenine 13 hours ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | zozbot234 11 hours ago | parent | prev [-] |
| > Once somebody cracks the memory problem and ships an agent that progressively understands the business and the codebase, then I'll start worrying. Um, you do realize that "the memory" is just a text file (or a bunch of interlinked text files) written in plain English. You can write these things out yourself. This is how you use AI effectively, by playing to its strengths and not expecting it to have a crystal ball. |