| ▲ | onion2k 5 hours ago |
| But you can’t just not review things! Actually you can. If you shift the reviews far to the left, and call them code design sessions instead, and you raise problems on dailys, and you pair programme through the gnarly bits, then 90% of what people think a review should find goes away. The expectation that you'll discover bugs and architecture and design problems doesn't exist if you've already agreed with the team what you're going to build. The remain 10% of things like var naming, whitespace, and patterns can be checked with a linter instead of a person. If you can get the team to that level you can stop doing code reviews. You also need to build a team that you can trust to write the code you agreed you'd write, but if your reviews are there to check someone has done their job well enough then you have bigger problems. |
|
| ▲ | Certhas 2 minutes ago | parent | next [-] |
| That's partly the point of the article, except the article acknowledges that this is organizationally hard: > You get things like the famous Toyota Production System where they eliminated the QA phase entirely. > [This] approach to manufacturing didn’t have any magic bullets. Alas, you can’t just follow his ten-step process and immediately get higher quality engineering. The secret is, you have to get your engineers to engineer higher quality into the whole system, from top to bottom, repeatedly. Continuously. > The basis of [this system] is trust. Trust among individuals that your boss Really Truly Actually wants to know about every defect, and wants you to stop the line when you find one. Trust among managers that executives were serious about quality. Trust among executives that individuals, given a system that can work and has the right incentives, will produce quality work and spot their own defects, and push the stop button when they need to push it. > I think we’re going to be stuck with these systems pipeline problems for a long time. Review pipelines — layers of QA — don’t work. Instead, they make you slower while hiding root causes. Hiding causes makes them harder to fix. |
|
| ▲ | alkonaut 2 hours ago | parent | prev | next [-] |
| This falls for the famous "hours of planning can save minutes of coding". Architecture can't (all) be planned out on a whiteboard, it's the response to the difficulty you only realize as you try to implement. If you can agree what to build and how to build it and then it turns out that actually is a working plan - then you are better than me. That hasn't happened in 20 years of software development. Most of what's planned falls down within the first few hours of implementation. Iterative architecture meetings will be necessary. But that falls into the pit of weekly meeting. |
| |
| ▲ | 2OEH8eoCRo0 5 minutes ago | parent | next [-] | | I've worked waterfall (defense) and while I hated it at the time I'd rather go back to it. Today we move much faster but often build the wrong thing or rewrite and refactor things multiple times. In waterfall we move glacially but what we would build sticks. Also, with so much up front planning the code practically writes itself. I'm not convinced there's any real velocity gains in agile when factoring in all the fiddling, rewrites, and refactoring. > Most of what's planned falls down within the first few hours of implementation. Not my experience at all. | |
| ▲ | AIorNot 7 minutes ago | parent | prev [-] | | “Everyone has a plan until they get punched in the mouth" - Mike Tyson |
|
|
| ▲ | loire280 5 hours ago | parent | prev | next [-] |
| I've seen engineers I respect abandon this way of working as a team for the productivity promise of conjuring PRs with a coding agent. It blows away years of trust so quickly when you realize they stopped reviewing their own output. |
| |
| ▲ | overfeed 4 hours ago | parent | next [-] | | Perhaps due to FOMO outbreak[1], upper management everywhere has demanded AI-powered productivity gains, based on LoC/PR metrics, it looks like they are getting it. 1. The longer I work in this industry, the more it becomes clear that CxO's aren't great at projecting/planning, and default to copy-cat, herd behaviors when uncertain. | | |
| ▲ | serial_dev an hour ago | parent | next [-] | | Software engineers are pushed to their limits (and beyond). Unrealistic expectations are established by Twitter "I shipped an Uber clone in 2 hours with Claude" forcing every developer to crank out PRs, managers are on the look out for any kind of perceived inefficiency in tools like GetDX and Span. If devs are expected to ship 10x faster (or else!), then they will find a way to ship 10x faster. | | |
| ▲ | pydry 5 minutes ago | parent [-] | | I always found it weird how most management would do almost anything other than ask their dev team "hey, is there any way to make you guys more productive?" Ive had metrics rammed down my throat, Ive had AI rammed down my throat, Scrum rammed down my throad and Ive had various other diktats rammed down my throat. 95% of these slowed us down. The only time ive been asked is when there is a deadline and it's pretty clear we arent going to hit it and even then they're interested in quick wins like "can we bring lunch to you for a few weeks?", not systemic changes. Im convinced that the companies which seek developer autonomy will leave the ones which seek to maximize token usage in the dust. |
| |
| ▲ | tripledry 3 hours ago | parent | prev | next [-] | | Would love to be a fly on the wall for a couple of months to see what corporate CxO's actually do. Surely I could do a mediocre job as a CxO by parroting whatever is hot on Linkedin. Probably wouldn't be a massively successful one, but good enough to survive 2 years and have millions in the bank for that, or get fired and get a golden parachute. (half) joking - most likely I'm massively trivializing the role. | | |
| ▲ | raphlinus an hour ago | parent | next [-] | | Funny enough, the author of this blog post wrote another one on exactly that topic, entitled "What do executives do, anyway?"[1]. If you read it, you'll find it's written from quite an interesting perspective, not quite "fly on the wall," but perhaps as close as you're going to get in a realistic scenario. [1]: https://apenwarr.ca/log/20190926 | |
| ▲ | arethuza 2 hours ago | parent | prev | next [-] | | "Surely I could do a mediocre job as a CxO by parroting whatever is hot on Linkedin" Having worked for a pretty decent CIO of a global business I'd say his main job was to travel about speak to other senior leaders and work out what business problems they had and try and work out, at a very high level, how technology would fit into that addressing those problems. Just parroting latest technology trends would, I suspect, get you sacked within a few weeks. | |
| ▲ | eptcyka 3 hours ago | parent | prev [-] | | A charitable explanation for what CxOs do is that they figure out their strategic goals and then focus really hard on ways to herd cats en masse to achieve the goals in an efficient manner. Some people end up doing a great job, some do so accidentally, other just end up doing a job. Sometimes parroting some linkadink drivel is enough to keep the ship on course - usually because the winds are blowing in the right direction or the people at the oars are working well enough on their own. |
| |
| ▲ | 3 hours ago | parent | prev [-] | | [deleted] |
| |
| ▲ | onion2k 5 hours ago | parent | prev | next [-] | | Putting too much trust in an agent is definitely a problem, but I have to admit I've written about a dozen little apps in the past year without bothering to look at the code and they've all worked really well. They're all just toys and utilities I've needed and I've not put them into a production system, but I would if I had to. Agents are getting really good, and if you're used to planning and designing up front you can get a ton of value from them. The main problem with them that I see today is people having that level of trust without giving the agent the context necessary to do a good job. Accepting a zero-shotted service to do something important into your production codebase is still a step too far, but it's an increasingly small step. | | |
| ▲ | camillomiller 4 hours ago | parent [-] | | >> Putting too much trust in an agent is definitely a problem, but I have to admit I've written about a dozen little apps in the past year without bothering to look at the code and they've all worked really well. They're all just toys and utilities I've needed and I've not put them into a production system, but I would if I had to. I have been doing this to, and I've forgotten half of them. For me the point is that this usage scenario is really good, but it also has no added value to it, really. The moment Claude Code raises it prices 2x this won't be viable anymore, and at the same time to scale this to enterprise software production levels you need to spend on an agent probably as much as hiring two SWEs, given that you need at least one to coordinate the agents. | | |
| ▲ | jeremyjh 3 hours ago | parent | next [-] | | Deepseek v3.2 tokens are $0.26/0.38 on OpenRouter. That model - released 4 months ago - isn't really good enough by today's standards, but its significantly stronger than Opus 4.1, which was only released last August! In 12 months I think its reasonable to expect there will be a model with less cost than that which is significantly stronger than anything available now. And no, it isn't ONLY because VC capital is being burned to subsidize cost. That is impossible for the dozen smaller providers offering service at that cost on OpenRouter who have to compete with each other for every request and also have to pay compute bills. Qwen3.5-9B is stronger than GPT-4o and it runs on my laptop. That isn't just benchmarks either. Models are getting smaller, cheaper and better at the same time and this is going to continue. | |
| ▲ | onion2k 3 hours ago | parent | prev [-] | | I think Claude could raise it's prices 100x and people would still use it. It'd just shift to being an enterprise-only option and companies would actually start to measure the value instead of being "Whee, AI is awesome! We're definitely going really fast now!" | | |
| ▲ | christophilus 2 hours ago | parent [-] | | 100x? You think people would pay $20k per month for Claude Code? Codex is as good (or very nearly) as Claude code. Open source models continue to improve. The open source harnesses will also continue to improve. Anthropic is good, but it has no moat. No way could they 100x their prices. |
|
|
| |
| ▲ | denkmoon an hour ago | parent | prev | next [-] | | I’m so disappointed to see the slip in quality by colleagues I think are better than that. People who used to post great PRs are now posting stuff with random unrelated changes, little structs and helpers all over the place that we already have in common modules etc :’( | |
| ▲ | nvardakas 2 hours ago | parent | prev [-] | | [dead] |
|
|
| ▲ | roncesvalles 2 hours ago | parent | prev | next [-] |
| >shift the reviews far to the left, and call them code design sessions instead, and you raise problems on dailys, and you pair programme through the gnarly bits hell in one sentence |
|
| ▲ | ap99 34 minutes ago | parent | prev | next [-] |
| Unless you're covering 100% of edge/corner cases during planning (including roughly how they're handled) then there is still value in code reviews. You conveniently brushed this under the rug of pair programming but of the handful of companies I've worked at, only one tried it and just as an experiment which in the end failed because no one really wanted to work that way. I think this "don't review" attitude is dangerous and only acceptable for hobby projects. |
| |
| ▲ | zingar 27 minutes ago | parent [-] | | Reviews are vital for 80% of the programmers I work with but I happily trust the other 20% to manage risk, know when merging is safe without review, and know how to identify and fix problems quickly. With or without pairing. The flip side is that if the programmer and the reviewer are both in the 80% then the review doesn’t decrease the risk (it may even increase it). |
|
|
| ▲ | riffraff 5 hours ago | parent | prev | next [-] |
| This is also the premise of pair programming/extreme programming: if code review is useful, we should do it all the time. |
| |
| ▲ | roncesvalles 2 hours ago | parent [-] | | Anyone who talks about pair programming has either never done them or just started doing them last week. | | |
| ▲ | orwin 33 minutes ago | parent | next [-] | | I like pair programming. Not everytime or even everyday, but to shadow a junior a few hours a week, or to work with another senior on a complex/new subject? It's fine. | |
| ▲ | interroboink an hour ago | parent | prev | next [-] | | My sense is that there is a narrow slice of software developers who genuinely do flourish in a pair programming environment. These are people who actually work through their thoughts better with another person in the loop. They get super excited about it and make the common mistake of "if it works for me, it will work for everybody" and shout it from the hilltops. Then there are the people who program best in a fugue state and the idea of having to constantly break that to transform their thoughts into words and human interaction is anathema. I say this as someone who just woke up in the wee hours of the morning when nobody else is around so I can get some work done (: | |
| ▲ | nicoburns 42 minutes ago | parent | prev [-] | | I like pair programming for certain problems: things that are genuinely hard / pushing the boundaries of both participants knowledge and abilities. In those scenarios sometimes two minds can fill in each other's gaps much more efficiently than either can work alone. |
|
|
|
| ▲ | Swizec 5 hours ago | parent | prev | next [-] |
| > You also need to build a team that you can trust to write the code you agreed you'd write I tell every hire new and old “Hey do your thing, we trust you. Btw we have your phone number. Thanks” Works like a charm. People even go out of their way to write tests for things that are hard to verify manually. And they verify manually what’s hard to write tests for. The other side of this is building safety nets. Takes ~10min to revert a bad deploy. |
| |
| ▲ | pdhborges 2 hours ago | parent | next [-] | | > The other side of this is building safety nets. Takes ~10min to revert a bad deploy. Does it? Reverting a bad deploy is not only about running the previous version. Did you mess up data? Did you take actions on third party services that that need to be reverted? Did it have legal reprecursions? | | | |
| ▲ | namanyayg 2 hours ago | parent | prev [-] | | How does the phone number help? | | |
| ▲ | chaboud 33 minutes ago | parent | next [-] | | That's the polite version of "we know where you live". Telling someone you have their phone number is a way of saying "we'll call you and expect immediacy if you break something." Wanna be treated like an adult? Cool. You'll also be held accountable like an adult. | |
| ▲ | swiftcoder 23 minutes ago | parent | prev | next [-] | | Never received a phone call at 5am on a Sunday because a bug is causing a valued customer to lose $10k/minute, and by the way, the SVP is also on the line? Lucky bastard | |
| ▲ | gabriel-uribe 2 hours ago | parent | prev [-] | | Presumably they will be contacted if there's a problem. So the hire has an interest in not creating problems. |
|
|
|
| ▲ | ozim 3 hours ago | parent | prev | next [-] |
| Then you spend all your budget on code design sessions and have nothing to show to the customer. |
|
| ▲ | totetsu 5 hours ago | parent | prev | next [-] |
| This seems to be a core of the problem with trying to leave things to autonomous agents .. The response to Amazons agents deleting prod was to implement review stages https://blog.barrack.ai/amazon-ai-agents-deleting-production... |
|
| ▲ | ramon156 4 hours ago | parent | prev | next [-] |
| I'm in a company that does no reviews and I'm medior. The tools we make is not interesting at all, so it's probably the best position I could ask for. I occasionally have time to explore some improvements, tools and side projects (don't tell my boss about that last one) |
|
| ▲ | thrwaway55 2 hours ago | parent | prev | next [-] |
| Okay but Claude is a fucking moron. |
|
| ▲ | froh 4 hours ago | parent | prev | next [-] |
| yes! and it also works for me when working with ai. that produces much better results, too, when I first so a design session really discussing what to build. then a planning session, in which steps to build it ("reviewability" world wonder). and then the instruction to stop when things get gnarly and work with the hooman. does anyone here have a good system prompt for that self observance "I might be stuck, I'm kinda sorta looping. let's talk with hooman!"? |
|
| ▲ | DeathArrow 2 hours ago | parent | prev | next [-] |
| The issue is that every review adds a lot of delay. A lot of alignment and pair programming won't be time expensive? |
|
| ▲ | hinkley 3 hours ago | parent | prev | next [-] |
| These systems make it more efficient to remove the actively toxic members for your team. Beligerence can be passively aggressively “handled” by additional layers but at considerable time and emotional labor cost to people who could be getting more work done without having to coddle untalented assholes. |
| |
|
| ▲ | rendall 4 hours ago | parent | prev | next [-] |
| Yes. This is the way. Declarative design contracts are the answer to A.I. coders. A team declares what they want, agents code it together with human supervision. Then code review is just answering the question "is the code conformant with the design contract?" But. The design contract needs review, which takes time. |
|
| ▲ | anal_reactor 5 hours ago | parent | prev | next [-] |
| I never review PRs, I always rubber-stamp them, unless they come from a certified idiot: 1. I don't care because the company at large fails to value quality engineering. 2. 90% of PR comments are arguments about variable names. 3. The other 10% are mistakes that have very limited blast radius. It's just that, unless my coworker is a complete moron, then most likely whatever they came up with is at least in acceptable state, in which case there's no point delaying the project. Regarding knowledge share, it's complete fiction. Unless you actually make changes to some code, there's zero chance you'll understand how it works. |
| |
| ▲ | swiftcoder 20 minutes ago | parent | next [-] | | > 2. 90% of PR comments are arguments about variable names. This sort of comment is meaningless noise that people add to PRs to pad their management-facing code review stats. If this is going on in your shop, your senior engineers have failed to set a suitable engineering culture. If you are one of the seniors, schedule a one-on-one with your manager, and tell them in no uncertain terms that code review stats are off-limits for performance reviews, because it's causing perverse incentives that fuck up the workflow. | |
| ▲ | recursivecaveat 5 hours ago | parent | prev | next [-] | | Do people really argue about variable names? Most reviews comments I see are fairly trivial, but almost always not very subjective. (Leftover debug log, please add comment here, etc) Maybe it helps that many of our seniors are from a team where we had no auto-formatter or style guide at all for quite a while. I think everyone should experience that a random mix of `){` and `) {` does not really impact you in any way beyond the mild irking of a crooked painting or something. There's a difference between aesthetically bothersome and actually harmful. Not to say that you shouldn't run a formatter, but just for some perspective. | | |
| ▲ | jffhn 3 hours ago | parent | next [-] | | >Do people really argue about variable names? Of course they do. A program's code is mostly a graph of names; they can be cornerstones of its clarity, or sources of confusion and bugs. The first thing I do when debugging is ensuring proper names, sometimes that's enough to make the bug obvious. | | |
| ▲ | vaylian an hour ago | parent [-] | | The greatest barrier to understanding is not lack of knowledge but incorrect knowledge. That's why good names matter. And naming things is hard, which is why it makes sense to comment on variable names in a review. |
| |
| ▲ | anal_reactor 4 hours ago | parent | prev [-] | | Yes. 80% of comments to my PRs are "change _ to -" or something like that. | | |
| ▲ | blitzar 2 hours ago | parent [-] | | PR #467 - Reformat code from tabs to spaces PR #515 - Reformat code from spaces to tabs |
|
| |
| ▲ | _kidlike 3 hours ago | parent | prev | next [-] | | I'm very surprised by these comments... I regularly review code that is way more complicated that it should. The last few days I was going back and forth on reviews on a function that had originally cyclomatic complexity of 23. Eventually I got it down to 8, but I had to call him into a pair programming session and show him how the complexity could be reduced. | | |
| ▲ | servo_sausage 3 hours ago | parent [-] | | Someone giving work like that should be either junior enough that there is potential for training them, so your time investment is worth it, or managed out. Or it didn't really matter that the function was complex if the structure of what's surrounding it was robust and testable; just let it be a refactor or bug ticket later. |
| |
| ▲ | worldsayshi 4 hours ago | parent | prev | next [-] | | People always makes mistakes. Like forgetting to include a change. The point of PRs for me is to try to weed out costly mistakes. Automated tests should hopefully catch most of them though. | | |
| ▲ | Fargren an hour ago | parent [-] | | The point of PRs is not to avoid mistakes (though sometimes this can happen). Automated tests are the tool to weed out those kinds of mistakes. The point of PRs is to spread knowledge. I try to read every PR, even if it's already approved, so I'm aware of what changes there are in code I'm going to own. They are the RSS feed of the codebase. |
| |
| ▲ | devmor 5 hours ago | parent | prev | next [-] | | I used to do this! I can’t anymore, not with the advent of AI coding agents. My trust in my colleagues is gone, I have no reason to believe they wrote the code they asked me to put my approval on, and so I certainly don’t want to be on a postmortem being asked why I approved the change. Perhaps if I worked in a different industry I would feel like you do, but payments is a scary place to cause downtime. | |
| ▲ | cindyllm 5 hours ago | parent | prev [-] | | [dead] |
|
|
| ▲ | jauntywundrkind 5 hours ago | parent | prev [-] |
| I wonder what delayed continuous release would be like. Trust folks to merge semi-responsibly, but have a two week delay before actually shipping to give yourself some time to find and fix issues. Perhaps kind of a pain to inject fixes in, have to rebase the outstanding work. But I kind of like this idea of the org having responsibility to do what review it wants, without making every person have to coral all the cats to get all the check marks. Make it the org's challenge instead. |