| ▲ | 2001zhaozhao 5 hours ago | |||||||||||||
That is exactly the impression I get from the claude code team, and by extension some of their recent launches like Cowork and Design. And of course with the growth team or whoever is in charge of the subscription and quota side of things. They just do the basic experiment -> ship workflow over and over again, doing whatever optimizes their product in the short term, and never seem to step back and think about the full long-term impact of their changes. They evidently seem to not even consider immediate regressions or negative blowback from users if it's not within the area of expertise of the guy who ships the change. That is despite their other teams (especially alignment) having a track record of being fairly well thought-out and intelligent. To the guys at Anthropic's product teams, every problem is a data science problem that you slap an A/B test onto, and they seem to think that the A/B test is all that's needed, and actual verification and thinking things through is overrated af. That's what leads to countless regressions in Claude Code as well as removing claude code from the pro plan in their product page for a few hours (lol). | ||||||||||||||
| ▲ | ffsm8 5 hours ago | parent [-] | |||||||||||||
Tbf, their harness was surprisingly ahead of the curve for most of the last year.. Are this point, the difference is mostly made up by issues like the OP has, so you're likely better off using eg pi (-agent) and writing your own custom skills and extensions (or any of the other harnesses the providers create, even copilot-cli has gotten decent nowadays) | ||||||||||||||
| ||||||||||||||