▲ | pqdbr 5 days ago | ||||||||||||||||||||||
Same here. Even with Opus in Claude Code I'm getting terrible results, sometimes feeling we went back to the GPT 3.5 eon. And it seems they are implementing heavily token-saving measures: the model does not read context anymore unless you force it to, making up method calls as it goes. | |||||||||||||||||||||||
▲ | mh- 5 days ago | parent | next [-] | ||||||||||||||||||||||
The simplest thing I frequently ask of regular Claude (not Code) in the desktop app: "Use your web search tool to find me the go-to component for doing xyz in $language $framework. Always link the GitHub repo in your response." Previously Sonnet 4 would return a good answer to this at least 80% of the time. Now even Opus 4.1 with extended thinking frequently ignores my ask for it to use the search tool, which allows it to hallucinate a component in a library. Or maybe an entire repo. It's gone backwards severely. (If someone from Anthropic sees this, feel free to reach out for chat IDs/share links. I have dozens.) | |||||||||||||||||||||||
| |||||||||||||||||||||||
▲ | j45 5 days ago | parent | prev [-] | ||||||||||||||||||||||
I’m running into this as well. Might be Claude optimizing for general use cases compared to code and that affecting the code side? Feels strange, because Claude api isn’t the same as the web tool so I didn’t expect Claude code to be the same. It might be a case of having to learn to read Claude best practice docs and keep up with them. Normally I’d have Claude read them itself and update an approach to use. Not sure that works as well anymore. |