▲ | tropicalfruit 5 days ago | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
reading all the shilling of Claude and GPT i see here often I feel like i'm being gaslighted i've been using premium tiers of both for a long time and i really felt like they've been getting worse especially Claude I find super frustrating and maddening, misunderstanding basic requests or taking liberties by making unrequested additions and changes i really had this sense of enshittification, almost as if they are no longer trying to serve my requests but do something else instead like i'm victim of some kind of LLM a/b testing to see how far I can tolerate or how much mental load can be transferred back onto me | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | tibbar 5 days ago | parent | next [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
While it's possible that the LLMs are intentionally throttled to save costs, I would also keep in mind that LLMs are now being optimized for new kinds of workflows, like long-running agents making tool calls. It's not hard to imagine that improving performance on one of those benchmarks comes at a cost to some existing features. | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | macawfish 5 days ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
I suspect that it may not necessarily be that they're getting objectively _worse_ as much as that they aren't static products. They're constantly getting their prompts/context engines tweaked in ways that surely break peoples' familiar patterns. There really needs to be a way to cheaply and easily anchor behaviors so that people can get more consistency. Either that or we're just going to have to learn to adapt. | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | simonw 5 days ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
Anthropic have stated on the record several times that they do not update the model weights once they have been deployed without also changing the model ID. | |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
▲ | TechDebtDevin 5 days ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||||||||||||||||||||||||||
If Anthropic made Deepthink 3.5 it would be AGI, I never use > 3.5 |