| ▲ | hackerInnen 10 hours ago |
| I just subscribed this month again because I wanted to have some fun with my projects. Tried out opus 4.6 a bit and it is really really bad. Why do people say it's so good? It cannot come up with any half-decent vhdl. No matter the prompt. I'm very disappointed. I was told it's a good model |
|
| ▲ | anon7000 10 hours ago | parent | next [-] |
| because they’re using it for different things where it works well and that’s all they know? |
|
| ▲ | adwn 9 hours ago | parent | prev | next [-] |
| And yet another "AI doesn't work" comment without any meaningful information. What were your exact prompts? What was the output? This is like a user of conventional software complaining that "it crashes", without a single bit of detail, like what they did before the crash, if there was any error message, whether the program froze or completely disappeared, etc. |
| |
| ▲ | emp17344 5 hours ago | parent | next [-] | | This is quite hostile. Yes, criticism is valid without an accompanying essay detailing every aspect of the associated environment, because these tools are still quite flawed. | |
| ▲ | 939373838 9 hours ago | parent | prev [-] | | [flagged] |
|
|
| ▲ | rurban 10 hours ago | parent | prev [-] |
| Because it was good until January 2026, then it detoriated into a opus-3.1. Probably given much less context windows or ram. |
| |
| ▲ | toomim 9 hours ago | parent [-] | | It released in February 2026. | | |
| ▲ | hxugufjfjf 8 hours ago | parent | next [-] | | I don’t think I’ve ever seen otherwise reasonable people go completely unhinged over anything like they do with Opus | | |
| ▲ | solenoid0937 8 hours ago | parent | next [-] | | I've seen a similar psychological phenomenon where people like something a lot, and then they get unreasonably angry and vocal about changes to that thing. Usage limits are necessary but I guess people expect more subsidized inference than the company can afford. So they make very angry comments online. For example, there is no evidence that 4.6 ever degraded in quality: https://marginlab.ai/trackers/claude-code-historical-perform... | | |
| ▲ | Capricorn2481 8 hours ago | parent [-] | | > Usage limits are necessary but I guess people expect more subsidized inference than the company can afford. So they make very angry comments online This is reductive. You're both calling people unreasonably angry but then acknowledging there's a limit in compute that is a practical reality for Anthropic. This isn't that hard. They have two choices, rate limit, or silently degrade to save compute. I have never hit a rate limit, but I have seen it get noticeably stupider. It doesn't make me angry, but comments like these are a bit annoying to read, because you are trying to make people sound delusional while, at the same time, confirming everything they're saying. I don't think they have turned a big knob that makes it stupider for everyone. I think they can see when a user is overtapping their $20 plan and silently degrade them. Because there's no alert for that. Which is why AI benchmark sites are irrelevant. | | |
| ▲ | scrawl 7 hours ago | parent [-] | | just my perspective: i pay $20/month and i hit usage limits regularly. have never experienced performance degradation. in fact i have been very happy with performance lately. my experience has never matched that of those saying model has been intentionally degraded. have been using claude a long time now (3 years). i do find usage limits frustrating. should prob fork out more... |
|
| |
| ▲ | unethical_ban 3 hours ago | parent | prev [-] | | That's what I thought today reading the comments in the Mozilla Thunderbolt thread today. Something about Mozilla absolutely sets people off. |
| |
| ▲ | ACCount37 9 hours ago | parent | prev [-] | | Doesn't matter. My vibes say it got bad in January 2026. Thus, they secretly nerfed Opus 4.6 in January 2026. The fact that it didn't exist back then is completely and utterly irrelevant to my narrative. | | |
| ▲ | MattSayar 7 hours ago | parent | next [-] | | I recognize the sarcasm. The data I can find says it's performing at baseline however? https://marginlab.ai/trackers/claude-code/ | | |
| ▲ | ACCount37 7 hours ago | parent [-] | | Yeah, that's my point. Humans are not reliable LLM evaluators. "Secret model nerfs" happen in "vibes" far more often than they do in any reality. |
| |
| ▲ | Der_Einzige 9 hours ago | parent | prev [-] | | This but unironically. "I reject your reality, and substitute my own". It worked for cheeto in chief, and it worked for Elon, so why not do it in our normal daily lives? |
|
|
|