| ▲ | g-mork 6 hours ago |
| My answer to this is simply rolling back to the pro plan for interactive usage in the coming month, and forcefully cutting myself over to one of the alternative Chinese models to just get over the hump and normalise API pricing at a sensible rate with sensible semantics. Dealing with Claude going into stupid mode 15 times a day, constant HTTP errors, etc. just isn't really worth it for all it does. I can't see myself justifying $200/mo. on any replacement tool either, the output just doesn't warrant it. I think we all jumped on the AI mothership with our eyes closed and it's time to dial some nuance back into things. Most of the time I'm just using Opus as a bulk code autocomplete that really doesn't take much smarts comparatively speaking. But when I do lean on it for actual fiddly bug fixing or ideation, I'm regularly left disappointed and working by hand anyway. I'd prefer to set my expectations (and willingness to pay) a little lower just to get a consistent slightly dumb agent rather than an overpriced one that continually lets me down. I don't think that's a problem fixed by trying to swap in another heavily marketed cure-all like Gemini or Codex, it's solved by adjusting expectations. In terms of pricing, $200 buys an absolute ton of GLM or Minimax, so much that I'd doubt my own usage is going to get anywhere close to $200 going by ccusage output. Minimax generating a single output stream at its max throughput 24/7 only comes to about $90/mo. |
|
| ▲ | Syntaf 2 hours ago | parent | next [-] |
| I put in probably thousands of Claude session hours a month, aggregated across work + personal. I must be missing something or supremely lucky because I feel like I’ve never hit these “stupid” moments. If I do, it’s probably because I forgot to switch off of haiku for some tiny side thing I was doing before going back to planning. |
|
| ▲ | SkyPuncher 6 hours ago | parent | prev | next [-] |
| I literally hit my 5 hour window limit in 1.5 hours every single day now. 2 weeks ago, I had only hit my limit a single time and that was when I had multiple agents doing codebase audits. |
| |
| ▲ | Aurornis 6 hours ago | parent | next [-] | | Anthropic had a special extra usage promotion going on during non-peak hours that ended recently. They didn’t do a great job of explaining it. I wonder how many people got used to the 2X limits and now think Anthropic has done something bad by going back to normal | | |
| ▲ | stavros 4 hours ago | parent | next [-] | | They also reduced the peak time limits, so it's not just the promotion. | |
| ▲ | SkyPuncher 3 hours ago | parent | prev | next [-] | | Naw, it's not that. This is business-day usage for all of it. | |
| ▲ | greenavocado an hour ago | parent | prev | next [-] | | Irrelevant. I had at least ten times more usage then at any time | |
| ▲ | paulddraper 4 hours ago | parent | prev [-] | | Could it also have anything to do with Anthropic being deliberately opaque about usage in general? |
| |
| ▲ | Razengan 2 hours ago | parent | prev | next [-] | | I've been using Codex extensively, 5.4 at "Extra High" and yet to hit a limit. The $20 plan | |
| ▲ | estimator7292 6 hours ago | parent | prev [-] | | They've been running a "double credits" promo for several weeks, which expired on the first of this month. |
|
|
| ▲ | bethekind 5 hours ago | parent | prev | next [-] |
| I think my next steps are:
1) try out openai $20/month. I've heard they're much more generous.
2) try out open router free models. I don't need geniuses, so long as I can see the thinking (something that Claude code obfuscates by default) I should be good. I've heard good things about the CLIO harness and want to try openrouter+clio |
| |
| ▲ | beering 5 hours ago | parent | next [-] | | Word on the street is that Opus is much much larger of a model than GPT-5.4 and that’s why the rate limits on Codex are so much more generous. But I guess you could also just switch to Sonnet or Haiku in Claude Code? | |
| ▲ | admiralrohan 4 hours ago | parent | prev [-] | | Openrouter free models have 50 requests per day limit + data collection. As per their doc. | | |
| ▲ | nodja 3 hours ago | parent [-] | | You can charge $10 on the account and get unlimited requests. I abused this last week with the nemotron super to test out some stuff and made probably over 10000 requests over a couple of days and didn't get blocked or anything, expect 5xx errors and slowdowns tho. |
|
|
|
| ▲ | danpalmer 6 hours ago | parent | prev | next [-] |
| Please don't use grossly offensive terms in this forum. That sort of language is not welcome here. |
| |
|
| ▲ | colechristensen 6 hours ago | parent | prev | next [-] |
| Every service is being sold at a deep discount chasing market share, but it's not lasting forever. |
| |
| ▲ | g-mork 6 hours ago | parent [-] | | Speaking only personally of course, I'm completely over the chat idiom in almost every way. Where is all this future demand coming from? By the time Android lands a God mode ultimate voice assistant it's pretty much guaranteed I will be well beyond the point where I'd want to use it. The whole thing is starting to remind me of 3G video calling where the networks thought it'd change everything, and by the end of it with all the infrastructure in place, the average user has made something like 0.001 3G-native video calls over the lifetime of their usage. Would really love some path forward where the AI parts only poke out as single fields in traditional user interfaces and we can forget this whole episode | | |
| ▲ | colechristensen 5 hours ago | parent [-] | | I don't understand this perspective. I can't imaging a point where I won't want to ask "what's the weather like?" "please turn off the lights" "what is the airspeed of an unladen swallow?" likewise chatting through directing it to build something or solve a problem, voice or typing will each have their place. And video calling did take off, plenty of people use facetime and almost everybody working in an office uses some form of video calls. Criticizing the early attempts at getting video calling working because they hadn't taken off yet (I remember them being advertised on "video phones" with 56k modems), of course someone was going to have the idea and implement before it was quite reasonable. | | |
| ▲ | neonstatic 4 hours ago | parent [-] | | > I can't imaging a point where I won't want to ask "what's the weather like?" "please turn off the lights" To help with understanding that perspective, I cannot imagine a scenario where I would ask a device connected to the internet to turn off the lights. I literally never wanted this. A physical switch is a 100% non negotiable for me. I feel the same way about non-mechanical car doors. Perhaps due to that outlook I was always puzzled about the entire idea of an "assistant". It's interesting for me to see, that there are people out there who actually want that "assistant". | | |
| ▲ | wat10000 3 hours ago | parent [-] | | You don't watch Iron Man and want a JARVIS? Current systems are pretty far away from that, but that's the overall draw. | | |
| ▲ | neonstatic 2 hours ago | parent [-] | | I don't watch superhero stuff. But even with a more classical example of Space Odyssey 2001 - a talking computer has never been something I found even remotely interesting. It took me months to give LLMs a serious try due to this. |
|
|
|
|
|
|
| ▲ | rhodysurf 5 hours ago | parent | prev | next [-] |
| This is what I did, downgraded to pro and pay for opencode zen for the open models. I like the combo of the two |
|
| ▲ | codybontecou 5 hours ago | parent | prev | next [-] |
| Are you using the Chinese models through their individual services or via an intermediary layer? |
|
| ▲ | Aurornis 5 hours ago | parent | prev | next [-] |
| > I think we all jumped on the AI mothership with our eyes closed and it's time to dial some nuance back into things. I’m kind of confused by these takes from HN readers. I could see LinkedIn bros getting reality checked when they finally discover that LLMs aren’t magic, but I’m confused about how a developer could go all-in on AI and not immediately realize the limitations of the output. |
|
| ▲ | zdragnar 6 hours ago | parent | prev [-] |
| > I think we all jumped on the AI mothership with our eyes closed Oh no, there's plenty of us willing to say we told you so. What's more interesting to me is what it's going to look like if big companies start removing "AI usage" from their performance metrics and cease compelling us to use it. More than anything else, that's been the dumbest thing to happen with this whole craze. |