| ▲ | Wowfunhappy 5 days ago |
| I'm probably not going to hit the weekly limit, but it makes me nervous that the limit is weekly as opposed to every 36 hours or something. If I do hit the limit, that's it for the entire week—a long time to be without a tool I've grown accustomed to! I feel like someone is going to reply that I'm too reliant on Claude or something. Maybe that's true, but I'd feel the same about the prospect of loosing ripgrep for a week, or whatever. Loosing it for a couple of days is more palatable. Also, I find it notable they said this will affect "less than 5% of users". I'm used to these types of announcements claiming they'll affect less than 1%. Anthropic is saying that one out of every 20 users will hit the new limit. |
|
| ▲ | el_benhameen 5 days ago | parent | next [-] |
| This is how I feel about the 100 msg/wk limit on o3 for the ChatGPT plus plan. There’s no way to see how much I’ve used, and it’s an important enough resource that my lizard brain wants to hoard it. The result is that I way underutilize my plan and go for one of the o4-mini models instead. I would much prefer a lower daily limit, but maybe the underutilization is the point of the weekly limit. *edited to change “pro” to “plus” |
| |
| ▲ | landl0rd 5 days ago | parent | next [-] | | You can tell how it’s intentional with both OpenAI and Anthropic by how they’re intentionally made opaque. I cant see a nice little bar with how much I’ve used versus have left on the given rate limits so it’s pressuring users to hoard. Because it prevents them from budgeting it out and saying “okay I’ve used 1/3 of my quota and it’s Wednesday, I can use more faster.” | | |
| ▲ | xpe 5 days ago | parent | next [-] | | > pressures users to hoard As a pedantic note, I would say 'ration'. Things you hoard don't magically go away after some period of time. | | |
| ▲ | zamadatix 4 days ago | parent | next [-] | | FWIW neither hoard nor ration imply anything about permanence of the thing to me. Whether you were rationed bread or you hoarded bread, the bread isn't going to be usable forever. At the same time whether you were rationed sugar or hoarded sugar, the sugar isn't going to expire (with good storage). Rationed/hoarded do imply, to me, something different about how the quantity came to be though. Rationed being given or setting aside a fixed amount, hoarded being that you stockpiled/amassed it. Saying "you hoarded your rations" (whether they will expire) does feel more on the money than "you ration your rations" from that perspective. I hope this doesn't come off too "well aktually", I've just been thinking about how I still realize different meanings/origins of common words later in life and the odd things that trigger me to think about it differently for the first time. A recent one for me was that "whoever" has the (fairly obvious) etymology of who+ever https://www.etymonline.com/word/whoever vs something like balloon, which has a comparatively more complex history https://www.etymonline.com/word/balloon | | |
| ▲ | mattkrause 4 days ago | parent | next [-] | | For me, the difference between ration and hoard is the uhh…rationality of the plan. Rationing suggests a deliberate, calculated plan: we’ll eat this much at these particular times so our food lasts that long. Hoard seems more ad hoc and fear-driven: better keep yet another beat-up VGA cable, just in case. | | |
| ▲ | jjani 4 days ago | parent [-] | | > Hoard seems more ad hoc and fear-driven: better keep yet another beat-up VGA cable, just in case. Counterexample: animals hoarding food for winter time, etc. | | |
| ▲ | nothrabannosir 4 days ago | parent [-] | | Rather a corroborating example than a counter, if you believe how many nuts squirrels lose sight of after burying them. | | |
| ▲ | xpe 4 days ago | parent [-] | | Exactly. How many random computer dongles and power supplies get buried in sundry boxes that are effectively lost to the world? |
|
|
| |
| ▲ | kanak8278 4 days ago | parent | prev | next [-] | | It can only happens in HackerNews that people talking about Claude Code limit, can start discussing what is a better a word for explaining it. :-) I just love this community for these silly things. | |
| ▲ | 14123newsletter 4 days ago | parent | prev | next [-] | | Isn't hoarding means you can get more bread ? While rationing means: "here is 1kg, use it however you want but you can't get more". | | |
| ▲ | zamadatix 4 days ago | parent [-] | | Hoarding doesn't really imply how you got it, just that you stockpile once you do. I think you're bang on rationing - it's about assigning the fixed amount. The LLM provider does the rationing, the LLM user hoards their rations. One could theoretically ration their rations out further... but that would require knowing the usage to the point to set the remaining fixed amounts - which is precisely whT's missing in the interface. |
| |
| ▲ | randomcarbloke 4 days ago | parent | prev [-] | | Bread can be rationed but cannot be hoarded. |
| |
| ▲ | nine_k 4 days ago | parent | prev | next [-] | | Rationing implies an ability to measure: this amount per day. But measuring the remaining amount is exactly what Claude Code API does not provide. So, back to hoarding. | |
| ▲ | landl0rd 4 days ago | parent | prev [-] | | Rationing is precisely what we want to do: I have x usage this week; let me determine precisely how much I can use without going over. Hoarding implies a less reasoned path of “I never know when I might run out so I must use as little as possible, save as much as I can.” One can hoard gasoline but it still expires past a point. |
| |
| ▲ | sothatsit 4 days ago | parent | prev | next [-] | | Anthropic also does this because they will dynamically change the limits to manage load. Tools like ccusage show you how much you've used and I can tell sometimes that I get limited with significantly lower usage than I would usually get limited for. | | |
| ▲ | TheOtherHobbes 4 days ago | parent [-] | | Which is a huge problem, because you literally have no idea what you're paying for. One day a few of hours of prompting is fine, another you'll hit your weekly limit and you're out for seven days. While still paying your subscription. I can't think of any other product or service which operates on this basis - where you're charged a set fee, but the access you get varies from hour to hour entirely at the provider's whim. And if you hit a limit which is a moving target you can't even check you're locked out of the service. It's ridiculous. Begging for a law suit, tbh. | | |
| ▲ | lukaslalinsky 4 days ago | parent [-] | | What happens when you have a gym membership, but you go there during their busy hours? What they could do is pay as you go, with pricing increasing with the demand (Uber style), but I don't think people would like that much. | | |
| ▲ | deeth_starr_v 4 days ago | parent [-] | | Your analogy would work if the gym would randomly suspend your membership for a week if you worked out too much during peak hours |
|
|
| |
| ▲ | canada_dry 4 days ago | parent | prev | next [-] | | OpenAI's "PRO" subscription is really a waste of money IMHO for this and other reasons. Decided to give PRO a try when I kept getting terrible results from the $20 option. So far it's perhaps 20% improved in complex code generation. It still has the extremely annoying ~350 line limit in its output. It still IGNORES EXPLICIT CONTINUOUS INSTRUCTIONS eg: do not remove existing comments. The opaque overriding rules that - despite it begging forgiveness when it ignores instructions - are extremely frustrating!! | | |
| ▲ | JoshuaDavid 4 days ago | parent | next [-] | | One thing that has worked for me when I have a long list of requirements / standards I want an LLM agent to stick to while executing a series of 5 instructions is to add extra steps at the end of the instructions like "6. check if any of the code standards are not met - if not, fix them and return to step 5" / "7. verify that no forbidden patterns from <list of things like no-op unit tests, n+1 query patterns, etc> exist in added code - if you find any, fix them and return to step 5" etc. Often they're better at recognizing failures to stick to the rules and fixing the problems than they are at consistently following the rules in a single shot. This does mean that often having an LLM agents so a thing works but is slower than just doing it myself. Still, I can sometimes kick off a workflow before joining a meeting, so maybe the hours I've spent playing with these tools will eventually pay for themselves in improved future productivity. | |
| ▲ | jmaker 4 days ago | parent | prev [-] | | There are things it’s great at and things it deceives you with. In many things I needed it to check something for me I knew was a problem, o3 kept insisting it were possible due to reasons a,b,c, and thankfully gave me links. I knew it used to be a problem so surprised I followed the links only to read black on white it still wasn’t. So I explained to o3 that it’s wrong. Two messages later we were back at square one. One week later it didn’t update its knowledge. Months later it’s still the same. But at things I have no idea about like medicine it feels very convincing. Am I in hazard? People don’t understand Dunning-Kruger. People are prone to biases and fallacies. Likely all LLMs are inept at objectivity. My instructions to LLMs are always strictness, no false claims, Bayesian likelihoods on every claim. Some modes ignore the instructions voluntarily, while others stick strictly to them. In the end it doesn’t matter when they insist on 99% confidence on refuted fantasies. | | |
| ▲ | namibj 4 days ago | parent [-] | | The problem is that all current mainstream LLMs are autoregressive decoder-only, mostly but not exclusively transformers.
Their math can't apply modifiers like "this example/attempt there is wrong due to X,Y,Z" to anything that came before the modifier clause in the prompt.
Despite how enticing these models are to train, these limitations are inherent. (For this specific situation people recommend going back to just before the wrong output and editing the message to reflect this understanding, as the confidently wrong output with no advisory/correcting pre-clause will "pollute the context": the model will look at the context for some aspects coded into high(-er)-layer token embeddings, inherently can't include the correct/wrong aspect because we couldn't apply the "wrong"/correction to the confidently-wrong tokens, thus retrieves the confidently-wrong tokens, and subsequently spews even more BS.
Similar to how telling a GPT2/GPT3 model it's an expert on $topic made it actually be better on said topic, this affirmation of that the model made an error will prime the model to behave in a way that it gets yelled at again... sadly.) |
|
| |
| ▲ | brookst 4 days ago | parent | prev [-] | | I think the simple product prioritization explanation makes way more sense than a a second-order conspiracy to trick people into hoarding. Reality is probably that there’s a backlog item to implement a view, but it’s hard to prioritize over core features. | | |
| ▲ | parineum 4 days ago | parent | next [-] | | > Reality is probably that there’s a backlog item to implement a view, but it’s hard to prioritize over core features. It's even harder to prioritize when the feature you pay to develop probably costs you money. | |
| ▲ | Zacharias030 4 days ago | parent | prev [-] | | I hear OpenAI and Anthropic are making tools that are supposedly pretty good at helping with creating a view from a backlog. Back to the conspiracy ^^ | | |
| ▲ | brookst 4 days ago | parent | next [-] | | If you’re on HN you’ve probably been around enough to know it’s never that simple. You implement the counter, now customer service needs to be able to provide documentation, users want to argue, async systems take hours to update, users complain about that, you move the batch accounting job to sync, queries that fail still end up counting, and on and on. They should have an indicator, for sure. But I at least have been around the block enough to know that declaring “it would be easy” for someone else’s business and tech stack is usually naive. | |
| ▲ | bluelightning2k 4 days ago | parent | prev [-] | | "We were going to implement a counter but hit out weekly Claude code limit before we could do it. Maybe next week? Anthropic." |
|
|
| |
| ▲ | hinkley 4 days ago | parent | prev | next [-] | | > it’s an important enough resource that my lizard brain wants to hoard it. I have zero doubt that this is working exactly as intended. We will keep all our users at 80% of what we sold them by keeping them anxious about how close they are to the limit. | |
| ▲ | oc1 5 days ago | parent | prev | next [-] | | They know this psychology. This dark pattern is intentional so you will use their costly service less. | | |
| ▲ | hn_throwaway_99 5 days ago | parent [-] | | I don't think this counts as a "dark pattern". The reality is that these services are resource constrained, so they are trying to build in resource limits that are as fair as possible and prevent people from gaming the system. | | |
| ▲ | const_cast 4 days ago | parent | next [-] | | The dark pattern isn't the payment pattern, that's fine. The dark pattern is hiding how much you're using, thereby tricking the human lizard brain into irrationally fearing they are running out. The human brain is stupid and remarkably exploitable. Just a teensy little bit of information hiding can illicit strange and self-destructive behavior from people. You aren't cut off until you're cut off, then it's over completely. That's scary, because there's no recourse. So people are going to try to avoid that as much as possible. Since they don't know how much they're using, they're naturally going to err on the side of caution - paying for more than they need. | | |
| ▲ | gorbypark 4 days ago | parent [-] | | I'm only on the $20 Pro plan, and I'm a big users of the /clear command. I don't really use Claude Code that much either, so the $20 plan is perfect for me. However, a few times I've gotten the "approaching context being full, auto compact coming soon" thing, so I manually do /compact and I run out of the 5hr usage window while compacting the context. It's extremely infuriating because if I could have a view into how close I was to being rate limited in the 5 hour window, I might make a different choice as to compact or finish the last little thing I was working on. |
| |
| ▲ | aspenmayer 4 days ago | parent | prev | next [-] | | > prevent people from gaming the system If I sit down for dinner at an all-you-can-eat buffet, I get to decide how much I’m having for dinner. I don’t mind if they don’t let me take leftovers, as it is already understood that they mean as much as I can eat in one sitting. If they don’t want folks to take advantage of an advertised offer, then they should change their sales pitch. It’s explicitly not gaming any system to use what you’re paying for in full. That’s your right and privilege as that’s the bill of goods you bought and were sold. | | |
| ▲ | Wowfunhappy 4 days ago | parent [-] | | I feel like using Claude Code overnight while you sleep or sharing your account with someone else is equivalent to taking home leftovers from an all-you-can-eat buffet. I also find it hard to believe 5% of customers are doing that, though. | | |
| ▲ | aspenmayer 4 days ago | parent | next [-] | | If that’s off-peak time, I’d argue the adjacent opposite point, that Anthropic et al could implement deferred and/or scheduled jobs natively so that folks can do what they’re going to do anyway in a way that comports with reasonable load management that all vendors must do. For example, I don’t mind that Netflix pauses playback after playing continuously for a few episodes of a show, because the options they present me with acknowledge different use cases. The options are: stop playing, play now and ask me again later, and play now and don’t ask me again. These options are kind to the user because they don’t disable the power user option. | | |
| ▲ | gorbypark 4 days ago | parent [-] | | Is there really an off peak time, though? I think Anthropic is running on AWS with the big investment from Amazon, right? I'm sure there's some peaks and valleys but with the Americas, Europe and Asia being in different time zones I'd expect there'd be a somewhat "baseline" usage with peaks where the timezones overlap (European afternoons and American mornings, for example). I know in my case I get the most 503 overloaded errors in the European afternoon. |
| |
| ▲ | closewith 4 days ago | parent | prev | next [-] | | I use Claude Code with Opus four days a week for about 5 hours a day. I've only once hit the limit. Yet the tool others mentioned here (ccusage) indicates I used about $120 in API equivalents per day or about $1,800 to date this month on a $200 subscription. That has to be a loss leader for Anthropic that they now want to wind back. I also wouldn't consider my usage extreme. I never use more than one instance, don't run overnight, etc. | |
| ▲ | kelnos 4 days ago | parent | prev | next [-] | | I think this is just a bad analogy. I've definitely set Claude Code on a task and then wandered off to do something else, and come back an hour or so later to see if it's done. If I'd chosen to take a nap, would you say I'm "gaming the system"? That's silly. I'm using an LLM agent to free up my own time; it's up to me to decide what I do with that time. | | |
| ▲ | Wowfunhappy 4 days ago | parent [-] | | No, this doesn't sound like gaming the system to me. However, if you were using a script to automatically queue up tasks so they can run as soon as your 5-hour-session expires to ensure you're using Claude 24/7, that's a different story. A project like this was posted to HN relatively recently. As I said, I have trouble believing this constitutes 5% of users, but it constitutes something and yeah, I feel Anthropic is justified in putting a cap on that. | | |
| ▲ | yunohn 4 days ago | parent [-] | | They always have had a “soft” sessions limit per month anyway, so it still doesn’t make sense to limit weekly. |
|
| |
| ▲ | benterix 4 days ago | parent | prev [-] | | I use Claude Code overnight almost exclusively, it's simply not worth my time during the day. It's just easier to prepare precise instructions, let it run and check the results in the morning. If it goes awry (it usually does), I can modify the instructions and start from scratch, without getting too attached to it. |
|
| |
| ▲ | hshdhdhj4444 4 days ago | parent | prev | next [-] | | Besides the click mazes to unsubscribe I’m struggling to think of a darker pattern than having usage limits but not showing usage. The dark pattern isn’t the usage limit. It’s the lack of information about current and remaining usage. | |
| ▲ | Timwi 4 days ago | parent | prev [-] | | The dark pattern is not telling users how much they've used so they can't plan or ration. |
|
| |
| ▲ | sitkack 5 days ago | parent | prev | next [-] | | Working as Intended. | | |
| ▲ | Wowfunhappy 5 days ago | parent [-] | | Well, kind of. If you don't use it at all you're going to unsubscribe. This isn't like a gym membership where people join aspirationally. No one's new year's resolution is "I'm going to use o3 more often." | | |
| ▲ | mattigames 5 days ago | parent | next [-] | | Yes it is, in the way of "I'm gonna work on X thing that is now much easier thanks to chatGPT" and then never work on it due lack of time or motivation or something else. | |
| ▲ | christina97 5 days ago | parent | prev [-] | | What makes you think it’s any different? |
|
| |
| ▲ | gfiorav 5 days ago | parent | prev | next [-] | | I nervously hover over the VSCode Copilot icon, watching the premium requests slowly accumulate. It’s not an enjoyable experience (whether you know how much you've used or not :) ) | | |
| ▲ | benjiro 4 days ago | parent [-] | | Noticed that my productive usage of CoPilot dropped like a brick, after they introduced those limits. You feel constantly on the clock, and being forced to constantly change models gets tiresome very fast. Unless you use "free" GPT 4.1 like MS wants you (not the same as Claude, even with Beast Mode). And how long is that going to be free, because it feels like a design to simply push you to a MS product (MS>OpenAI) instead of third party. So what happens a year from now? Paid GPT 5.1? With 4.1 being removed? If it was not for the insane prices of actual large mem GPUs and the slowness of large models, i will be using LLMs at home. Right now MS/Antropic/OpenAI are right in that zone where its not too expensive yet to go full local LLM. |
| |
| ▲ | milankragujevic 5 days ago | parent | prev | next [-] | | Where did you find this info? I am unable to find in on OpenAI's website. https://help.openai.com/en/articles/6950777-what-is-chatgpt-... I haven't yet run into this limit... | | | |
| ▲ | wiseowise 4 days ago | parent | prev | next [-] | | > There’s no way to see how much I’ve used Hover on it on a desktop, it’ll show how many requests you have left. | |
| ▲ | littlestymaar 5 days ago | parent | prev | next [-] | | > This is how I feel about the 100 msg/wk limit on o3 for the ChatGPT Do I read this correctly? Only 100 messages per week, on the pro plan worth a few hundred buck a month?! | | |
| ▲ | CSMastermind 5 days ago | parent | next [-] | | That's definitely not correct because I'm on the pro plan and make extensive use of o3-pro for coding. I've sent 100 messages in a single day with no limitation. Per their website: https://help.openai.com/en/articles/9793128-what-is-chatgpt-... There are no usage caps on pro users (subject to some common sense terms of use). | |
| ▲ | mhl47 5 days ago | parent | prev | next [-] | | No it's 100 a week for plus users. | | | |
| ▲ | doorhammer 5 days ago | parent | prev [-] | | I think it’s just a mistype I have a pro plan and I hammer o3–I’d guess more than a hundred a day sometimes—and have never run into limits personally Wouldn’t shock me if something like that happened but haven’t seen evidence of it yet |
| |
| ▲ | _giorgio_ 4 days ago | parent | prev | next [-] | | Not sure. But o3 seems to be 200/10 days, not weekly anymore in my opinion. | |
| ▲ | artursapek 5 days ago | parent | prev | next [-] | | Just curious, what do people use these expensive reasoning models for? | | | |
| ▲ | clownpenis_fart 4 days ago | parent | prev | next [-] | | [dead] | |
| ▲ | jstummbillig 5 days ago | parent | prev [-] | | If it behaves anything like the GPT-4.5 Limit, it will let you know when you near the limit. | | |
|
|
| ▲ | forty 5 days ago | parent | prev | next [-] |
| It makes me sad that devs start relying on proprietary online services to be able to work. We have been lucky enough to have FOSS tools to do everything and not to have to rely on any specific service or company to work and some of us are deciding to become like Monsanto-addicted farmers who forgot how to do their jobs without something they have to pay every month. |
| |
| ▲ | brookst 4 days ago | parent | next [-] | | I remember being beholden to commercial (cough, pirated, cough) compilers and assemblers back in the day. FOSS is awesome but often lags because capital sees a chance to make money and can move faster. It will change. There will be FOSS models, once it no longer takes hundreds of millions of dollars to train them. | |
| ▲ | pythonguython 4 days ago | parent | prev | next [-] | | Do you mind sharing what industry you’re in where you can fully rely on FOSS? In my industry we’re dependent on MATLAB, Xilinx tools, closed source embedded software and more. To name a few industries: game devs might be stuck with unity, finance quant devs might be stuck with Bloomberg terminals, iOS app devs are stuck with apple’s tooling etc… this isn’t just an LLM problem IMO. | | |
| ▲ | ozgung 4 days ago | parent | next [-] | | I think this is different. Yes, 3rd party tools and services are nothing new. Depending on 3rd party libraries is also a standard thing, although minimum dependency is generally considered a good practice. But all these services provide something you don't want to do yourself and are willing to pay. They all complement what you do, and don't replace your core competency. Apple is your business partner, doing marketing and distribution for you, and shares its user base. Bloomberg terminals provide real time data and UI to non-technical finance people. Github provides you Git hosting service so you don't need to setup and maintain servers. MATLAB (although there are Octave, Python and open alternatives) sells numerical computation environment to non-CS engineers. Xilinx is sells its hardware and dev tools. Game devs use Unity because they want to focus on gameplay and not game engine development. These are all the examples of Division of Labor. This time, however, you have to pay for your core competency, because you cannot compete with a good AI coder in the long run. The value you provide diminishes to almost nothing. Yes you can write prompts, but anyone, even a mediocre LLM can write prompts these days. If you need some software, you don't need to hire SW engineers anymore. A handful of vendors dominate the SW development market. Yes, you can switch. But only between the 3 or 4 tech giants. It's an Oligopoly. If we have FOSS alternatives, at least we can build new services around them and can move on to this new era. We can adapt. Otherwise, we become a human frontend between the client and the AI giants. | |
| ▲ | forty 4 days ago | parent | prev | next [-] | | That's a good remark. As my sibling said, backend and web dev. But indeed it always struck me that some developpers decided to become Apple developpers and sacrifice 30% of everything they ever produce to Apple. I would argue that it might a bit different though, because when doing iOS development it's possible that you don't lose you core skill, which is building software, and that you can switch to another platform with relative ease. What I think might happen with LLM is that people will lose the core skill (maybe not for the generation who did do LLM-less development, but some devs might eventually not ever know other ways to work, and will become digital vassals of whatever service managed to kill all others) | | |
| ▲ | yunohn 4 days ago | parent | next [-] | | Android also charges a fee for Play Store distribution, as do most platforms - not just mobile app ones. The cost of doing business on a platform does not mean no one should do it - like literally all other businesses. | |
| ▲ | unsupp0rted 4 days ago | parent | prev [-] | | > and sacrifice 30% of everything they ever produce to Apple In exchange for 500% more paid users | | |
| ▲ | forty 4 days ago | parent [-] | | The users would exist even if Apple did not exist wouldn't they? | | |
| ▲ | unsupp0rted 3 days ago | parent [-] | | No, they wouldn’t. Apple largely makes paying users exist that otherwise wouldn’t be there. | | |
|
|
| |
| ▲ | aprdm 4 days ago | parent | prev [-] | | backend web services
devops
frontend javascript just three possible examples | | |
| ▲ | forty 4 days ago | parent [-] | | Yes. Web is more and more debatable as as time goes though, there is more and more only one company providing the software to access the web :) | | |
|
| |
| ▲ | epolanski 4 days ago | parent | prev | next [-] | | It makes me sad that couriers start relying on third party wheeled machines like cars and motorbikes. | | |
| ▲ | conradfr 4 days ago | parent | next [-] | | Do their motorbikes stop working in the middle of the week if they make too many deliveries or ask them to pay multiple the base rate for the remaining time? | |
| ▲ | hvb2 4 days ago | parent | prev [-] | | You ignore the fact that there are many suppliers there with plenty of competition. The 'take me from A to N' is a pretty broad problem that can have many different solutions. Is that comparable? We can all see this end up in a oligopoly, no? |
| |
| ▲ | arvinsim 4 days ago | parent | prev | next [-] | | I think it's less about reliance and more about competition. Sure, devs can still work without AI. But if the developer who uses AI has more output than the one that doesn't, it naturally incentives everyone to leverage AI more and more. | | |
| ▲ | forty 4 days ago | parent [-] | | That's a big if. People might feel faster because there is more "movement", but it's not clear if overall they are actually signicantly faster (though everyone would like us to believe so). And note that I objected online services, local LLM don't have the same issues. | | |
| ▲ | epolanski 4 days ago | parent [-] | | People who use it extensively may know whether it makes them significantly faster or not. | | |
| ▲ | forty 4 days ago | parent | next [-] | | I assume they have an opinion on the topic, but it doesn't mean they are right (or wrong). Think of driving a car. If the shortest path (in term to time of travel) is through traffic jam, and there is a longer path where you can drive must faster, it's very likely that most people will have the feeling to be more efficient with the longer path. Also the slow down of using LLM might be more subtle and harder to measure. They might happen at code review time, handling more bugs and incident, harder maintainance, recovering your deleted DB ;)... | | |
| ▲ | epolanski 4 days ago | parent | next [-] | | Apologies, but from antirez[1] to many other brilliant 1000x developers advocate for LLMs speeding up the process. I can see the impact on my own input both in quantity and quality (LLMs can come up with ideas I would not come up to, and are very useful for tinkering and quickly testing different solutions). As any tool it is up to the user to make the best out of it and understand the limits. At this point it is clear that naysayers: 1) either don't understand our job 2) or haven't given AI tools the proper stress testing in different conditions 3) or are luddites being defensive about the "old" world [1] https://www.antirez.com/news/154 | | |
| ▲ | forty 4 days ago | parent | next [-] | | From your source ``` The fundamental requirement for the LLM to be used is: don’t use agents or things like editor with integrated coding agents. You want to: * Always show things to the most able model, the frontier LLM itself. * Avoid any RAG that will show only part of the code / context to the LLM. This destroys LLMs performance. You must be in control of what the LLM can see when providing a reply. * Always be part of the loop by moving code by hand from your terminal to the LLM web interface: this guarantees that you follow every process. You are still the coder, but augmented. ``` Not sure about you, but I think this process, which your source seems to present as a prerequisites to use LLM efficiently (and seems good advice to me too, and actually very similar of how I use LLM myself) must be followed by less than 1% of LLM users. | |
| ▲ | forty 4 days ago | parent | prev | next [-] | | I wish I had only antirezs working on my projects, and would for sure be much more confident that some significant time might be saved with llms if that was the case. | |
| ▲ | zelphirkalt 4 days ago | parent | prev [-] | | 1000x developers, ahahaha! Come on now, this is too comical. Even 10x is extremely rare. The deciding factor is not speed. It is knowledge. Will I be able to dish out a great compiler in a week? Probably not. But an especially knowledgeable compiler engineer might just do it, for a simple language. Situations like this are the only 10x we have in our profession, if we don't count completely incapable people. The use of AI doesn't make you 1000x. It might make you output an infinite factor of AI slop more, but then you are just pushing the maintenance burden to a later point in time. In total it might make your output completely useless in the long run, making you a 0x dev in the worst case. |
| |
| ▲ | eichin 4 days ago | parent | prev [-] | | We've known for decades that self-reported time perception in computer interactions is drastically off (Jef Raskin, The Humane Interface in particular) so unless they have some specifically designed external observations, they are more likely to be wrong. (There have been more recent studies - discussed here on HN - about perception wrt chat interfaces for code specifically - that confirm the effect on modern tools.) |
| |
| ▲ | hvb2 4 days ago | parent | prev [-] | | There's 2 problems there
1. 'faster' is subjective since you cannot do the same task twice without the second time being biased by the learnings from the first pass
2. While speed might be one measure, I've rarely found speed to be the end goal. Unless you're writing something that's throw away, you'll be reading what was written many times over. Long term maintainability is at odds with speed, in most cases. | | |
| ▲ | epolanski 4 days ago | parent [-] | | You're implying that LLMs make maintainability worst when the opposite could happen if you know how to use the tools. | | |
| ▲ | zelphirkalt 4 days ago | parent [-] | | But the tools are trained on tons and tons of mediocre work and will have a strong tendency to output such. Please share your prompts aimed at preventing mediocre code entering the code bases you work on. So far almost no code I got from LLMs was acceptable to stay as suggested. I found it useful in cases, when I myself didn't know what a typical (!) way is to do things with some framework, but even then often opted for another way, depending on my project's goals and design. Sometimes useful to get you unstuck, but oh boy I wouldn't let it code for me. Then I would have to review so much bad code, it would be very frustrating. |
|
|
|
|
| |
| ▲ | stingraycharles 4 days ago | parent | prev | next [-] | | Devs have been doing this for years. If Github goes down, it has a huge impact on dev productivity. | | |
| ▲ | forty 4 days ago | parent [-] | | Has it? We don't use GitHub in my company, but our self hosted Gitlab occasionally goes down, and while it prevents us from merging (including doing code reviews) and deploying code, it does not prevent us from doing most of our work (ie designing and creating software). It merely delays the moment when things are shipped. If you meant goes down for good, then I'm sure it would be annoying for a few weeks for the FOSS ecosystem, just the time to migrate elsewhere, but there is not much GitHub specific we would really miss. |
| |
| ▲ | danielbln 5 days ago | parent | prev | next [-] | | The very recent agentic open weights models seem to be shaping up, so if all fails you can host one of these yourself (if you have the vram) or host it yourself somewhere. | |
| ▲ | sneak 5 days ago | parent | prev | next [-] | | We can and did work without it. It just makes us many times faster. Nothing about using an LLM removes skills and abilities you already had before it. | | |
| ▲ | 14123newsletter 4 days ago | parent | next [-] | | >Nothing about using an LLM removes skills and abilities you already had before it. Funny story: The widespread of Knorr soup stock already made people unable to cook their own stock soup, or even worse, the skill to season their soup from just basic, fresh ingredients. Source: my mom. | | |
| ▲ | windexh8er 4 days ago | parent [-] | | I'm always surprised when people buy vegetable stock. So many people I know cook "from scratch" with base ingredients like stock out of a box. And just as with cooking: most people won't care - and the same goes with LLMs. It can be good enough... Less efficient? Meh - cloud. AI slop image? Meh - cheaper than paying an artist. LLMs to get kids through school? Meh - something something school-of-life. I look around and see many poorly educated people leaning hard into LLMs. These people are confusing parroting their prompt output as knowledge, especially in the education realm. And while LLMs may not "remove skills and abilities you already had before it" - you damn sure will lose any edge you had over time. It's a slippery slope of trading a honed skill for convenience. And in some cases that may be a worthwhile trade. In others that is a disaster waiting to happen. |
| |
| ▲ | 3836293648 5 days ago | parent | prev | next [-] | | For the industry as a whole it absolutely does. And for the individual it absolutely does kill your ability to do it unless you actually do practice. And yes, the goal might be to only use it for boilerplate or first draft. But that's today, people are lazy, just wait for the you of tomorrow | | |
| ▲ | epolanski 4 days ago | parent [-] | | > And for the individual it absolutely does kill your ability to do it unless you actually do practice. Just because you state it, it doesn't make it true. I could tell you that taking buses or robotaxis doesn't change a bit your ability to drive. | | |
| ▲ | forty 4 days ago | parent | next [-] | | Depends how often you do drive. I can guarantee you that not driving absolutely affected my ability to drive (I can still drive but certainly not nearly as well as if I drove daily) | |
| ▲ | 3836293648 4 days ago | parent | prev [-] | | Are you seriously claiming that not driving has no affect on your ability to drive? Taking a bus sometimes is fine, but that is missing literally the entire point that I made. |
|
| |
| ▲ | umbra07 4 days ago | parent | prev | next [-] | | If you don't use a skill, it atrophies. Now, maybe that is the future (no more/extremely little human-written code). Maybe that's a good thing in the same way that "x technological advancement means y skill is no longer necessary" - like how the advent of readily-accessible live maps means you don't need to memorize street intersections and directions or whatever. But it is true. | | |
| ▲ | brookst 4 days ago | parent [-] | | I am terrible at computing sine and cosine, for sure. It doesn’t bother me. | | |
| ▲ | zelphirkalt 4 days ago | parent [-] | | On the surface, this comparison might hold, but when you look at software development as a craft, and therefore containing aspects of creativity and art, the comparison no longer holds. | | |
| ▲ | brookst 4 days ago | parent [-] | | Interesting take. So AI is the Ikea-ization of software, producing far cheaper / lower quality / less durable / more accessible product that is completely good enough for most people, but unacceptable to those who have the expertise to do it themselves, or the wealth to not care about price? |
|
|
| |
| ▲ | hoppp 4 days ago | parent | prev [-] | | My experience was that reviewing generated code can take longer than writing it from scratch. There was research about vibe coding that had similar conclusion. Feels productive but can take longer to review. the moment you generate code you don't instantly understand you are better off reading the docs and writing it yourself |
| |
| ▲ | yunohn 4 days ago | parent | prev [-] | | Huh? We’re discussing the equivalent of tradespeople who need a high quality drill to work instead of cheap knockoffs or worse, no tools. |
|
|
| ▲ | blitzar 5 days ago | parent | prev | next [-] |
| > Anthropic is saying that one out of every 20 users will hit the new limit. I regularly hit the the Pro limits 3 times a day using sonnet. If I use claude code & claude its over in about 30 minutes. No multi 24/7 agent whatever, no multiple windows open (except using Claude to write a letter between claude code thoughts). I highly doubt I am a top 5%er - but wont be shocked if my week ends on a wednessday. I was just starting to use Claude chat more as it is in my subscription but if I can not rely on it to be available for multiple days its functionally useless - I wont even bother. |
| |
| ▲ | Aurornis 5 days ago | parent | next [-] | | > If I use claude code & claude its over in about 30 minutes. Can you share what you're doing? I've been experimenting with Claude Code and I feel like I have to be doing a lot with it before I even start seeing the usage warning limits on the $20/month plan. When I see people claiming they're getting rate limited after 30 minutes on the $100/month plan I have a hard time understanding what they're doing so different. For what it's worth I don't use it every day, so maybe there's a separate rate that applies to heavy and frequent users? | | |
| ▲ | gorbypark 4 days ago | parent | next [-] | | As a $20 month user, I can tell you in my experience it's "refactoring" jobs that really smash through those tokens quickly. If you do a "write a component that does this" kinda thing, you can use the $20 plan almost an unlimited amount of time. If you are doing "find all instances of ComponentFoo, change to ComponentBar, refactor each screen for correct usage of ComponentBar" kinda things, it's going to grep through your code, find multiple files, read all of them into context and start making changes one by one and/or spin up a subagent to do it. You'll be rate limited pretty quick doing things that way. | | |
| ▲ | mike_hearn 4 days ago | parent | next [-] | | Are people really using LLMs this way? Can't your IDE at least attempt global refactorings? I get that it's harder with dynamic languages like Python but using an LLM to rename a class seems like using a nuke to crack a nut. | | |
| ▲ | gorbypark 2 days ago | parent [-] | | For sure, it was a bit of a simplified example, but you can imagine prompts that are asking the LLM to touch dozens of files at once. A lot of people will use fairly high level requests, "make a screen in my app that does this" and that ends up making screens, components, queries, local state changes, etc etc. It adds up quick having so much in the context at once (and if you /clear the context and ask it to do anything more than one off changes in a file, it's gonna grep and glob it's way through your code looking for the right files) |
| |
| ▲ | nkassis 4 days ago | parent | prev | next [-] | | That kind of gives a point in favor of AI enabled IDEs, these kind of operations could often be achieved with simple refactoring tools available without AI today and save massive amount of trees. | |
| ▲ | gaws 2 days ago | parent | prev [-] | | > If you are doing "find all instances of ComponentFoo, change to ComponentBar, refactor each screen for correct usage of ComponentBar" kinda things, it's going to grep through your code, find multiple files, read all of them into context and start making changes one by one and/or spin up a subagent to do it. You'll be rate limited pretty quick doing things that way. Huh?? grep and sed do this for free; you don't need A.I. for that. |
| |
| ▲ | flutas 5 days ago | parent | prev [-] | | $20/mo plan doesn't include opus (the larger model) like the $100+ plans do, it's likely they are hitting the opus limit which is fairly low. | | |
| ▲ | r053bud 4 days ago | parent | next [-] | | The $20/month plan most definitely does include Opus. Just not a ton of usage allowed. | | | |
| ▲ | flowerthoughts 4 days ago | parent | prev [-] | | I'm on the $20 and was hitting the limit quite often. It lasted 2.5 h out of 5 h if I went all in. So even before this questionable change, we were at 50% utilization at most. And I guess it'll go downhill from here. Anthropic, I wish you the best. Claude is a great tool at good value. But if you keep changing the product after my purchase, that's bad value. |
|
| |
| ▲ | bogtog 5 days ago | parent | prev | next [-] | | > I highly doubt I am a top 5%er - but wont be shocked if my week ends on a wednessday. I was just starting to use Claude chat more as it is in my subscription but if I can not rely on it to be available for multiple days its functionally useless - I wont even bother. You very well might be a top 5%er among people only on the Pro rather than Max plan | |
| ▲ | ketzo 5 days ago | parent | prev | next [-] | | What does your Claude code usage look like if you’re getting limited in 30 minutes without running multiple instances? Massive codebase or something? | | |
| ▲ | blitzar 5 days ago | parent [-] | | I set claude about writing docstrings on a handful of files - 4/5 files couple 100 lines each - couple of classes in each - it didnt need to scan codebase (much). Low danger task so I let it do as it pleased - 30 minutes and was maxed out. Could probably have reduced context with a /clear after every file but then I would have to participate. | | |
| ▲ | tlbsofware 5 days ago | parent | next [-] | | You can tell it to review and edit each file within a Task/subagent and can even say to run them in parallel and it will use a separate context for each file without having to clear them manually | | |
| ▲ | blitzar 5 days ago | parent [-] | | Every day is a school day - I feel like this is a quicker way to burn usage but it does manage context nicely. | | |
| ▲ | tlbsofware 5 days ago | parent [-] | | I haven’t ran any experiments about token usage with tasks, but if you ran them all together without tasks, then each files full operation _should_ be contributing as cached tokens for each subsequent request. But if you use a task then only the summary returned from that task would contribute to the cached tokens. From my understanding it actually might save you usage rates (depending on what else it’s doing within the task itself). I usually use Tasks for running tests, code generation, summarizing code flows, and performing web searches on docs and summarizing the necessary parts I need for later operations. Running them in parallel is nice if you want to document code flows and have each task focus on a higher level grouping, that way each task is hyper focused on its own domain and they all run together so you don’t have to wait as long, for example: - “Feature A’s configuration”
- “Feature A’s access control”
- “Feature A’s invoicing” |
|
| |
| ▲ | stuaxo 5 days ago | parent | prev | next [-] | | I hope you thoroughly go through these as a human, purely AI written stuff can be horrible to read. | | |
| ▲ | blitzar 5 days ago | parent [-] | | Docstring slop is better than code slop - anyway that is what git commits are for - and I have 4.5 hours to do that till next reset. | | |
| |
| ▲ | Kurtz79 4 days ago | parent | prev [-] | | If I understand correctly, looking at API pricing for Sonnet, output tokens are 5 times more expensive than input tokens. So, if rate limits are based on an overall token cost, it is likely that one will hit them first if CC reads a few files and writes a lot of text as output (comments/documentation) rather than if it analyzes a large codebase and then makes a few edits in code. |
|
| |
| ▲ | rapind 4 days ago | parent | prev [-] | | I think you'll want to specify your /model to not use opus. Strangely unintuitive, but I opted out of opus on the max plan myself and aren't really having any usage issues since. | | |
| ▲ | blitzar 4 days ago | parent [-] | | no opus in claude code on my el'cheapo plan ($20 - pro) - used it occasionally in claude desktop and its as expensive as they advertise. |
|
|
|
| ▲ | _jab 5 days ago | parent | prev | next [-] |
| > Anthropic is saying that one out of every 20 users will hit the new limit. Very good point, I find it unlikely that 1/20 users is account sharing or running 24/7 agentic workflows. |
| |
| ▲ | Terretta 5 days ago | parent | next [-] | | Moreover, if you run a SaaS, generally somewhere from 1 in 5 to 1 in 20 users are using you for real, while the others are mostly not using you. The stat would be more interesting if instead of 1 in 20 users, they said x in y of users with at least one commit per business day, or with at least one coding question per day, or whatever. I suspect this could be a significantly higher percentage of professional users they plan to throttle. Be careful of defining Pro like Apple does if you market to actual professionals who earn based on using your product. Your DAUs might be a different ratio than you expect. | | |
| ▲ | comebhack 4 days ago | parent | next [-] | | I would probably show up in their metrics as an active user and one of the 95% but I barely use the product. I have a Pro subscription which I use for personal projects but I do very little, maybe using it once a week for a short session. At work I use Cursor via a corporate account. I imagine there are lots of people like me who have a subscription to be aware of the product and do some very light work, but the "real" users who rely on the tool might be badly affected by this. | |
| ▲ | 0cf8612b2e1e 5 days ago | parent | prev [-] | | … if you run a SaaS, generally somewhere from 1 in 5 to 1 in 20 users are using you for real, while the others are mostly not using you.
That is a hilarious and believable stat. Has anyone published such numbers or is it a dirty secret about how many corporate licenses are purchased and never used by the rank and file?I can personally think of a few internally licensed products, announced with huge fan fare, which never get used beyond the demo to a VP. | | |
| |
| ▲ | rapind 4 days ago | parent | prev [-] | | A decent chunk (more than 1/20) account shared netflix. Also there are probably some who are account sharing with more than one other person. I don't really doubt it. | | |
| ▲ | yifanl 4 days ago | parent [-] | | Sharing Netflix passwords was literally a common enough phonomena that it purt of Netflix's advertising, that's a very unique case that I don't expect to be happening for Anthropic. |
|
|
|
| ▲ | arghwhat 5 days ago | parent | prev | next [-] |
| > but I'd feel the same about the prospect of loosing ripgrep for a week, or whatever. Loosing it for a couple of days is more palatable. Just to nitpick: When the limit is a week, going over it does not mean losing access for a week, but for the remaining time which would assuming the limits aren't overly aggressive mean losing access for at most a couple of days (which you say is more palatable). I wouldn't say you're too reliant, but it's still good to stay sharp by coding manually every once in a while. |
|
| ▲ | furyofantares 5 days ago | parent | prev | next [-] |
| > I'm probably not going to hit the weekly limit, but it makes me nervous that the limit is weekly as opposed to every 36 hours or something. If I do hit the limit, that's it for the entire week—a long time to be without a tool I've grown accustomed to! Well, not the entire week, however much of it is left. You said you probably won't hit it -- if you do, it's very likely to be in the last 36 hours (20% of a week) then, right? And you can pay for API usage anyway if you want. |
|
| ▲ | arach 5 days ago | parent | prev | next [-] |
| if it affects only a minority of accounts, why not figure out how to special case them without affecting everyone else is the primary question I would ask myself if I worked on this the principle: let's protect against outliers without rocking the behavior of the majority, not at this stage of PMF and market discovery i'd also project out just how much the compute would cost for the outlier cohort - are we talking $5M, $100M, $1B per year? And then what behaviors will simply be missed by putting these caps in now - is it worth missing out on success stories coming from elite and creative users? I'm sure this debate was held internally but still... |
| |
| ▲ | vineyardmike 5 days ago | parent | next [-] | | Because the goal is to extract more money from the people who have significant usage. These users are the actual targets of the product. The idea that it’s a few bad actors is misdirection of blame to distract “power users”. They undercharged for this product to collect usage data to build better coding agents in the future. It was a ploy for data. Anecdotally, I use Claude Code with the $20/mo subscription. I just use it for personal projects, so I figured $20 was my limit on what I’d be willing to spend to play around with it. I historically hit my limits just a few times, after ~4hrs of usage (resets every 5hrs). They recently updated the system and I hit my limits consistently within an hour or two. I’m guessing this weekly limit will affect me. I found a CLI tool (which I found in this thread today) that estimates I’m using ~$150/mo in usage if I paid through the API. Obviously this is very different from my payments. If this was a professional tool, maybe I’d pay, but not as a hobbyist. | | | |
| ▲ | Uehreka 5 days ago | parent | prev | next [-] | | > why not figure out how to special case them without affecting everyone else I’m guessing that they did, and that that’s what this policy is. If you’re talking about detecting account sharing/reselling, I’m guessing they have some heuristics, but they really don’t want the bad press from falsely accusing people of that stuff. | | |
| ▲ | arach 5 days ago | parent [-] | | fair enough - DS probably ran through data and came up with 5% and some weekly cutoff as a good starting point until they have better measures in place my point is that 5% still a large cohort and they happen to be your most excited/creative cohort. they might not all want to pay a surchage yet while everyone is discovering the use cases / patterns / etc having said that, entirely possible burn rate math and urgency requires this approach |
| |
| ▲ | data-ottawa 5 days ago | parent | prev | next [-] | | They did have several outages last week, it would be good to find better plans for those huge users but I can also see them wanting to just stop the bleeding. | | |
| ▲ | arach 5 days ago | parent [-] | | I've noticed the frequent perf issues and I'm on the 20x plan myself - good point that you'd want to stop the bleeding or bad actors to make sure the majority have a better experience |
| |
| ▲ | Aurornis 5 days ago | parent | prev | next [-] | | > why not figure out how to special case them without affecting everyone else is the primary question I would ask myself if I worked on this The announcement says that using historical data less than 5% of users would even be impacted. That seems kind of clear: The majority of users will never notice. | | |
| ▲ | arach 5 days ago | parent [-] | | 5% of a large number is a large number - this why it's both a significant problem for them and why I'm thinking out loud about the downsides of discouraging good actors who happen to be power users. that 5% is probably the most creative and excited cohort. obviously it's critical to not make the experience terrible for the 95% core, but i'd hate to lose even a minority of the power users who want to build incredible things on the platform having said that, the team is elite, sure they are thinking about all angles of this issue | | |
| ▲ | 0cf8612b2e1e 5 days ago | parent [-] | | 5% seems like a huge number of previously ecstatic customers who may suddenly be angry. Especially when it is trivial to identify the top 0.1% of users who are doing something insane. |
|
| |
| ▲ | bananapub 5 days ago | parent | prev | next [-] | | > if it affects only a minority of accounts, why not figure out how to special case them without affecting everyone else that's exactly what they have done - the minority of accounts that consume many standard deviations above the mean of resources will be limited, everyone else will be unaffected. | | |
| ▲ | arach 5 days ago | parent [-] | | "You're absolutely right!" i misread the announcement - thought everyone moved to primarily a weekly window but seems like 5hr window still in place and they're putting in place another granularity level control that DS teams will adjust to cutoff mostly bad actors. correct me if I'm wrong, it's not like we have visibility into the token limit logic, even on the 5hr window? |
| |
| ▲ | nharada 5 days ago | parent | prev [-] | | What do you think they should have done instead? | | |
| ▲ | actsasbuffoon 5 days ago | parent | next [-] | | At a bare minimum there needs to be some way to understand how close you are to these limits. People shouldn’t be wondering if this is going to impact them or not. | |
| ▲ | arach 5 days ago | parent | prev [-] | | It’s tricky without seeing the actual data. 5% of a massive user base can still be a huge number so I get that it’s hard to be surgical. But those power users are often your most creative, most productive, and most likely to generate standout use cases or case studies. Unless they’re outright abusing the system, I’d lean toward designing for them, not against them. if the concern is genuine abuse, that feels like something you handle with escalation protocols: flag unusual usage, notify users, and apply adaptive caps if needed. Blanket restrictions risk penalizing your most valuable contributors before you’ve even discovered what they might build | | |
| ▲ | smileysteve 5 days ago | parent [-] | | 5% of a massive user base could also be huge if 50% of users are on an enterprise plan and barely using it. | | |
| ▲ | arach 5 days ago | parent [-] | | in other words, these limits will help introduce Enterprise (premium) plans? |
|
|
|
|
|
| ▲ | thephyber 4 days ago | parent | prev | next [-] |
| (1) I use Claude Code as a solo iOS developer and don’t hit the 5 hour limits much. I suspect the users hitting it more often are throwing way more tokens from much larger code repos and are probably asking for larger incremental changes. (2) I interpret this change as targeting people who are abusing the single Pro account, but using it more like a multi-developer business would maximizing the number of tokens (multiple sessions running 24/7 always hitting the limits). Anthropic has a business interest in pushing those users to use the API (paying per token) or upgrade to the $200/mo subscription. (3) While I fear they might regularly continue to push the top x% usage tier users into the higher subscription rate, I also realize this is the first adjustment for token rates of Claude Pro since Claude Code became available on that subscription. (4) If you don’t want to wait for the next unthrottling, you can always switch to the API usage and pay per token until you are unblocked. |
|
| ▲ | jonas21 5 days ago | parent | prev | next [-] |
| You can use an API key to get more usage on a pay-as-you-go basis. |
| |
| ▲ | blitzar 5 days ago | parent | next [-] | | You can set cash on fire if you want to. | | |
| ▲ | thejazzman 5 days ago | parent | next [-] | | i've gotten months of usage out of openai and claude where i seeded each with only $5 but if you use an agent and it tries to include a 500kb json file, yeah, you will light cash on fire (this happened to me today but the rate limit bright it to my attention.) | | |
| ▲ | blitzar 5 days ago | parent [-] | | Fortunately Ai companies are not (currently) like AWS - you can only burn as much money as you have on account. For brief, low context interactions, it is crazy how far your money goes. |
| |
| ▲ | fullstackwife 5 days ago | parent | prev [-] | | you spend 400$ per month on api usage, but your AI builds the next unicorn worth billions, where is the problem? | | |
| ▲ | blitzar 5 days ago | parent | next [-] | | For $2 you can have a lottery ticket that will win you a quater of a billion dollars. | |
| ▲ | edude03 5 days ago | parent | prev [-] | | You might have better odds buying scratchers - assuming you even have $400/m to invest in an enterprise without cash flow |
|
| |
| ▲ | Wowfunhappy 5 days ago | parent | prev | next [-] | | Yes, but that's so expensive I will never do it! | | |
| ▲ | tqwhite 5 days ago | parent [-] | | With one months exception, I've never gotten past $150 with API. I plan to do the $100 plan and use the API for overflow. I think I will come out ahead. | | |
| ▲ | Wowfunhappy 5 days ago | parent [-] | | Well, lucky you! Before Claude Max was an option, I burned a lot of money using Claude Code, and that was while I was trying my best to use it as little as possible. |
|
| |
| ▲ | wahnfrieden 4 days ago | parent | prev [-] | | Just buy several Max tier subscriptions... |
|
|
| ▲ | gaws 2 days ago | parent | prev | next [-] |
| > If I do hit the limit, that's it for the entire week—a long time to be without a tool I've grown accustomed to! These limits will only get worse. Let this be a wake-up call for you to not put all your development eggs in the A.I. basket. |
|
| ▲ | swalsh 5 days ago | parent | prev | next [-] |
| I used about $300 worth of credits based on ccusage ($20 pro plan). It's pretty easy to hit the limit once you get going. |
| |
|
| ▲ | mrits 5 days ago | parent | prev | next [-] |
| I imagine they will add some features soon where you have more control. It could get complicated quickly. Before they put this in I think they should have at least given you an easy way to buy more credits at a hugely discounted rate. I know entire offices in Bangladesh share some of these accounts, so I can see how it is a problem. |
| |
| ▲ | fuzzzerd 5 days ago | parent [-] | | That is exactly the use case they're trying to stop. Sharing accounts. | | |
| ▲ | mrits 5 days ago | parent [-] | | Which is why I said I see how it could be a problem :) |
|
|
|
| ▲ | matt3210 4 days ago | parent | prev | next [-] |
| >If I do hit the limit, that's it for the entire week Now your vibes can be at the beach. |
| |
|
| ▲ | jcims 5 days ago | parent | prev | next [-] |
| Cursor sent me a note yesterday that at my current usage rate I was going to exceed whatever cap was in place at some date in the future. I thought that was very helpful. |
|
| ▲ | Teknomadix 4 days ago | parent | prev | next [-] |
| Speaking of ripgrep, adding `rg`, `fzf` and a few other tools to my Claude config seems to have been helpful in reducing token consumption and improving agents capabilities. I'm curious what other tools people are giving their agents instructions to use and what the experience is like? |
|
| ▲ | j45 5 days ago | parent | prev | next [-] |
| Getting in early on a plan may not have as much of an upside where the computational costs and guaranteed heavily can look different than other services. If it's affecting 5% of users, it might be people who are really pushing it and might not know (hopefully they get a specialized notice that they may see usage differences). |
|
| ▲ | stingraycharles 4 days ago | parent | prev | next [-] |
| Yeah, if this happens, I’m probably going to do the inevitable: get two accounts. If I could pay $500 a month for Claude 100x I would probably do it at this point. Given that I rarely hit the session limits I’m hopeful I won’t be affected, but the complete and utter lack of transparency is really frustrating. |
| |
| ▲ | Aeolun 4 days ago | parent [-] | | Like, I understand why they do it, usage patterns aren’t the same all over the world, and they paid for the GPU’s anyway, so they want to utilize them. People in parts of the world that see less CC utilization get to do more with their CC plan than people in very busy areas of the world. |
|
|
| ▲ | theshrike79 4 days ago | parent | prev | next [-] |
| From what I gathered from their post, to hit the weekly limit you have to hit the daily limit MULTIPLE times during the week. I'm pretty sure they calibrated it so that only the people who max out every 5 hour window consistently get hit by the weekly quota. |
|
| ▲ | lerchmo 4 days ago | parent | prev | next [-] |
| Time to start using Claude code proxy and other models. This black box rate limit is really lame. |
|
| ▲ | tqwhite 5 days ago | parent | prev | next [-] |
| Don't you also have an API subscription to provide overflow capacity? |
|
| ▲ | wahnfrieden 4 days ago | parent | prev | next [-] |
| It is becoming typical to counter this (and the multi-hour limits) simply by purchasing multiple Max tier subscriptions. |
|
| ▲ | mattlangston 4 days ago | parent | prev | next [-] |
| I have both a Claude subscription plan and console credits as my backup, which I thought was a reasonable solution. |
|
| ▲ | jonfw 4 days ago | parent | prev | next [-] |
| GitHub copilot has a monthly rate limit for premium models- much worse! I ran into mine within hours of using Claude |
|
| ▲ | motbus3 4 days ago | parent | prev | next [-] |
| i believe this is intended to combat abuse.
that said, i do not like the idea. as many other services did, and even some tangible products are implementing, the introduced limit will later on be used to create more tiers and charge you more for the same without providing anything extra. #shrinkflation |
| |
| ▲ | oblio 4 days ago | parent [-] | | Shrinkflation usually assumes that the product is profitable. Do we even know the Anthropic financials? My guess is that they're probably losing money on all their tiers. |
|
|
| ▲ | prolly97 4 days ago | parent | prev | next [-] |
| > Loosing it for a couple of days is more palatable Sorry, I'll just be "that guy" for a moment.
Assuming that access is cut at a random time during the week, the average number of days without Claude would be 3.5.
That's not reasonable as it's dependant on usage.
So assume that you've always been just shy of hitting the limit, and you increase usage by 50%, then you'd hit the limit 4.67 days in. Just 2-3 hours shy of the weekend - a sort of reward for the week's increased effort. Have a blessed Thuesday. |
| |
| ▲ | Wowfunhappy 4 days ago | parent [-] | | My original post wasn't clear and that's my fault, but I am taking that into account. When I said "that's it for the week" I meant "the remainder of the week." It's just a mouthful to write "the remainder of the week" each time. And... well, I am worried that I could do something stupid or accidental on Monday and Tuesday and then loose access with more than half of the week left, especially since, as everyone else has noted, Claude doesn't show you how close to the limit you are until you've almost hit it. (That said, I also appreciate that it doesn't warn me too early, or I would be constantly watching the usage tick up well before I actually need to worry about it.) A couple of weeks ago, Claude got into some kind of weird loop where it just kept saying something like "now I need to add the code continuation" over and over and over again. I had auto-accept turned on, and it was chugging for several hours before I realized something was wrong and stopped it. Who knows how much usage that burned! Luckily, it happened late in the evening, so I knew my usage would reset the next day anyway. But IIRC it was also a Monday evening... |
|
|
| ▲ | jey 5 days ago | parent | prev | next [-] |
| Get two subscriptions if it's delivering that much value and you hit the limits? |
|
| ▲ | ekianjo 5 days ago | parent | prev | next [-] |
| the first rule of APIs is do not expect them to work 24/7. and you are never in control of any change that can occur. Thats why its really important to cultivate local LLMs. |
|
| ▲ | tkiolp4 5 days ago | parent | prev | next [-] |
| Don’t worry, just give them more money. |
|
| ▲ | douglaswlance 5 days ago | parent | prev | next [-] |
| pay for another account to double your limit. |
|
| ▲ | TacticalCoder 5 days ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | draxil 5 days ago | parent | prev | next [-] |
| Of course ripgrep runs on your machine and you control it. |
| |
| ▲ | dewey 5 days ago | parent [-] | | ripgrep is an example for "tool I've grown accustomed to", where it runs is irrelevant. |
|
|
| ▲ | sva_ 5 days ago | parent | prev [-] |
| [flagged] |
| |
| ▲ | Wowfunhappy 5 days ago | parent | next [-] | | On a Mac, you can use option-shift-dash to insert an emdash, which is muscle memory for me. If I had used an LLM, maybe I wouldn't have misspelled "losing" not once but twice and not noticed until after the edit window. <_< | |
| ▲ | adastra22 5 days ago | parent | prev | next [-] | | Was it necessary to post this? FYI many input methods (including my own) turn two hyphens into an em dash. The em-dash-means-LLM thing is bogus. | | | |
| ▲ | tomhow 4 days ago | parent | prev | next [-] | | Please don't do this here. If a comment seems unfit for HN, please flag it and email us at hn@ycombinator.com so we can have a look. | |
| ▲ | fredoliveira 5 days ago | parent | prev [-] | | You can't possibly think that using an em dash is exclusive to AI-generated output. |
|