| ▲ | dakiol 4 hours ago |
| We dropped Claude. It's pretty clear this is a race to the bottom, and we don't want a hard dependency on another multi-billion dollar company just to write software We'll be keeping an eye on open models (of which we already make good use of). I think that's the way forward. Actually it would be great if everybody would put more focus on open models, perhaps we can come up with something like the "linux/postgres/git/http/etc" of the LLMs: something we all can benefit from while it not being monopolized by a single billionarie company. Wouldn't it be nice if we don't need to pay for tokens? Paying for infra (servers, electricity) is already expensive enough |
|
| ▲ | ahartmetz 4 hours ago | parent | next [-] |
| >we don't want a hard dependency on another multi-billion dollar company just to write software One of two main reasons why I'm wary of LLMs. The other is fear of skill atrophy. These two problems compound. Skill atrophy is less bad if the replacement for the previous skill does not depend on a potentially less-than-friendly party. |
| |
| ▲ | post-it 4 hours ago | parent | next [-] | | I was worried about skill atrophy. I recently started a new job, and from day 1 I've been using Claude. 90+% of the code I've written has been with Claude. One of the earlier tickets I was given was to update the documentation for one of our pipelines. I used Claude entirely, starting with having it generate a very long and thorough document, then opening up new contexts and getting it to fact check until it stopped finding issues, and then having it cut out anything that was granular/one query away. And then I read what it had produced. It was an experiment to see if I could enter a mature codebase I had zero knowledge of, look at it entirely through an AI, and come to understand it. And it worked! Even though I've only worked on the codebase through Claude, whenever I pick up a ticket nowadays I know what file I'll be editing and how it relates to the rest of the code. If anything, I have a significantly better understanding of the codebase than I would without AI at this point in my onboarding. | | |
| ▲ | estetlinus 3 hours ago | parent | next [-] | | Yeah, +1. I will never be working on unsolved problems anyhow. Skill atrophy is not happening if you stay curious and responsible. | | |
| ▲ | stringfood 3 hours ago | parent | next [-] | | I have never learned so quickly in my entire life than to post a forum thread in its entirety into a extended think LLM and then be allowed to ask free form questions for 2 hours straight if I want to. Having my questions answered NOW is so important for me to learn. Back in the day by the time I found the answer online I forgot the question | | |
| ▲ | lobf 2 hours ago | parent [-] | | Same. I work in the film industry, but I’ve always been interested in computers and have enjoyed tinkering with them since I was about 5. However, coding has always been this insurmountably complicated thing- every time I make an effort to learn, I’m confronted with concepts that are difficult for me to understand and process. I’ve been 90% vibe coding for a year or so now, and I’ve learned so much about networking just from spinning up a bunch of docker containers and helping GPT or Claude fix niggling issues. I essentially have an expert (well, maybe not an expert but an entity far more capable than I am on my own) who’s shoulder I can look over and ask as many questions I want to, and who will explain every step of the process to me if I want. I’m finally able to create things on my computer that I’ve been dreaming about for years. |
| |
| ▲ | idopmstuff 3 hours ago | parent | prev | next [-] | | Some people talk like skill atrophy is inevitable when you use LLMs, which strikes me as pretty absurd given that you are talking about a tool that will answer an infinite number of questions with infinite patience. I usually learn way more by having Claude do a task and then quizzing it about what it did than by figuring out how to do it myself. When I have to figure out how to do the thing, it takes much more time, so when I'm done I have to move on immediately. When Claude does the task in ten minutes I now have several hours I can dedicate entirely to understanding. | | |
| ▲ | onemoresoop 3 hours ago | parent | next [-] | | You lose some, you win some. The win could be short-term much higher, however imagine that the new tool suddenly gets ragged pulled from under your feet. What do you do then? Do you still know how to handle it the old way or do you run into skill atrophy issues? I’m using Claude/Codex as well, but I’m a little worried that the environment we work in will become a lot more bumpy and shifty. | | |
| ▲ | visarga 3 hours ago | parent [-] | | > the new tool suddenly gets ragged pulled from under your feet If that happened at this point, it would be after societal collapse. | | |
| |
| ▲ | hdjrudni an hour ago | parent | prev | next [-] | | The "infinite patience" thing I find particularly interesting. Every now and then I pause before I ask an LLM to undo something it just did or answer something I know it answered already, somewhere. And then I remember oh yeah, it's an LLM, it's not going to get upset. | |
| ▲ | dlopes7 an hour ago | parent | prev | next [-] | | Asking infinite questions about something does not make you good at “doing” that thing, you get pretty good at asking questions | |
| ▲ | techpression an hour ago | parent | prev [-] | | Understanding is not learning. Zero effort gives zero rewards, I ask Claude plenty of things, I get answers but not learnings. |
| |
| ▲ | bdangubic 2 hours ago | parent | prev [-] | | I used to speak Russian like I was born in Russia. I stopped talking Russian … every day I am curious ans responsible but I can hardly say 10 words in Russian today. if you don’t use it (not just be curious and responsible) you will lose it - period. | | |
| ▲ | thih9 2 hours ago | parent [-] | | Programming language is not just syntax, keywords and standard libraries, but also: processes, best practices and design principles. The latter group I guess is more difficult to learn and harder to forget. | | |
| ▲ | bdangubic an hour ago | parent [-] | | I respectfully completely disagree. not only will you just as easily lose thr processed, best practices and design principles but they will be changing over time (what was best practice when I got my first gig in 1997 is not a best practice today (even just 4-5 years ago not to go all the back to the 90’s)). all that is super easy to both forget and lose unless you live it daily |
|
|
| |
| ▲ | root_axis an hour ago | parent | prev | next [-] | | I have also found LLMs are a great tool for understanding a new code base, but it's not clear to me what your comment has to do with skill atrophy. | |
| ▲ | Ifkaluva an hour ago | parent | prev | next [-] | | What do you mean “cut out anything that was granular/one query away”? This was a very cool workflow to hear about—I will be applying it myself | |
| ▲ | SpicyLemonZest 3 hours ago | parent | prev | next [-] | | Are you sure you would know if it didn't work? I use Claude extensively myself, so I'm not saying this from a "hater" angle, but I had 2 people last week who believe themselves to be in your shoes send me pull requests which made absolutely no sense in the context of the codebase. | | |
| ▲ | therealdrag0 3 hours ago | parent [-] | | That’s always been the case, AI or not. | | |
| ▲ | Jweb_Guru an hour ago | parent | next [-] | | No, it hasn't. I did not have a problem before AI with people sending in gigantic pull requests that made absolutely no sense, and justifying them with generated responses that they clearly did not understand. This is not a thing that used to happen. That's not to say people wouldn't have done it if it were possible, but there was a barrier to submitting a pull request that no longer exists. | |
| ▲ | viccis 25 minutes ago | parent | prev | next [-] | | In my experience, the people sending me garbage PRs with Claude are the same ones who wrote garbage code beforehand. Now there's just 10x more of it. | |
| ▲ | windexh8er 2 hours ago | parent | prev [-] | | It just happens to be a lot worse now. Confidence through ignorance has come into the spotlight with the commoditization of LLMs. |
|
| |
| ▲ | viccis 11 minutes ago | parent | prev | next [-] | | It's good that it's working for you but I'm not sure what this has to do with skill atrophy. It sounds like you never had this skill (in this case, working with that particular system) to begin with. >I have a significantly better understanding of the codebase than I would without AI at this point in my onboarding One of the pitfalls of using AI to learn is the same as I'd see students doing pre-AI with tutoring services. They'd have tutors explain the homework to do them and even work through the problems with them. Thing is, any time you see a problem or concept solved, your brain is tricked into thinking you understand the topic enough to do it yourself. It's why people think their job interview questions are much easier than they really are; things just seem obvious when you've thought about the solution. Anyone who's read a tutorial, felt like they understood it well, and then struggled for a while to actually start using the tool to make something new knows the feeling very well. That Todo List app in the tutorial seemed so simple, but the author was making a bunch of decisions constantly that you didn't have to think about as you read it. So I guess my question would be: If you were on a plane flight with no wifi, and you wanted to do some dev work locally on your laptop, how comfortable would you be vs if you had done all that work yourself rather than via Claude? | |
| ▲ | throwaway613746 3 hours ago | parent | prev [-] | | [dead] |
| |
| ▲ | ljm 4 hours ago | parent | prev | next [-] | | Not so much atrophy as apathy. I've worked with people who will look at code they don't understand, say "llm says this", and express zero intention of learning something. Might even push back. Be proud of their ignorance. It's like, why even review that PR in the first place if you don't even know what you're working with? | | |
| ▲ | psygn89 3 hours ago | parent | next [-] | | I cringed when I saw a dev literally copy and paste an AI's response to a concern. The concern was one that had layers and implications to it, but instead of getting an answer as to why it was done a certain way and to allay any potential issues, that dev got a two paragraph lecture on how something worked on the surface of it, wrapped in em dashes and joviality. A good dev would've read deeper into the concern and maybe noticed potential flaws, and if he had his own doubts about what the concern was about, would have asked for more clarification. Not just feed a concern into AI and fling it back. Like please, in this day and age of AI, have the benefit of the doubt that someone with a concern would have checked with AI himself if he had any doubts of his own concern... | |
| ▲ | oremj 3 hours ago | parent | prev | next [-] | | Is this the same subset of people who copy/paste code directly from stack overflow without understanding ? I’m not sure this is a new problem. | | |
| ▲ | foobarchu 3 hours ago | parent | next [-] | | It's a new problem in the sense that now executive management at many (if not most) software companies is pushing for all employees to work this way as much as possible. Those same people probably don't know what stack overflow even is. | |
| ▲ | pizza234 3 hours ago | parent | prev | next [-] | | In my experience, no - I think the ability to build more complete features with less/little/no effort, rather than isolated functions, is (more) appealing to (more) developers. | |
| ▲ | dingaling 2 hours ago | parent | prev | next [-] | | It's difficult to copy & paste an entire app from Stack Overflow | |
| ▲ | malnourish 3 hours ago | parent | prev | next [-] | | I don't think so. I'll spend a ton of time and effort thinking through, revising, and planning out the approach, but I let the agent take the wheel when it comes to transpiling that to code. I don't actually care about the code so long as it's secure and works. I spent years cultivating expertise in C++ and .NET. And I found that time both valuable and enjoyable. But that's because it was a path to solve problems for my team, give guidance, and do so with both breadth and depth. Now I focus on problems at a higher level of abstraction. I am certain there's still value in understanding ownership semantics and using reflection effectively, but they're broadly less relevant concerns. | | | |
| ▲ | sroussey 2 hours ago | parent | prev | next [-] | | Copied and pasted without noting the license that stack overflow has on code published there, no doubt | |
| ▲ | trinsic2 3 hours ago | parent | prev [-] | | Hey. I resemble that remark sometimes!! quit being a hater (sarcasm) :P |
| |
| ▲ | kilroy123 3 hours ago | parent | prev | next [-] | | We've had such developers around, long before LLMs. | | | |
| ▲ | RexM 3 hours ago | parent | prev | next [-] | | It’s a lot like someone bragging that they’re bad at math tossing around equations. | |
| ▲ | monkpit 3 hours ago | parent | prev | next [-] | | If I wanted to know what the LLM says, I would have asked it myself, thanks… | |
| ▲ | redanddead 3 hours ago | parent | prev [-] | | What is it in the broader culture that's causing this? | | |
| ▲ | groundzeros2015 3 hours ago | parent | next [-] | | People who got into the job who don’t really like programming | | | |
| ▲ | mattgreenrocks 3 hours ago | parent | prev [-] | | These people have always existed. Hell, they are here, too. Now they have a new thing to delegate responsibility to. And no, I don't understand them at all. Taking responsibility for something, improving it, and stewarding it into production is a fantastic feeling, and much better than reading the comment section. :) |
|
| |
| ▲ | tossandthrow 4 hours ago | parent | prev | next [-] | | You can argu that you will have skill atrophy by not using LLMs. We have gone multi cloud disaster recovery on our infrastructure. Something I would not have done yet, had we not had LLMs. I am learning at an incredible rate with LLMs. | | |
| ▲ | mgambati 4 hours ago | parent | next [-] | | I kind feel the same. I’m learning things and doing things in areas that would just skip due to lack of time or fear. But I’m so much more detached of the code, I don’t feel that ‘deep neural connection’ from actual spending days in locked in a refactor or debugging a really complex issue. I don’t know how a feel about it. | | |
| ▲ | Fire-Dragon-DoL 4 hours ago | parent | next [-] | | I strongly agree on the refactor, but for debugging I have another perspective: I think debugging is changing for the better, so it looks different. Sure, you don't know the code by heart, but people debugging code translated to assembly already do that. The big difference is being able to unleash scripts that invalidate enormous amount of hypothesis very fast and that can analyze the data. Used to do that by hand it took hours, so it would be a last resort approach. Now that's very cheap, so validating many hypothesis is way cheaper! I feel like my "debugging ability" in terms of value delivered has gone way up. For skill, it's changing. I cannot tell, but the value i am delivering for debugging sessions has gone way up | |
| ▲ | afzalive 4 hours ago | parent | prev [-] | | As someone who's switched from mobile to web dev professionally for the last 6 months now. If you care about code quality, you'll develop that neural connection after some time. But if you don't and there's no PR process (side projects), the motivation to form that connection is quite low. | | |
| ▲ | hombre_fatal 2 hours ago | parent [-] | | > If you care about code quality, you'll develop that neural connection after some time. No, because you can get LLMs to produce high quality code that has gone through an infinite number of refinement/polish cycles and is far more exhaustive than the code you would have written yourself. Once you hit that point, you find yourself in a directional/steering position divorced from the code since no matter what direction you take, you'll get high quality code. |
|
| |
| ▲ | ori_b 4 hours ago | parent | prev | next [-] | | Yes, you certainly can argue that, but you'd be wrong. The primary selling point of LLMs is that they solve the problem of needing skill to get things done. | | |
| ▲ | tossandthrow 4 hours ago | parent | next [-] | | That is not the entire selling point - so you are very wrong. You very much decide how you employ LLMs. Nobody are keeping a gun to your head to use them. In a certain way. Sonif you use them in a way that increase you inherent risk, then you are incredibly wrong. | | |
| ▲ | ori_b 4 hours ago | parent [-] | | I suggest you read the sales pitches that these products have been making. Again, when I say that this is the selling point, I mean it: This is why management is buying them. | | |
| ▲ | SpicyLemonZest 3 hours ago | parent | next [-] | | I've read the sales pitches, and they're not about replacing the need for skill. The Claude Design announcement from yesterday (https://www.anthropic.com/news/claude-design-anthropic-labs) is pretty typical in my experience. The pitch is that this is good for designers, because it will allow them to explore a much broader range of ideas and collaborate on them with counterparties more easily. The tool will give you cool little sliders to set the city size and arc width, but it doesn't explain why you would want to adjust these parameters or how to determine the correct values; that's your job. I understand why a designer might read this post and not be happy about it. If you don't think your management values or appreciates design skill, you'd worry they're going to glaze over the bullet points about design productivity, and jump straight to the one where PMs and marketers can build prototypes and ignore you. But that's not what the sales pitch is focused on. | | |
| ▲ | ori_b 3 hours ago | parent [-] | | The majority of examples in the document you linked describe 'person without<skill> can do thing needing <skill>'. It's very much selling 'more output, less skill' |
| |
| ▲ | trinsic2 3 hours ago | parent | prev [-] | | Sales pitches dont mean jack, WTF are you talking about? | | |
| ▲ | foobarchu 3 hours ago | parent [-] | | Sales pitches are literally the same thing as "the selling point". Neither of those is necessarily a synonym for why you personally use them |
|
|
| |
| ▲ | andy_ppp 4 hours ago | parent | prev | next [-] | | I see it completely the opposite way, you use an LLM and correct all its mistakes and it allows you to deliver a rough solution very quickly and then refine it in combination with the AI but it still gets completely lost and stuck on basic things. It’s a very useful companion that you can’t trust, but it’s made me 4-5x more productive and certainly less frustrated by the legacy codebase I work on. | |
| ▲ | trinsic2 3 hours ago | parent | prev | next [-] | | Yeah I whole hardheartedly disagree with this. Because I understand the basics of coding I can understand where the model gets stuck and prompt it in other directions. If you don't know whats going on through the whole process, good luck with the end product. | |
| ▲ | Forgeties79 3 hours ago | parent | prev [-] | | They purportedly solve the problem of needing skill to get things done. IME, this is usually repeated by VC backed LLM companies or people who haven’t knowingly had to deal with other people’s bad results. This all bumps up against the fact that most people default to “you use the tool wrong” and/or “you should only use it to do things where you already have firm grasp or at least foundational knowledge.” It also bumps against the fact that the average person is using LLM’s as a replacement for standard google search. |
| |
| ▲ | weego 2 hours ago | parent | prev | next [-] | | You're learning at your standard rate of learning, you're just feeding yourself over-confidence on how much you're absorbing vs what the LLM is facilitating you rolling out. | | |
| ▲ | tossandthrow 2 hours ago | parent [-] | | This is such a weird statement in so many levels. The latent assumption here is that learning is zero sum. That you can take a 30 year old from 1856 bring them into present day and they will learn whatever subject as fast as a present day 20 year old. That teachers doesn't matter. That engagement doesn't matter. Learning is not zero sum. Some cultural background makes learning easier, some mentoring makes is easier, and some techniques increases engagement in ways that increase learning speed. |
| |
| ▲ | Wowfunhappy 2 hours ago | parent | prev | next [-] | | > We have gone multi cloud disaster recovery on our infrastructure. Something I would not have done yet, had we not had LLMs. That’s product atrophy, not skill atrophy. | |
| ▲ | bluefirebrand 4 hours ago | parent | prev | next [-] | | > I am learning at an incredible rate with LLMs Could you do it again without the help of an LLM? If no, then can you really claim to have learned anything? | | |
| ▲ | Paradigma11 2 hours ago | parent | next [-] | | So, you havent really learned anything from any teacher if you could not do it again without them? | | |
| ▲ | sho_hn an hour ago | parent | next [-] | | Yes. | |
| ▲ | falkensmaize 2 hours ago | parent | prev | next [-] | | I mean...yeah? If your child says they've learned their multiplication tables but they can't actually multiply any numbers you give them do they actually know how to do multiplication? I would say no. | | |
| ▲ | Jweb_Guru an hour ago | parent [-] | | For some reason people are perfectly able to understand this in the context of, say, cursive, calculator use, etc., but when it comes to their own skillset somehow it's going to be really different. |
| |
| ▲ | UncleMeat an hour ago | parent | prev | next [-] | | Yes that's exactly right. | |
| ▲ | techpression an hour ago | parent | prev [-] | | That would be the definition of learning something, yes. |
| |
| ▲ | tossandthrow 4 hours ago | parent | prev | next [-] | | I could definitely maintain the infrastructure without an llm. Albeit much slower. And yes. If LLMs disappear, then we need to hire a lot of people to maintain the infrastructure. Which naturally is a part of the risk modeling. | | |
| ▲ | bluefirebrand 3 hours ago | parent [-] | | > I could definitely maintain the infrastructure without an llm Not what I asked, but thanks for playing. | | |
| ▲ | tossandthrow 2 hours ago | parent [-] | | You literally asked that question > Could you do it again without the help of an LLM? | | |
|
| |
| ▲ | danw1979 4 hours ago | parent | prev | next [-] | | I think this is a bit dismissive. It’s quite possible to be deep into solving a problem with an LLM guiding you where you’re reading and learning from what it says. This is not really that different from googling random blogs and learning from Stack Overflow. Assuming everyone just sits there dribbling whilst Claude is in YOLO mode isn’t always correct. | |
| ▲ | subscribed 3 hours ago | parent | prev | next [-] | | >> I am learning a new skill with instructor at an incredible rate > Could you do it again on your own? Can you you see how nonsensical your stance is? You're straight up accusing GP of lying they are learning something at the increased rate OR suggesting if they couldn't learn that, presumably at the same rate, on they own, they're not learning anything. That's not very wise to project your own experiences on others. | | |
| ▲ | sroussey 2 hours ago | parent [-] | | Actually, it’s much like taking a physics or engineering course, and after the class being fully able to explain the class that day, and yet realize later when you are doing the homework that you did not actually fully understand like you thought you did. |
| |
| ▲ | _blk 3 hours ago | parent | prev [-] | | The challenge is not if you could do all of it without AI but any of it that you couldn't before. Not everyone learns at the same pace and not everyone has the same fault tolerance threshold. In my experiencd some people are what I call "Japanese learners" perfecting by watching. They will learn with AI but would never do it themselves out of fear of getting something wrong while they understand most of it, others that I call "western learners" will start right away and "get their hands dirty" without much knowledge and also get it wrong right away. Both are valid learning strategies fitting different personalities. |
| |
| ▲ | jjallen 4 hours ago | parent | prev | next [-] | | Also AI could help you pick those skills up again faster, although you wouldn’t need to ever pick those skills up again unless AI ceased to exist. What an interesting paradox-like situation. | | |
| ▲ | estetlinus 3 hours ago | parent [-] | | I believe some professor warned us about being over reliant on Google/reddit etc: “how would you be productive if internet went down” dilemma. Well, if internet is down, so is our revenue buddy. Engineering throughput would be the last of our concerns. |
| |
| ▲ | i_love_retros 4 hours ago | parent | prev | next [-] | | >I am learning at an incredible rate with LLMs. I don't believe it. Having something else do the work for you is not learning, no matter how much you tell yourself it is. | | |
| ▲ | margalabargala 4 hours ago | parent | next [-] | | If you've seen further it's only because you've stood on the shoulders of giants. Having other people do work for you is how people get to focus on things they actually care about. Do you use a compiler you didn't write yourself? If so can you really say you've ever learned anything about computers? | | |
| ▲ | butterisgood 3 hours ago | parent [-] | | You have to build a computer to learn about computers! | | |
| ▲ | viccis 6 minutes ago | parent [-] | | I would argue that if you've just watched videos about building computers and haven't sat down and done one yourself, then yeah I don't see any evidence that you've learned how to build a computer. |
|
| |
| ▲ | tossandthrow 4 hours ago | parent | prev [-] | | It is easy to not believe if you only apply an incredibly narrow world view. Open your eyes, and you might become a believer. | | |
| ▲ | nothinkjustai 3 hours ago | parent [-] | | What is this, some sort of cult? | | |
| ▲ | subscribed 2 hours ago | parent | next [-] | | You mean the cult of "I can't see the viruses therefore they dint exist"? As in "I can't imagine something so it means it's a lie"? Indeed, quite weird and no imagination. | |
| ▲ | tossandthrow 3 hours ago | parent | prev [-] | | No, it is an as snarky response to a person being snarky about usefulness of AI agents. It does seem like there is a cult of people who categorically see LLMs as being poor at anything without it being founded in anything experience other than their 2023 afternoon to play around with it. | | |
| ▲ | nothinkjustai 2 hours ago | parent [-] | | Who cares? Why are people so invested in trying to “convert” others to see the light? Can’t you be satisfied with outcompeting “non believers”? What motivates you to argue on the internet about it? Deep down are you insecure about your reliance on these tools or something, and want everyone else to be as well? | | |
| ▲ | tossandthrow 2 hours ago | parent [-] | | Why do people invest themselves so hard in interjecting themselves into conversations about Ai telling people it doesn't work? It feels so off rebuilding serious SaaS apps in days for production, only to be told it is not possible? |
|
|
|
|
| |
| ▲ | deadbabe 4 hours ago | parent | prev [-] | | Using LLMs as a learning tool isn’t what causes skill atrophy. It’s using them to solve entire problems without understanding what they’ve done. And not even just understanding, but verifying that they’ve implemented the optimal solution. | | |
| ▲ | tehjoker 2 hours ago | parent [-] | | It's partly that, but also reading and surface level understanding something vs generating yourself are different skills with different depths. If you're learning a language, you can get good at listening without getting good at speaking for example. |
|
| |
| ▲ | IgorPartola an hour ago | parent | prev | next [-] | | Yeah I am worried about skill atrophy too. Everyone uses a compiler these days instead of writing assembly. Like who the heck is going to do all the work when people forget how to use the low level tools and a compiler has a bug or something? And don’t get me started on memory management. Nobody even knows how to use malloc(), let alone brk()/mmap(). Everything is relying on automatic memory management. I mean when was the last time you actually used your magnetized needle? I know I am pretty rusty with mine. | | |
| ▲ | otabdeveloper4 an hour ago | parent | next [-] | | > an LLM is exactly like a compiler if a compiler was a black box hosted in a proprietary cloud and metered per symbol Yeah, exactly. | |
| ▲ | techpression an hour ago | parent | prev [-] | | Snark aside, this is an actual problem for a lot of developers in varying degrees, not understand anything about the layers below make for terrible layers above in very many situations. |
| |
| ▲ | solarengineer 3 hours ago | parent | prev | next [-] | | https://hex.ooo/library/power.html When future humans rediscover mathematics. | |
| ▲ | boxingdog 4 hours ago | parent | prev [-] | | [dead] |
|
|
| ▲ | dgellow 3 hours ago | parent | prev | next [-] |
| Another aspect I haven’t seen discussed too much is that if your competitor is 10x more productive with AI, and to stay relevant you also use AI and become 10x more productive. Does the business actually grow enough to justify the extra expense? Or are you pretty much in the same state as you were without AI, but you are both paying an AI tax to stay relevant? |
| |
| ▲ | xixixao 3 hours ago | parent | next [-] | | This is the “ad tax” reasoning, but ultimately I think the answer is greater efficiency. So there is a real value, even if all competitors use the tools. It’s like saying clothing manufacturers are paying the “loom tax” tax when they could have been weaving by hand… | | |
| ▲ | SlinkyOnStairs 3 hours ago | parent | next [-] | | Software development is not a production line, the relationship between code output and revenue is extremely non-linear. Where producing 2x the t-shirts will get you ~2x the revenue, it's quite unlikely that 10x the code will get you even close to 2x revenue. With how much of this industry operates on 'Vendor Lock-in' there's a very real chance the multiplier ends up 0x. AI doesn't add anything when you can already 10x the prices on the grounds of "Fuck you. What are you gonna do about it?" | | |
| ▲ | groundzeros2015 3 hours ago | parent [-] | | Yep and in a vendor lock in scenario, fixing deep bugs or making additions in surgical ways is where the value is. And Claude helps you do that, by giving you more information, analyzing options, but it doesn’t let you make that decision 10x faster. |
| |
| ▲ | bigbadfeline 2 hours ago | parent | prev [-] | | We already know how to multiply the efficiency of human intelligence to produce better quality than LLMs and nearly match their productivity - open source - in fact coding LLMs wouldn't even exist without it. Open source libraries and projects together with open source AI is the only way to avoid the existential risks of closed source AI. |
| |
| ▲ | dakiol 2 hours ago | parent | prev | next [-] | | Where's the evidence of competitors being 10x more productive? So far, everyone is simply bragging about how much code they have shipped last week, but that has zero relevance when it comes to productivity | | |
| ▲ | dgellow an hour ago | parent | next [-] | | Read it as just a given rate. The number doesn’t matter too much here, if company B does believe claims from company A they are N times more productive that’s enough to force B to adopt the same tooling. | |
| ▲ | Silhouette an hour ago | parent | prev [-] | | I feel like a lot of the AI advocacy today is like the Cloud advocacy of a few years ago or the Agile advocacy before that. It's this season's silver bullet to make us all 10x more effective according to metrics that somehow never translate into adding actually useful functionality and quality 10x as fast. The evangelists told us 20 years ago that if we weren't doing TDD then we weren't really professional programmers at all. The evangelists told us 10 years ago that if we were still running stuff locally then we must be paying a fortune for IT admin or not spending our time on the work that mattered. The evangelists this week tell us that we need to be using agents to write all our code or we'll get left in the dust by our competitors who are. I'm still waiting for my flying car. Would settle for some graphics software on Linux that matches the state of the art on Windows or even reliable high-quality video calls and online chat rooms that don't make continental drift look fast. |
| |
| ▲ | redanddead 3 hours ago | parent | prev | next [-] | | The alternative is probably also true. If your F500 competitor is also handicapped by AI somehow, then you're all stagnant, maybe at different levels. Meanwhile Anthropic is scooping up software engineers it supposedly made irrelevant with Mythos and moving into literally 2+ new categories per quarter | |
| ▲ | Lihh27 3 hours ago | parent | prev | next [-] | | it's worse than a tie. 10x everyone just floods the market and tanks per-unit price. you pay the AI tax and your output is worth less. | |
| ▲ | JambalayaJimbo 3 hours ago | parent | prev | next [-] | | If the business doesn’t grow then you shed costs like employees | |
| ▲ | otabdeveloper4 an hour ago | parent | prev | next [-] | | > your competitor is 10x more productive with AI This doesn't happen. Literally zero evidence of this. | | |
| ▲ | dgellow an hour ago | parent [-] | | The actual rate isn’t relevant for the discussion | | |
| ▲ | Miner49er an hour ago | parent | next [-] | | Well it might. If the actual rate is .9x then it matters a lot. Or even if it's like 1.1x, is the cost worth the return? | |
| ▲ | surgical_fire an hour ago | parent | prev [-] | | What if the rate is negative? Would it matter? |
|
| |
| ▲ | senordevnyc 3 hours ago | parent | prev [-] | | Either the business grows, or the market participants shed human headcount to find the optimal profit margin. Isn’t that the great unknown: what professions are going to see headcount reduction because demand can’t grow that fast (like we’ve seen in agriculture), and which will actually see headcount stay the same or even expand, because the market has enough demand to keep up with the productivity gains of AI? Increasingly I think software writ large is the latter, but individual segments in software probably are the former. |
|
|
| ▲ | michaelje 3 hours ago | parent | prev | next [-] |
| Open models keep closing the eval gap for many tasks, and local inference continues to be increasingly viable. What's missing isn't technical capability, but productized convenience that makes the API path feel like the only realistic option. Frontier labs are incentivized to keep it that way, and they're investing billions to make AI = API the default. But that's a business model, not a technical inevitability. |
| |
| ▲ | trueno an hour ago | parent [-] | | im hoping and praying that local inference finds it's way to some sort of baseline that we're all depending on claude for here. that would help shape hardware designs on personal devices probably something in the direction of what apple has been doing. ive had to like tune out of the LLM scene because it's just a huge mess. It feels impossible to actually get benchmarks, it's insanely hard to get a grasp on what everyone is talking about, bots galore championing whatever model, it's just way too much craze and hype and misinformation. what I do know is we can't keep draining lakes with datacenters here and letting companies that are willing to heel turn on a whim basically control the output of all companies. that's not going to work, we collectively have to find a way to make local inference the path forward. everyone's foot is on the gas. all orgs, all execs, all peoples working jobs. there's no putting this stuff down, and it's exhausting but we have to be using claude like _right now_. pretty much every company is already completely locked in to openai/gemini/claude and for some unfortunate ones copilot. this was a utility vendor lock in capture that happened faster than anything ive ever seen in my life & I already am desperate for a way to get my org out of this. | | |
| ▲ | hakfoo 23 minutes ago | parent [-] | | I'm frustrated that there's not "solid" instructional tooling. I either see people just saying "keep trying different prompts and switching models until you get lucky" or building huge cantilevered toolchains that seems incredibly brittle, and even then, how well do they really work? I get choice paralysis when you show me a prompt box-- I don't know what I can reasonably ask for and how to best phrase it, so I just panic. It doesn't help when we see articles saying people are getting better outcomes by adding things like "and no bugs plz owo" I'm sure this is by design-- anything with clear boundaries and best practices would discourage gacha style experimentation. Can you trust anyone who sells you a metered service to give you good guidance on how to use it efficiently? | | |
| ▲ | trueno 8 minutes ago | parent [-] | | yea that is probably the worst part of these techs becoming mainstream services and local-LLM'ing taking off in general: working with them at many points in any architecture no longer feels... deterministic i guess. way too fucking much "heres what i use" but no real best practices yet, just a lot of vague gray area and everyones still in discovery-mode on how to best find some level of determinism or workflow and ways we are benchmarking is seriously a moving target. everyone has their own branded take on what the technology is and their own branded approach on how to use it, and it's probably the murkiest and foggiest time to be in technology fields that i've ever seen :\ seems like weekly/monthly something is outdated, not just the models but the tooling people are parroting as the current best tooling to use. incredibly frustrating. there's simply too much ground to cover for any one person to have any absolute takes on any of it, and because a handful of entities are currently leading the charge draining lakes and trying to compete for every person and every businesses money, there's zero organized frameworks at the top to make some sense of this. they all are banking on their secret sauce, and i _really_ want us all to get away from this. local inference has to succeed imo but goddamn there needs to be some collective working together to rally behind some common strats/frameworks here. im sure there's already countless committees that have been established to try and get in front of this but even that's messy. i don't know how else to phrase it: this feels like such an unstable landscape, "beta" software/services are running rampant in every industry/company/org/etc and there's absolutely no single resource we can turn to to help stay ahead of & plan for the rapidly-evolving landscape. every, and i mean every company, is incredibly irresponsible for using this stuff. including my own. once again though, cat's already out of the bag. now we fight for our lives trying to contain it and ensure things are well understood and implemented properly...which seems to be the steepest uphill battle of my life |
|
|
|
|
| ▲ | dewarrn1 3 hours ago | parent | prev | next [-] |
| I'm hopeful that new efficiencies in training (Deepseek et al.), the impressive performance of smaller models enhanced through distillation, and a glut of past-their-prime-but-functioning GPUs all converge make good-enough open/libre models cheap, ubiquitous, and less resource-intensive to train and run. |
|
| ▲ | leonidasv 2 hours ago | parent | prev | next [-] |
| >perhaps we can come up with something like the "linux/postgres/git/http/etc" of the LLMs I fear that this may not be feasible in the long term. The open-model free ride is not guaranteed to continue forever; some labs offer them for free for publicity after receiving millions in VC grants now, but that's not a sustainable business model. Models cost millions/billions in infrastructure to train. It's not like open-source software where people can just volunteer their time for free; here we are talking about spending real money upfront, for something that will get obsolete in months. Current AI model "production" is more akin to an industrial endeavor than open-source arrangements we saw in the past. Until we see some breakthrough, I'm bearish on "open models will eventually save us from reliance on big companies". |
| |
| ▲ | falkensmaize an hour ago | parent [-] | | "get obsolete in months" If you mean obsolete in the sense of "no longer fit for purpose" I don't think that's true. They may become obsolete in terms of "can't do hottest new thing" but that's true of pretty much any technology. A capable local model that can do X will always be able to do X, it just may not be able to do Y. But if X is good enough to solve your problem, why is a newer better model needed? I think if we were able to achieve ~Opus 4.6 level quality in a local model that would probably be "good enough" for a vast number of tasks. I think it's debatable whether newer models are always better - 4.7 seems to be somewhat of a regression for example. |
|
|
| ▲ | tossandthrow 4 hours ago | parent | prev | next [-] |
| The lock in is so incredibly poor. I could switch to whatever provider in minuets. But it requires that one does not do something stupid. Eg. For recurring tasks: keep the task specification in the source code and just ask Claude to execute it. The same with all documentation, etc. |
|
| ▲ | aliljet 4 hours ago | parent | prev | next [-] |
| What open models are truly competing with both Claude Code and Opus 4.7 (xhigh) at this stage? |
| |
| ▲ | parinporecha 3 hours ago | parent | next [-] | | I've had a good experience with GLM-5.1. Sure it doesn't match xhigh but comes close to 4.6 at 1/3rd the cost | |
| ▲ | Someone1234 3 hours ago | parent | prev | next [-] | | That's a lame attitude. There are local models that are last year's SOTA, but that's not good enough because this year's SOTA is even better yet still... I've said it before and I'll say it again, local models are "there" in terms of true productive usage for complex coding tasks. Like, for real, there. The issue right now is that buying the compute to run the top end local models is absurdly unaffordable. Both in general but also because you're outbidding LLM companies for limited hardware resources. You have a $10K budget, you can legit run last year's SOTA agentic models locally and do hard things well. But most people don't or won't, nor does it make cost effective sense Vs. currently subsidized API costs. | | |
| ▲ | gbro3n 3 hours ago | parent | next [-] | | I completely see your point, but when my / developer time is worth what it is compared to the cost of a frontier model subscription, I'm wary of choosing anything but the best model I can. I would love to be able to say I have X technique for compensating for the model shortfall, but my experience so far has been that bigger, later models out perform older, smaller ones. I genuinely hope this changes through. I understand the investment that it has taken to get us to this point, but intelligence doesn't seem like it's something that should be gated. | | |
| ▲ | Someone1234 3 hours ago | parent | next [-] | | Right; but every major generation has had diminishing returns on the last. Two years ago the difference was HUGE between major releases, and now we're discussing Opus 4.6 Vs. 4.7 and people cannot seem to agree if it is an improvement or regression (and even their data in the card shows regressions). So my point is: If you have the attitude that unless it is the bleeding edge, it may have well not exist, then local models are never going to be good enough. But truth is they're now well exceeding what they need to be to be huge productivity tools, and would have been bleeding edge fairly recently. | | |
| ▲ | gbro3n 3 hours ago | parent [-] | | I feel like I'm going to have to try the next model. For a few cycles yet. My opinion is that Opus 4.7 is performing worse for my current work flow, but 4.6 was a significant step up, and I'd be getting worse results and shipping slower if I'd stuck with 4.5. The providers are always going to swear that the latest is the greatest. Demis Hassabis recently said in an interview that he thinks the better funded projects will continue to find significant gains through advanced techniques, but that open source models figure out what was changed after about 6 months or so. We'll see I guess. Don't get me wrong, I'd love to settle down with one model and I'd love it to be something I could self host for free. |
| |
| ▲ | dakiol 2 hours ago | parent | prev [-] | | > I completely see your point, but when my / developer time is worth what it is compared to the cost of a frontier model subscription, I'm wary of choosing anything but the best model I can. Don't you understand that by choosing the best model we can, we are, collectively, step by step devaluating what our time is worth? Do you really think we all can keep our fancy paychecks while keep using AI? | | |
| ▲ | gbro3n an hour ago | parent [-] | | Do you think if you or me stopped using AI that everyone else will too? We're still what we always were - problem solvers who have gained the ability to learn and understand systems better that the general population, communicate clearly (to humans and now AIs). Unfortunately our knowledge of language APIs and syntax has diminished in value, but we have so many more skills that will be just as valuable as ever. As the amount of software grows, so will the need for people who know how to manage the complexity that comes with it. |
|
| |
| ▲ | aliljet 3 hours ago | parent | prev | next [-] | | First, making sure to offer an upvote here. I happen to be VERY enthusiastic about local models, but I've found them to be incredibly hard to host, incredibly hard to harness, and, despite everything, remarkably powerful if you are willing to suffer really poor token/second performance... | |
| ▲ | HWR_14 2 hours ago | parent | prev | next [-] | | $10k is a lot of tokens. | | | |
| ▲ | wellthisisgreat 2 hours ago | parent | prev [-] | | > that are last year's SOTA Early last year or late last year? opus 4.5 was quite a leap |
| |
| ▲ | esafak 3 hours ago | parent | prev [-] | | GLM 5.1 competes with Sonnet. I'm not confident about Opus, though they claim it matches that too. | | |
| ▲ | ojosilva 3 hours ago | parent [-] | | I have it as failover to Opus 4.6 in a Claude proxy internally. People don't notice a thing when it triggers, maybe a failed tool call here and there (harness remains CC not OC) or a context window that has gone over 200k tokens or an image attachment that GLM does not handle, otherwise hunky-dory all the way. I would also use it as permanent replacement for haiku at this proxy to lower Claude costs but have not tried it yet. Opus 4.7 has shaken our setup badly and we might look into moving to Codex 100% (GLM could remain useful there too). |
|
|
|
| ▲ | sourya4 an hour ago | parent | prev | next [-] |
| yep!! had similar thoughts on the the "linux/postgres/git/http/etc" of the LLMs made a HN post of my X article on the lock-in factor and how we should embrace the modular unix philosophy as a way out: https://news.ycombinator.com/item?id=47774312 |
|
| ▲ | 2 hours ago | parent | prev | next [-] |
| [deleted] |
|
| ▲ | GaryBluto 3 hours ago | parent | prev | next [-] |
| > open models Google just released Gemma 4, perhaps that'd be worth a try? |
|
| ▲ | ben8bit 4 hours ago | parent | prev | next [-] |
| Any recommendations on good open ones? What are you using primarily? |
| |
| ▲ | culi 3 hours ago | parent | next [-] | | LMArena actually has a nice Pareto distribution of ELO vs price for this model elo $/M
---------------------------------------
glm-5.1 1538 2.60
glm-4.7 1440 1.41
minimax-m2.7 1422 0.97
minimax-m2.1-preview 1392 0.78
minimax-m2.5 1386 0.77
deepseek-v3.2-thinking 1369 0.38
mimo-v2-flash (non-thinking) 1337 0.24
https://arena.ai/leaderboard/code?viewBy=plot&license=open-s... | | |
| ▲ | logicprog 2 hours ago | parent [-] | | LMArena isn't very useful as a benchmark, however I can vouch for the fact that GLM 5.1 is astonishingly good. Several people I know who have a $100/mo Claude Code subscription are considering cancelling it and going all in on GLM, because it's finally gotten (for them) comparable to Opus 4.5/6. I don't use Opus myself, but I can definitely say that the jump from the (imvho) previous best open weight model Kimi K2.5 to this is otherworldly — and K2.5 was already a huge jump itself! |
| |
| ▲ | blahblaher 4 hours ago | parent | prev | next [-] | | qwen3.5/3.6 (30B) works well,locally, with opencode | | |
| ▲ | zozbot234 4 hours ago | parent | next [-] | | Mind you, a 30B model (3B active) is not going to be comparable to Opus. There are open models that are near-SOTA but they are ~750B-1T total params. That's going to require substantial infrastructure if you want to use them agentically, scaled up even further if you expect quick real-time response for at least some fraction of that work. (Your only hope of getting reasonable utilization out of local hardware in single-user or few-users scenarios is to always have something useful cranking in the background during downtime.) | | |
| ▲ | pitched 4 hours ago | parent | next [-] | | For a business with ten or more engineers/people-using-ai, it might still make sense to set this up. For an individual though, I can’t imagine you’d make it through to positive ROI before the hardware ages out. | | |
| ▲ | zozbot234 3 hours ago | parent | next [-] | | It's hard to tell for sure because the local inference engines/frameworks we have today are not really that capable. We have barely started exploring the implications of SSD offload, saving KV-caches to storage for reuse, setting up distributed inference in multi-GPU setups or over the network, making use of specialty hardware such as NPUs etc. All of these can reuse fairly ordinary, run-of-the-mill hardware. | |
| ▲ | DeathArrow 3 hours ago | parent | prev [-] | | Since you need at least a few of H100 class hardware, I guess you need at least few tens of coders to justify the costs. |
| |
| ▲ | wuschel 3 hours ago | parent | prev | next [-] | | What near SOTA open models are you referring to? | |
| ▲ | cyberax 3 hours ago | parent | prev [-] | | I'm backing up a big dataset onto tapes, so I wanted to automate it. I have an idle 64Gb VRAM setup in my basement, so I decided to experiment and tasked it with writing an LTFS implementation. LTFS is an open standard for filesystems for tapes, and there's an implementation in C that can be used as the baseline. So far, Qwen 3.6 created a functionally equivalent Golang implementation that works against the flat file backend within the last 2 days. I'm extremely impressed. | | |
| ▲ | Gareth321 17 minutes ago | parent [-] | | It is surprisingly competent. It's not Opus 4.6 but it works well for well structured tasks. |
|
| |
| ▲ | pitched 4 hours ago | parent | prev | next [-] | | I want to bump this more than just a +1 by recommending everyone try out OpenCode. It can still run on a Codex subscription so you aren’t in fully unfamiliar territory but unlocks a lot of options. | | |
| ▲ | zozbot234 4 hours ago | parent | next [-] | | The Codex TUI harness is also open source and you can use open models with it, so you can stay in even more familiar territory. | |
| ▲ | pwython 4 hours ago | parent | prev [-] | | pi-coding-agent (pi.dev) is also great. I've been using it with Gemma 4 and Qwen 3.6. |
| |
| ▲ | jherdman 4 hours ago | parent | prev | next [-] | | Is this sort of setup tenable on a consumer MBP or similar? | | |
| ▲ | Gareth321 16 minutes ago | parent | next [-] | | The Mac Minis (probably 64GB RAM) are the most cost effective. | |
| ▲ | danw1979 4 hours ago | parent | prev | next [-] | | Qwen’s 30B models run great on my MBP (M4, 48GB) but the issue I have is cooling - the fan exhaust is straight onto the screen, which I can’t help thinking will eventually degrade it, given the thermal cycling it would go through. A Mac Studio makes far more sense for local inference just for this reason alone. | |
| ▲ | pitched 4 hours ago | parent | prev [-] | | For a 30B model, you want at least 20GB of VRAM and a 24GB MBP can’t quite allocate that much of it to VRAM. So you’d want at least a 32GB MBP. | | |
| ▲ | richardfey 3 hours ago | parent | next [-] | | I have 24GB VRAM available and haven't yet found a decent model or combination.
Last one I tried is Qwen with continue, I guess I need to spend more time on this. | |
| ▲ | zozbot234 4 hours ago | parent | prev | next [-] | | It's a MoE model so I'd assume a cheaper MBP would simply result in some experts staying on CPU? And those would still have a sizeable fraction of the unified memory bandwidth available. | | |
| ▲ | pitched 3 hours ago | parent [-] | | I haven’t tried this myself yet but you would still need enough non-vram ram available to the cpu to offload to cpu, right? This is a fully novice question, I have not ever tried it. |
| |
| ▲ | _blk 3 hours ago | parent | prev [-] | | Is there any model that practically compares to Sonnet 4.6 in code and vision and runs on home-grade (12G-24G) cards? | | |
| ▲ | macwhisperer an hour ago | parent [-] | | im currently running a custom Gemma4 26b MoE model on my 24gb m2... super fast and it beat deepseek, chatgpt, and gemini in 3 different puzzles/code challenges I tested it on. the issue now is the low context... I can only do 2048 tokens with my vram... the gap is slowly closing on the frontier models |
|
|
| |
| ▲ | cpursley 4 hours ago | parent | prev [-] | | How are you running it with opencode, any tips/pointers on the setup? |
| |
| ▲ | cmrdporcupine 4 hours ago | parent | prev | next [-] | | GLM 5.1 via an infra provider. Running a competent coding capable model yourself isn't viable unless your standards are quite low. | | | |
| ▲ | DeathArrow 3 hours ago | parent | prev [-] | | I am using GLM 5.1 and MiniMax 2.7. |
|
|
| ▲ | crgk 2 hours ago | parent | prev | next [-] |
| Who’s your “we,” if you don’t mind sharing? I’m curious to learn more about companies/organizations with this perspective. |
|
| ▲ | finghin 2 hours ago | parent | prev | next [-] |
| I’m imagining a (private/restricted) tracker style system where contributors “seed” compute and users “leech”. |
|
| ▲ | atleastoptimal an hour ago | parent | prev | next [-] |
| Open models are only near SOTA because of distillation from closed models. |
|
| ▲ | giancarlostoro 2 hours ago | parent | prev | next [-] |
| > I think that's the way forward. Actually it would be great if everybody would put more focus on open models, I'm still surprised top CS schools are not investing in having their students build models, I know some are, but like, when's the last time we talked about a model not made by some company, versus a model made by some college or university, which is maintained by the university and useful for all. It's disgusting that OpenAI still calls itself "Open AI" when they aren't truly open. |
|
| ▲ | i_love_retros 4 hours ago | parent | prev | next [-] |
| > we don't want a hard dependency on another multi-billion dollar company just to write software My manager doesn't even want us to use copilot locally. Now we are supposed to only use the GitHub copilot cloud agent. One shot from prompt to PR. With people like that selling vendor lock in for them these companies like GitHub, OpenAI, Anthropic etc don't even need sales and marketing departments! |
| |
| ▲ | tossandthrow 4 hours ago | parent [-] | | You are aware that using eg. Github copilot is not one shot? It will start an agentic loop. | | |
| ▲ | dgellow 3 hours ago | parent [-] | | Unnecessary nitpicking | | |
| ▲ | tossandthrow 3 hours ago | parent [-] | | Why? One shoting has a very specific meaning, and agentic workflows are not it? What is the implied meaning I should understand from them using one shot? They might refer to the lack of humans in the loop. | | |
| ▲ | dgellow an hour ago | parent [-] | | You give a prompt, you get a PR. If it is ready to merge with the first attempt, that’s a one shot. The agentic loop is a detail in their context | | |
| ▲ | tossandthrow an hour ago | parent [-] | | I would recommend reading up on this. Inside of the agentic loop the agent will enrich the context from reading files or by searching the internet. What you are writing is a severe display of ignorance in the area. |
|
|
|
|
|
|
| ▲ | Frannky 3 hours ago | parent | prev | next [-] |
| Opencode go with open models is pretty good |
|
| ▲ | sergiotapia 3 hours ago | parent | prev | next [-] |
| I can recommend this stack. It works well with the existing Claude skills I had in my code repos: 1. Opencode 2. Fireworks AI: GLM 5.1 And it is SIGNIFICANTLY cheaper than Claude. I'm waiting eagerly for something new from Deepseek. They are going to really show us magic. |
| |
| ▲ | dirasieb 3 hours ago | parent [-] | | it is also significantly less capable than claude | | |
| ▲ | dakiol 2 hours ago | parent [-] | | That's fine. When the "best of the best" is offered only by a couple of companies that are not looking into our best interests, then we can discard them |
|
|
|
| ▲ | OrvalWintermute 2 hours ago | parent | prev | next [-] |
| I'm increasingly thinking the same as our spend on tokens goes up. If you have HPC or Supercompute already, you have much of the expertise on staff already to expand models locally, and between Apple Silicon and Exo there are some amazingly solutions out there. Now, if only the rumors about Exo expanding to Nvidia are true.. |
|
| ▲ | somewhereoutth 2 hours ago | parent | prev | next [-] |
| My understanding is that the major part of the cost of a given model is the training - so open models depend on the training that was done for frontier models? I'm finding hard to imagine (e.g.) RLHF being fundable through a free software type arrangement. |
| |
| ▲ | zozbot234 2 hours ago | parent [-] | | No, the training between proprietary and open models is completely different. The speculation that open models might be "distilled" from proprietary ones is just that, speculation, and a large portion of it is outright nonsense. It's physically possible to train on chat logs from another model but that's not "distilling" anything, and it's not even eliciting any real fraction of the other model's overall knowledge. | | |
| ▲ | tehjoker an hour ago | parent [-] | | I don't know what to make of it, I am skeptical of OpenAI/Anthropic claims about distillation, but I did notice DeepSeek started sounding a lot like Claude recently. |
|
|
|
| ▲ | wahnfrieden an hour ago | parent | prev | next [-] |
| or just use codex |
|
| ▲ | DeathArrow 3 hours ago | parent | prev | next [-] |
| >perhaps we can come up with something like the "linux/postgres/git/http/etc" of the LLMs: something we all can benefit from while it not being monopolized by a single billionarie company Training and inference costs so we would have to pay for them. |
| |
|
| ▲ | SilverElfin 3 hours ago | parent | prev | next [-] |
| Is that why they are racing to release so many products? It feels to me like they want to suck up the profits from every software vertical. |
| |
| ▲ | Bridged7756 3 hours ago | parent [-] | | Yeah it seems so. Anthropic has entered the enshittification phase. They got people hooked onto their SOTAs so it's now time to keep releasing marginal performance increase models at 40% higher token price. The problem is that both Anthropic and OpenAI have no other income other than AI. Can't Google just drown them out with cheaper prices over the long run? It seems like an attrition battle to me. |
|
|
| ▲ | boxingdog 4 hours ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | throwaway613746 3 hours ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | gbgarbeb 3 hours ago | parent | prev | next [-] |
| [dead] |
|
| ▲ | sky2224 2 hours ago | parent | prev [-] |
| This is part of the reason why I'm really worried that this is all going to result in a greater economic collapse than I think people are realizing. I think companies that are shelling out the money for these enterprise accounts could honestly just buy some H100 GPUs and host the models themselves on premises. Github CoPilot enterprise charges $40 per user per month (this can vary depending on your plan of course), but at this price for 1000 users that comes out to $480,000 a year. Maybe I'm missing something, but that's roughly what you're going to be spending to get a full fledged hosting setup for LLMs. |
| |
| ▲ | subarctic an hour ago | parent | next [-] | | Out of curiosity, how many concurrent users could you get with a hosting setup at that price? If let's say 10% of those 1000 users were using it at the same time would it handle it? What about 30% or 100%? | |
| ▲ | merlinoa an hour ago | parent | prev [-] | | Most companies don't want to host it themselves. They want someone to do it for them, and they are happy to pay for it. If it makes their lives easier and does not add complexity, then it has a lot of value. |
|