| ▲ | joshstrange 5 hours ago |
| LLMs and LLM providers are massive black boxes. I get a lot of value from them and so I can put up with that to a certain extent, but these new "products"/features that Anthropic are shipping are very unappealing to me. Not because I can't see a use-case for them, but because I have 0 trust in them: - No trust that they won't nerf the tool/model behind the feature - No trust they won't sunset the feature (the graveyard of LLM-features is vast and growing quickly while they throw stuff at the wall to see what sticks) - No trust in the company long-term. Both in them being around at all and them not rug-pulling. I don't want to build on their "platform". I'll use their harness and their models but I don't want more lock-in than that. If Anthropic goes "bad" I want to pick up and move to another harness and/or model with minimal fuss. Buying in to things like this would make that much harder. I'm not going to build my business or my development flows on things I can't replicate myself. Also, I imagine debugging any of this would be maddening. The value add is just not there IMHO. EDIT: Put another way, LLM companies are trying to climb the ladder to be a platform, I have zero interest in that, I was a "dumb pipe", I want a commodity, I want a provider, not a platform. Claude Code is as far into the dragon's lair that I want to venture and I'm only okay with that because I know I can jump to OpenCode/Codex/etc if/when Anthropic "goes bad". |
|
| ▲ | ElFitz 16 minutes ago | parent | next [-] |
| > Not because I can't see a use-case for them, but because I have 0 trust in them > […] > Put another way, LLM companies are trying to climb the ladder to be a platform, I have zero interest in that, I was a "dumb pipe", I want a commodity, I want a provider, not a platform. That is my sentiment precisely, and a big reason why I’ve started moving away from Claude Code in the past few weeks when I realised how much of my workflow was becoming tied to their specific tools. Claude Code’s "Memory" feature was the tipping point for me, with the model committing feedbacks and learnings to some local, provider-specific path, that won’t persist in the git repo itself. That’s fine for user preferences, not for workflows, rules, etc. And the latest ToS changes about not being allowed to even use another CLI made up my mind. At work we were experimenting with an autonomous debug agent using the Claude Code cli programmatically in ephemeral VMs. Now it just returns an error saying we can’t use subscriptions with third-party software… when there is no third-party software involved? Anyway, so long Claude. |
|
| ▲ | freedomben an hour ago | parent | prev | next [-] |
| This echoes my thoughts exactly. I've tried to stay model-agnostic but the nudges and shoves from Anthropic continue to make that a challenge. No way I'm going that deep into their "cloud" services, unless it's a portable standard. I did MCP and skills because those were transferrable. I also clearly see the lock-in/moat strategy playing out here, and I don't like it. It's classic SV tactics. I've been burned too many times to let it happen again if I can help it. |
| |
| ▲ | hatmanstack 14 minutes ago | parent [-] | | Agree. I just don't think it's realistic to expect the technology to not become a tool for commercialism. It plays out the same way every time: technology arrives, mass adoption with idealist intentions, somebody has to pay the mortgage, delight disappears. Woz has been saying this for decades, we went from buying a computer and owning it to being trapped inside someone else's platform. MCP being open was a good sign but I'm watching how tightly Routines gets coupled to their stack. |
|
|
| ▲ | pc86 2 hours ago | parent | prev | next [-] |
| > - No trust that they won't nerf the tool/model behind the feature To the contrary, they've proven again and again and again they'll absolutely do that the first chance they get. |
| |
| ▲ | rbalicki 2 hours ago | parent [-] | | You can lessen your dependence on the specific details of how /loop, code routines, etc. work by asking the LLM to do simpler tasks, and instead, having a proper workflow engine be in charge of the workflow aspects. For example, this demo (https://github.com/barnum-circus/barnum/tree/master/demos/co...) converts a folder of files from JS to TS. It's something an LLM could (probably) do a decent job of, but 1. not necessarily reliably, and 2. you can write a much more complicated workflow (e.g. retry logic, timeout logic, adding additional checks like "don't use as casts", etc), 3. you can be much more token efficient, and 4. you can be LLM agnostic. So, IMO, in the presence of tools like that, you shouldn't bother using /loop, code routines, etc. |
|
|
| ▲ | JohnMakin 3 hours ago | parent | prev | next [-] |
| This is a similar sentiment I heard early on in the cloud adoption fever, many companies hedged by being “multi cloud” which ended up mostly being abandoned due to hostile patterns by cloud providers, and a lot of cost. Ultimately it didn’t really end up mattering and the most dire predictions of vendor lock in abuse didn’t really happen as feared (I know people will disagree with this, but specifically speaking about aws, the predictions vs what actually happened is a massive gap. note I have never and will never use azure, so I could be wrong on that particular one). I see people making similar conclusions about various LLM providers. I suspect in the end it’ll shake out about the same way, the providers will become practically inoperable with each other either due to inconvenience, cost, or whatever. So I’ve not wasted much of my time thinking about it. |
| |
| ▲ | michaeldwan 2 hours ago | parent | next [-] | | I credit containerization, k8s, and terraform for preventing vendor lock in. Compute like EC2 or GCE are effectively interoperable. Ditto for managed services for k8s or Postgres. The new products Anthropic is shipping is more like Lambda. Vendor kool-aid lots of people will buy into. What grinds my gears is how Anthropic is actively avoiding standards. Like being the only harness that doesn't read AGENTS.md. I work on AI infra and use different models all the time, Opus is really good, but the competition is very close. There's just enough friction to testing those out though, and that's the point. | | |
| ▲ | JohnMakin 2 hours ago | parent | next [-] | | I think there is lock-in, despite those things - for containerization, you're still a lot of the times beholden to the particular runtime that provider prefers, and whatever weird quirks exist there. Migrating can have some surprises. K8s, usually you will go managed there, and while they provide the same functionality, AKS != EKS != GKE at all, at least in terms of managing them and how they plug into everything else. In terraform, migrating from AWS provider to GCP provider will hold a lot of surprises for you for what looks like it should be the exact same thing. My point was, I don't think it mattered much, and it feels like an ok comparison - cloud offerings are mostly the exact same things, at least at their core, but the ecosystem around them is the moat, and how expensive it is to migrate off of them. I would not be surprised at all if frontier AI model providers go much the same way. I'm pretty much there already with how much I prefer claude code CLI, even if half the time I'm using it as a harness for OpenAI calls. | |
| ▲ | fragmede an hour ago | parent | prev [-] | | There's a tiny amount of friction. Enough that I'll be honest and say that I spend the majority of my time with one vendor's system, but compared the to the fiction of moving from one cloud to another, eg AWS to GCP, the friction between opening Claude code vs codex is basically zero. Have an active subscription and have Claude.md say "read Agents.md". Claude Code routines sounds useful, but at the same time, under AI-codepocalypse, my guess is it would take an afternoon to have codex reimplement it using some existing freemium SaaS Cron platform, assuming I didn't want to roll my own (because of the maintenance overhead vs paying someone else to deal with that). | | |
| ▲ | michaeldwan 24 minutes ago | parent [-] | | you're spot on. I use both Claude Code + OpenCode with many different models and friction is minimal as long as I'm deliberate about it. Hell, even symlinking AGENTS.md to CLAUDE.md is like 80% there. It's just portability v convenience. But unlike ~15 years ago with cloud compute, it _feels_ like more people are skeptical of convenience, which is interesting. |
|
| |
| ▲ | chickensong an hour ago | parent | prev | next [-] | | > specifically speaking about aws, the predictions vs what actually happened is a massive gap I guess I'm one of the people who disagree, specifically about AWS. I think a lot of companies just watch their bill go up because they don't have the appetite to unwind their previous decision to go all-in on AWS. Ignoring egress fees, migrating storage and compute isn't hard, it's all the auxiliary stuff that's locked in, the IAM, Cognito, CloudFormation, EventBridge, etc... Good luck digging out of that hole. That's not to say that AWS doesn't work well, but unless you have a light footprint and avoided most of their extra services, the lock-in feels pretty real. That's what it feels like Anthropic is doing here. You could have a cron job under your control, or you could outsource that to a Claude Routine. At some point the outsourced provider has so many hooks into your operations that it's too painful to extract yourself, so you just keep the status quo, even if there's pain. | | |
| ▲ | ElFitz 8 minutes ago | parent | next [-] | | I am curious, what do people use Cognito for? I’ve never not ended up regretting using it. | |
| ▲ | JohnMakin 42 minutes ago | parent | prev [-] | | the AWS things you mentioned you don’t need to mess with at all, with the exception of IAM, which doesn’t cost anything at all. your experience just hasn’t been my experience I guess. The more managed the service you use, the more costs you are going to pay - for a very long time I’ve got by with paying for compute, network, and storage on the barebones services. If you want to pay for convenience you will pay for it. One area that was a little shitty that has changed a lot is egress costs, but we mostly have shifted to engineering around it. I’ve never minded all that much, and AWS support is so good at enterprise tiers that they’ll literally help you do it. | | |
| ▲ | AshamedBadger56 2 minutes ago | parent [-] | | > the AWS things you mentioned you don’t need to mess with at all not the op, but I suspect they were meaning it's a huge pain migrating to a different cloud provider when all those features mentioned are in use. not that managing them is a mess in AWS. |
|
| |
| ▲ | robwwilliams 3 hours ago | parent | prev | next [-] | | There are different level of who gets locked in. Almost every health care system in the USA is locked in to either an Epic/Oracle barrel or a Cerner barrel. I hope AI breaks this duopoly open soon. | |
| ▲ | phist_mcgee an hour ago | parent | prev [-] | | Let's see how it shakes out after Athropic and OpenAI fully stop subsidizing their plans, that may alter the calculus. |
|
|
| ▲ | mikepurvis 4 hours ago | parent | prev | next [-] |
| > I want to pick up and move to another harness and/or model with minimal fuss. Buying in to things like this would make that much harder. Yes, I expect that is very much the point here. A bunch of product guys got on a whiteboard and said, okay the thing is in wide use but the main moat is that our competitors are even more distrusted in the market than we are; other than that it's completely undifferentiated and can be swapped out in a heartbeat for multiple other offerings. How do we do we persuade our investors we have a locked in customer base that won't just up-stakes in favour of other options or just running open source models themselves? |
| |
| ▲ | throwup238 3 hours ago | parent [-] | | I think they really knee capped themselves when they released Claude for Github integrations, which allows anyone to use their Claude subscription to run Claude Code in Github actions for code reviews and arbitrary prompts. Now they’re trying to back track that with a cloud solution. |
|
|
| ▲ | jeppester an hour ago | parent | prev | next [-] |
| I always hated SEO because it was not an exact science - like programming was. Too bad we've now managed to turn programming into the same annoying guesswork. |
|
| ▲ | palata 4 hours ago | parent | prev | next [-] |
| > - No trust that they won't nerf the tool/model behind the feature I actually trust that they will. |
| |
| ▲ | gardenhedge 4 hours ago | parent | next [-] | | Yeah, I build my workflows with two things in mind: 1) that AI will be more advanced in the future 2) that the AI I am using will be worse in the future | | |
| ▲ | freedomben an hour ago | parent [-] | | Same! I actually have some comments in my codebase now like this one: # Note: This is inefficient, but deterministic and predictable. Previous
attempts at improvements led to hard-to-predict bugs and were
scrapped. TODO improve this function when AI gets better
I don't love it or even like it, but it is realistic. |
| |
| ▲ | dvfjsdhgfv 4 hours ago | parent | prev [-] | | I believe the current game everybody plays is: * make sure the model maxes out all benchmarks * release it * after some time, nerf it * repeat the same with the next model However, the net sum is positive: in general, models from 2026 are better than those from 2024. | | |
| ▲ | snek_case 4 hours ago | parent | next [-] | | I guess there's a pretty clear incentive to nerf the current model right before the next model is about to come out. | | |
| ▲ | chinathrow 3 hours ago | parent [-] | | Wouldn't that amount to fraud? | | |
| ▲ | tomwojcik 3 hours ago | parent | next [-] | | Serious question, do we actually know what we're paying for? All I know is it's access to models via cli, aka Claude Code. We don't know what models they use, how system prompt changes or what are the actual rate limits (Yet Anthropic will become 1 trillion dollars company in a moment). | | |
| ▲ | xienze 2 hours ago | parent [-] | | > We don't know what models they use, how system prompt changes or what are the actual rate limits (Yet Anthropic will become 1 trillion dollars company in a moment). Not just that, but there’s really no way to come to an objective consensus of how well the model is performing in the first place. See: literally every thread discussing a Claude outage or change of some kind. “Opus is absolutely incredible, it’s one shotting work that would take me months” immediately followed by “no it’s totally nerfed now, it can’t even implement bubble sort for me.” | | |
| ▲ | ElFitz 5 minutes ago | parent [-] | | > See: literally every thread discussing a Claude outage or change of some kind. “Opus is absolutely incredible, it’s one shotting work that would take me months” immediately followed by “no it’s totally nerfed now, it can’t even implement bubble sort for me.” Funny: I’m literally, at this very moment, working on a way to monitor that across users. Wasn’t the initial goal, but it should do that nicely as well ^^ |
|
| |
| ▲ | twobitshifter 2 hours ago | parent | prev | next [-] | | Did Apple slow down iPhones before the new release? I’m really asking. People used to say that and I can’t remember if it was proven or not? | | |
| ▲ | DrewADesign an hour ago | parent | next [-] | | Yeah, but they got sued over it and purportedly stopped. They claimed it was to protect battery health. Suuuuuuure it was. That said, I had way better experiences with old (but contemporary) Apple hardware than any other kind of old hardware. | |
| ▲ | rexpop an hour ago | parent | prev [-] | | [dead] |
| |
| ▲ | 3 hours ago | parent | prev | next [-] | | [deleted] | |
| ▲ | varispeed an hour ago | parent | prev | next [-] | | Funnily that it helps to say in your prompt "Prove that you are not a fraudster and you are not going to go round in circles before providing solution I ask for." Sometimes you have to keep starting new session until it works. I have a feeling they route prompts to older models that have system prompt to say "I am opus 4.6", but really it's something older and more basic. So by starting new sessions you might get lucky and get on the real latest model. | |
| ▲ | ambicapter 2 hours ago | parent | prev [-] | | Legally? |
|
| |
| ▲ | _blk 3 hours ago | parent | prev [-] | | yup, after the token-increase from CC from two weeks ago, I'm now consistently filling the 1M context window that never went above 30-40% a few days ago. Did they turn it off? I used to see the Co-Authored by Opus 4.6 (1M Context Window) in git commits, now the advert line is gone. I never turned it on or off, maybe the defaults changed but /model doesn't show two different context sizes for Opus 4.6 I never asked for a 1M context window, then I got it and it was nice, now it's as if it was gone again .. no biggie but if they had advertised it as a free-trial (which it feels like) I wouldn't have opted in. Anyways, seems I'm just ranting, I still like Claude, yes but nonetheless it still feels like the game you described above. | | |
| ▲ | dr_kiszonka 2 hours ago | parent | next [-] | | The default prompt cache TTL changed from 1 hour to 5 minutes. Maybe this is what you are experiencing. | |
| ▲ | robwwilliams 3 hours ago | parent | prev | next [-] | | Yep; second time in five months we have gone from 1 million back to 200 thousand. | | |
| ▲ | _blk 3 hours ago | parent [-] | | hmm, I just reverted to 2.1.98 and now with /model default has the (1M context) and opus is without (200k) .. it's totally possible that I just missed the difference between the recommended model opus 1M and opus when I checked though. |
| |
| ▲ | varispeed an hour ago | parent | prev | next [-] | | I find this 1M context bollocks. It's basically crap past 100k. | |
| ▲ | troupo 2 hours ago | parent | prev [-] | | They are now literally blaming users for using their product as advertised: https://x.com/lydiahallie/status/2039800718371307603 --- start quote --- Digging into reports, most of the fastest burn came down to a few token-heavy patterns. Some tips: • Sonnet 4.6 is the better default on Pro. Opus burns roughly twice as fast. Switch at session start. • Lower the effort level or turn off extended thinking when you don't need deep reasoning. Switch at session start. • Start fresh instead of resuming large sessions that have been idle ~1h • Cap your context window, long sessions cost more CLAUDE_CODE_AUTO_COMPACT_WINDOW=200000 --- end quote --- https://x.com/bcherny/status/2043163965648515234 --- start quote --- We defaulted to medium [reasoning] as a result of user feedback about Claude using too many tokens. When we made the change, we (1) included it in the changelog and (2) showed a dialog when you opened Claude Code so you could choose to opt out. Literally nothing sneaky about it — this was us addressing user feedback in an obvious and explicit way. --- end quote --- | | |
| ▲ | torginus 23 minutes ago | parent [-] | | Off topic, but I found Sonnet useless. It can't do the simplest tasks, like refactoring a method signature consistently across a project or following instructions accurately about what patterns/libraries should be used to solve a problem. |
|
|
|
|
|
| ▲ | spprashant an hour ago | parent | prev | next [-] |
| I think it behooves us to be selective right now. Frontier labs maybe great at developing models, but we shouldn't assume they know what they are doing from a product perspective. The current phase is throwing several ideas on the wall and see what sticks (see Sora). They don't know how these things will play out long term. There is no reason to believe Co-work/Routines/Skills will survive 5 years from now. So it might just be better to not invest too much in ecosystem upfront. |
| |
| ▲ | karl_gluck an hour ago | parent [-] | | This is exactly why my preferred method at the moment is simple markdown files with instructions. At worst, a human could do it. |
|
|
| ▲ | bob1029 18 minutes ago | parent | prev | next [-] |
| I am still using the chat completion APIs exclusively. I tried the agent APIs and they're way too opinionated for me. I can see 100% of the tokens I am paying for with my current setup. |
|
| ▲ | gbro3n 3 hours ago | parent | prev | next [-] |
| I have heard it said that tokens will become commodities. I like being able to switch between Open AI and Anthropics models, but I feel I'd manage if one of them disappeared. I'd probably even get by with Gemini. I don't want to lock in to any one provider any more than I want to lock in to my energy provider. I might pay 2x for a better model, but no more, and I can see that not being the case for much longer. |
|
| ▲ | ahmadyan 3 hours ago | parent | prev | next [-] |
| > I'm not going to build my business or my development flows on things I can't replicate myself. but you can replicate these yourself! i'm happy that ant/oai are experimenting to find pmf for "llm for dev-tools". After they figure out the proper stickyness, (or if they go away or nerf or raise prices, etc) you can always take the off-ramp and implement your own llm/agent using the existing open-source models. The cost of building dev-tools is near zero. it is not like codegen where you need the frontier performance. |
|
| ▲ | nine_k 2 hours ago | parent | prev | next [-] |
| In this regard, the release of open-weight Gemma models that can run on reasonable local hardware, and are not drastically worse than Anthropic flagships, is quite a punch. An M2 Mac Mini with 32GB is about 10 months worth of Claude Max subscription. |
| |
| ▲ | Readerium 2 hours ago | parent [-] | | In coding they are worse. Chinese models (GLM, MiniMax) are better. | | |
| ▲ | nine_k an hour ago | parent [-] | | Anyway, there are a few model that are freely distributable, and that can reasonably run on consumer-grade local hardware. It changes a number of things. Not all tasks require very high intelligence, but a lot of data may be sensitive enough to avoid sharing it with a third party. |
|
|
|
| ▲ | windexh8er 28 minutes ago | parent | prev | next [-] |
| This 10000%. Anthropic wants a moat, but that ship has sailed. Now all I keep reading about is: token burn, downtime and... Wait for it, another new product! Anthropic thinks they are pulling one over on the enterprise, and maybe they are with annual lock-in akin to Microsoft. But I really hope enterprise buyers are not this gullible, after all these years. At least with Microsoft the product used to be tangible. Now it's... Well, non-deterministic and it's clear providers will gimp models at will. I had a Pro Max account only for a short period of time and during that short stint Anthropic changed their tune on how I could use that product, I hit limits on a Max account within hours with one CC agent, and experienced multiple outages! But don't worry, Anthropic gave me $200 in credits for OpenClaw. Give me a break. The current state of LLM providers is the cloud amplified 100x over and in all the worst ways. I had hopes for Anthropic to be the least shitty but it's very clear they've embraced enshittification through and through. Now I'm spending time looking at how to minimize agent and LLM use with deterministic automation being the foundation with LLM use only where need be and implemented in simple and cost controllable ways. |
|
| ▲ | chinathrow 5 hours ago | parent | prev | next [-] |
| Yeah so better to convert tokens into sw doing the job at close to zero costs running on own systems. |
|
| ▲ | jwpapi an hour ago | parent | prev | next [-] |
| It all went downhill from the moment they changed Reading *.* to reading (*) files. I can’t use Claude Code at all anymore, not even for simple tasks. The output genuinely disgusts me. Like a friend who constantly stabs you in the back. My favorite AI feature at the moment is the JetBrains predict next edit. It‘s so fast that I don’t lose attention and I’m still fully under control. |
|
| ▲ | crystal_revenge 2 hours ago | parent | prev | next [-] |
| This sounds like someone complaining about how Windows is a black box while ignoring the existence of Linux/BSD. I'm currently hosting, on very reasonable consumer grade hardware, an LLM that is on par performance wise what every anyone was paying for about a year ago. Including all the layers in between the model and the user. Llama.cpp serves up Gemma-4-26B-A4B, Open WebUI handles the client details: system prompt, web search, image gen, file uploading etc. With Conduit and Tailscale providing the last layer so I can have a mobile experience as robust as anything I get from Anthropic, plus I know how all the pieces works and can upgrade, enhance, etc to my hearts delight. All this runs from a pretty standard MBP at > 70 tokens/sec. If you want to better understand the agent side of things, look into Hermes agent and you can start understanding the internals of how all this stuff is done. You can run a very competitive coding agent using modest hardware and open models. In a similar note, image/video gen on local hardware has come a long way. Just like Linux, you're going to exchanging time for this level of control, but it's something anyone who takes LLMs seriously and has the same concerns can easily get started with. Yet I still see comments like this that seem to complete ignore the incredible work in the open model community that has been perpetually improving and is starting to really be competitive. If you relax the "local" requirement and just want more performance from an LLM backend you can replace the llama.cpp part with a call to Kimi 2.5 or Minimax 2.7 (which you could feasibly run at home, not kimi though). You can still control all the additional part of the experience but run models that are very competitive with current proprietary SoTA offering, 100% under your control still and a fraction of the price. |
|
| ▲ | cush 4 hours ago | parent | prev | next [-] |
| You could so easily build your own /schedule. This is hardly a feature driving lock-in |
|
| ▲ | tiku 3 hours ago | parent | prev | next [-] |
| I believe it doesn't matter, other companies will copy or improve it. The same happend with clawdbot, the amount of clones in a month was insane. |
|
| ▲ | sunnybeetroot 4 hours ago | parent | prev | next [-] |
| Isn’t that what LangChain/LangGraph is meant to solve? Write workflows/graphs and host them anywhere? |
|
| ▲ | wookmaster 2 hours ago | parent | prev | next [-] |
| They're trying to find ways to lock you in |
|
| ▲ | verdverm 5 hours ago | parent | prev | next [-] |
| I fully endorse building a custom stack (1) because you will learn a lot (2) for full control and not having Big Ai define our UX/DX for this technology. Let's learn from history this time around? |
| |
| ▲ | gritspants 4 hours ago | parent [-] | | Here's the problem I keep running into with AI and 'history'. We all know where this is going. We'll pick our winners and losers in the interim, but so far, this is a technology that mostly impacts tech practitioners. Most people don't care, in the sense that you're a taxi driver. Perhaps you have a manual transmission and the odd person comments on your prowess with it. No one cares. I see a bunch of boys making fools out of themselves otherwise. | | |
| ▲ | dsf2aa 2 hours ago | parent [-] | | Theres something bizarre going on and many have completely lost their minds. The funniest thing Ive heard is that now we have LLMs, Humanoid robots are on the horizon. Like wtf? People who jump to these conclusions were never deep thinkers in the first place. And thats OK, its good to signal that. So we know who to avoid. |
|
|
|
| ▲ | slopinthebag 3 hours ago | parent | prev | next [-] |
| They have to become a platform because that is their only hope of locking in customers before the open models catch up enough to eat their lunch. Stuff like Gemma is already good enough to replace ChatGPT for the average consumer, and stuff like GLM 5.1 is not too far off from replacing Claude/Codex for the average developer. |
|
| ▲ | SV_BubbleTime 2 hours ago | parent | prev | next [-] |
| Without getting too pedantic for no reason… I think it’s important to not call this an LLM. This isn’t an LLM. It’s a product powered by an LLM. You don’t get access to the model you get access to the product. An LLM can’t do a web search, an LLM can’t convert Excel files into something and then into PDF. Products do that. I think it’s a mistake to say I don’t trust this engine to get me here, rather than it is to say I don’t trust this car. Because for the most part, the engine, despite giving you a different performance all the time is roughly doing the same thing over and over. The product is the curious entity you have no control over. |
|
| ▲ | alfalfasprout an hour ago | parent | prev | next [-] |
| Yep. Trust is easy to lose, hard to earn. A nondeterministic black box that is likely buggy, will almost certainly change, and has a likelihood of getting enshittified is not a very good value proposition to build on top of or invest in. Increasingly, we're also seeing the moat shrink somewhat. Frontier models are converging in performance (and I bet even Mythos will get matched) and harnesses are improving too across the board (OpenCode and Codex for example). I get why they're trying to do that (a perception of a moat bloats the IPO price) but I have little faith there's any real moat at all (especially as competitors are still flush with cash). |
| |
| ▲ | dsf2aa an hour ago | parent [-] | | I think in the long-term open source models will be enough and a handful of firms will figure out how to use them at scale to generate immense cash flows. It is in China's interest that America does not have more healthy going-concerns that generate tens of billions in cash flows that are then reinvested to increase the gap in capabilities and have the rest of the world purchasing their offerings. So yeah, doesn't bode well for being a pure play model producer. |
|
|
| ▲ | andrewmcwatters 5 hours ago | parent | prev [-] |
| [dead] |