| ▲ | mpalmer 3 hours ago |
| > Claude Mythos Preview’s large increase in capabilities has led us to decide not to make it generally available. A month ago I might have believed this, now I assume that they know they can't handle the demand for the prices they're advertising. |
|
| ▲ | IceWreck 2 hours ago | parent | next [-] |
| Didn't OpenAI say something similar about GPT-3? Too dangerous to open source and then afew years later tehy were open sourcing gpt-oss because a bunch of oss labs were competing with their top models. |
| |
| ▲ | FeepingCreature 2 hours ago | parent | next [-] | | OpenAI didn't release GPT-2 initially because they were worried it would make it too easy to generate spam. Which it kinda did. | |
| ▲ | abroszka33 an hour ago | parent | prev [-] | | OpenAI said that GPT-5 was too dangerous to release... And look where we are now. It's mostly hype. |
|
|
| ▲ | wg0 3 hours ago | parent | prev | next [-] |
| That's for the investors basically. Scarcity and FOMO. |
| |
| ▲ | causal an hour ago | parent [-] | | *Until GPT-6 comes out, at which point Mythos will coincidentally be sufficiently safety-tested to release :) |
|
|
| ▲ | skippyboxedhero 3 hours ago | parent | prev | next [-] |
| GPT-2, o1, Opus...been here so many times. The reason they do this is because they know it works (and they seem to specifically employ credulous people who are prone to believe AGI is right around the corner). There haven't been significant innovations, the code generated is still not good but the hype cycle has to retrigger. I remember when OpenAI created the first thinking model with o1 and there were all these breathless posts on here hyperventilating about how the model had to be kept secret, how dangerous it was, etc. Fell for it again award. All thinking does is burn output tokens for accuracy, it is the AI getting high on its own supply, this isn't innovation but it was supposed to super AGI. Not serious. |
| |
| ▲ | chaos_emergent 2 hours ago | parent | next [-] | | > All thinking does is burn output tokens for accuracy “All that phenomenon X does is make a tradeoff of Y for Z” It sounds like you’re indignant about it being called thinking, that’s fine, but surely you can realize that the mechanism you’re criticizing actually works really well? | |
| ▲ | b65e8bee43c2ed0 3 hours ago | parent | prev | next [-] | | >I remember when OpenAI created the first thinking model with o1 and there were all these breathless posts on here hyperventilating about how the model had to be kept secret, how dangerous it was, etc. I've read that about Llama and Stable Diffusion. AI doomers are, and always have been, retarded. | |
| ▲ | simianwords 3 hours ago | parent | prev | next [-] | | Incredible that people still think like this. | | |
| ▲ | skippyboxedhero 3 hours ago | parent [-] | | You're completely right. | | |
| ▲ | simianwords 3 hours ago | parent [-] | | uhh the model found actual vulnerabilities in software that people use. either you believe that the vulnerabilities were not found or were not serious enough to warrant a more thoughtful release | | |
| ▲ | mlsu 2 hours ago | parent [-] | | So did GPT-4. https://arxiv.org/html/2402.06664v1 Like think carefully about this. Did they discover AGI? Or did a bunch of investors make a leveraged bet on them "discovering AGI" so they're doing absolutely anything they can to make it seem like this time it's brand new and different. If we're to believe Anthropic on these claims, we also have to just take it on faith, with absolutely no evidence, that they've made something so incredibly capable and so incredibly powerful that it cannot possibly be given to mere mortals. Conveniently, that's exactly the story that they are selling to investors. Like do you see the unreliable narrator dynamic here? | | |
| ▲ | mgfist an hour ago | parent | next [-] | | On the other hand I've gotten to use opus-4.6 and claude code and the quality is off the charts compared to 2023 when coding agents first hit the scene. And what you're saying is essentially "If they haven't created God, I'm not impressed". You don't think there's some middleground between those two? Also they just hit a $30B run-rate, I don't think they're that needy for new hype cycles. | |
| ▲ | simianwords 2 hours ago | parent | prev [-] | | I don't see the problem here. How would you have handled it differently? If you released this model as such without any safety concern, the vulnerabilities might be found by bad actors and used for wrong things. What do you find surprising here? | | |
| ▲ | mlsu an hour ago | parent [-] | | Vulnerabilities were found, probably a few by bad actors, when GPT4 was released. Every vulnerability found now is probably found with AI assistance at the very least. Should they have never released GPT4? Should we have believed claims that GPT4 was too dangerous for mere mortals to access? I believe openAI was making similar claims about how GPT4 was a step function and going to change white collar work forever when that model was released. The point is that this whole "the model is too powerful" schtick is a bunch of smoke and mirrors. It serves the valuation. | | |
| ▲ | simianwords an hour ago | parent [-] | | Its far more simple to believe that they are releasing it step by step. Release to trusted third parties first, get the easy vulnerabilities fixed, work on the alignment and then release to public. Do you don't believe that the vulnerabilities found by these agents are serious enough to warrant staggered release? |
|
|
|
|
|
| |
| ▲ | vonneumannstan 3 hours ago | parent | prev [-] | | Lol you haven't used a model since GPT2 is what it sounds like. | | |
| ▲ | skippyboxedhero 3 hours ago | parent [-] | | Just checked my subscription start date for Anthropic. September 2023, I believe before they announced public launch. Sorry kid. | | |
| ▲ | SyneRyder 2 hours ago | parent | next [-] | | Genuine question - if you don't think the models are improved or that the code is any good, why do you still have a subscription? You must see some value, or are you in a situation where you're required to test / use it, eg to report on it or required by employer? (I would disagree about the code, the benefits seem obvious to me. But I'm still curious why others would disagree, especially after actively using them for years.) | | |
| ▲ | skippyboxedhero 2 hours ago | parent [-] | | The assumption that the other person made was that I would only use it for coding. If you look through my other comments today, I suggest that they are useful for performing repetitive tasks i.e. checking lint on PR, etc. Also, can be used for throwaway code, very useful. I don't think the issue is with the model, it is with the implication that AGI is just around the corner and that is what is required for AI to be useful...which is not accurate. The more grey area is with agentic coding but my opinion (one that I didn't always hold) is that these workflows are a complete waste of time. The problem is: if all this is true then how does the CTO justify spending $1m/month on Anthropic (I work somewhere where this has happened, OpenAI got the earlier contract then Cursor Teams was added, now they are adding Anthropic...within 72 hours of the rollout, it was pulled back from non-engineering teams). I think companies will ask why they need to pay Anthropic to do a job they were doing without Anthropic six months ago. Also, the code is bad. This is something that is non-obvious to 95% of people who talk about AI online because they don't work in a team environment or manage legacy applications. If I interview somewhere and they are using agentic workflow, the codebase will be shit and the company will be unable to deliver. At most companies, the average developer is an idiot, giving them AI is like giving a monkey an AK-47 (I also say this as someone of middling competence, I have been the monkey with AK many times). You increase the ability to produce output without improving the ability to produce good output. That is the reality of coding in most jobs. AI isn't good enough to replace a competent human, it is fast enough to make an incompetent human dangerous. |
| |
| ▲ | vonneumannstan 3 hours ago | parent | prev [-] | | So you are doubly stupid, by not seeing any improvement in the models and also paying for models you believe are terrible? lol | | |
| ▲ | skippyboxedhero 2 hours ago | parent [-] | | That doesn't follow logically from what I said. You should ask your AI for help with this. You are in need of some artificial intelligence. |
|
|
|
|
|
| ▲ | b65e8bee43c2ed0 3 hours ago | parent | prev [-] |
| you would be a fool to believe it at any point in time. Amodei is anthropomorphic grease, even more so than Altman. Anthropic is burning through billions of VC cash. if this model was commercially viable, it would've been released yesterday. |
| |
| ▲ | landtuna 2 hours ago | parent [-] | | If there's limited hardware but ample cash, it doesn't make sense to sell compute-intensive services to the public while you're still trying to push the frontier of capability. | | |
| ▲ | b65e8bee43c2ed0 2 hours ago | parent [-] | | that's more or less what I'm saying. "Claude Mythos Preview’s large increase in capabilities has led us to decide not to make it generally available", translated from bullshit, means "It would've cost four digits per 1M tokens to run this model without severe quantization, and we think we'll make more money off our hardware with lighter models. Cool benchmarks though, right?" |
|
|