▲ | epolanski 6 days ago | ||||||||||||||||
No, the problem is that HN is blind to the fact that there are multiple definitions of "best". It isn't just about "more powerful", it's also about "cheaper" or "faster". Mistral models are faster than anything out of US (bar Gemini Flash) and are cost competitive with them. For me, having to produce financial news in a short time span for tens of thousands of users speed and cost are important, and the fact that Opus 4.1 is "more intelligent" is worthless. That's like telling me that a Ryzen Threadripper with 64 cores is faster than than my raspberry pi for controlling the appliances in my kitchen. It's irrelevant when it's much more expensive and energy hungry. | |||||||||||||||||
▲ | BoorishBears 6 days ago | parent [-] | ||||||||||||||||
Pretty far off the mark. I've spent the last year building an AI product in a situation with really cut throat margins: I've post-trained every model Mistral has released in that time frame that was either open-weights or supported fine-tuning via Le Platforme (so I've gotten them at their absolute best case) Mistral's models are not competitive anymore, and haven't been for most of that time. Gemma 27b has better world knowledge, Deepseek obsoleted their dense models, Gemini Flash is faster and their models are not even close to cost competitive with it (shocking claim otherwise tbh). Mistral's platform is not fast (Mistral Medium is slower than Sonnet 4, which is just straight up insane!). Cerebras is fast, but there are both competitors offering similar speeds (Samba Nova and Groq), and other models that are faster on Cerebras (people really sleep on gpt-oss after the launch jitters) You're inventing a snowman with your analogy: their models are just irrelevant, and that's informed by using everything from dots.llm to Minimax-Text to Jamba (which is really underestimated btw, and not Chinese if sinophobia has a grip on your org) to Seed-OSS, in production. tl;dr: the only way to justify Mistral's models is in fact to reject the best solutions in any dimension that can be described as model performance. If you're still using them and it really isn't for non-performance reasons, I assume you're overindexing on benchmarks or behind on the last year or so of open-weight progress and would recommend actually trying some other offerings. | |||||||||||||||||
|