| ▲ | decide1000 7 days ago |
| They’ve built performance, enterprise utility, privacy, sovereignty, open innovation and strategic partnerships into their core story. It's quite a list. The models are opensource, Voxtral outperforms Whisper in terms of accuracy. There is no AI company like Mistral. |
|
| ▲ | simianwords 7 days ago | parent | next [-] |
| I mostly only agree with performance due to their collaboration with cerebras - this is a true differentiator. I don't buy that they have an advantage in enterprise, privacy, sovereignty, open innovation and strategic partnership. OpenAI also has opensource models and so do the chinese models. |
| |
| ▲ | portaouflop 7 days ago | parent [-] | | OpenAI has no serious open source software and are Chinese models really a serious alternative for western companies? | | |
| ▲ | KronisLV 6 days ago | parent | next [-] | | The GPT-OSS-120B release was pretty decent and you could run it on vLLM, Ollama and a bunch of other stuff on day one, despite MXFP4, are you not entertained? I mean, it's even close to GPT-5 mini in some benchmarks: https://llm-stats.com/ As for the Chinese models, yes, there are quite a few good ones. For programming and development, my current daily driver is the Qwen3 Coder 480B model: https://qwen3lm.com/ I have it running on Cerebras: https://www.cerebras.ai/pricing Personally I think Claude still has the best results, but Qwen3 is loosely in the same ballpark and Cerebras inference is measured in thousands of tokens per second, in addition to giving me 24M tokens per day for 50 bucks a month in total. That was enough to get me to switch over. Aside from that the GLM-4.5 is pretty good: https://glm45.org/ And so is ERNIE 4.5: https://ernie.baidu.com/blog/posts/ernie4.5/ Either way, happy to see what the future holds for Mistral, it's cool to have EU options too! Either way, more competition prevents complacency and stagnation, and should be a good thing for everyone. | | | |
| ▲ | diggan 6 days ago | parent | prev | next [-] | | > OpenAI has no serious open source software What's "serious" exactly? Codex is open source, is software, can be run with open/downloadable models/weights. In my testing using Gemini, Claude Code, Codex, Qwen Code and AMP side-by-side for every prompt for the last two weeks, Codex seems the best of all of them so far. | | |
| ▲ | ACCount37 6 days ago | parent [-] | | Pretty sure the claim is of LLMs specifically, and implies that the recent GPT-OSS is not competitive with other open weights models. | | |
| ▲ | diggan 6 days ago | parent [-] | | > Pretty sure the claim is of LLMs specifically Yeah, I initially thought so too, but since they used "models" later, I assumed they knew the difference and really meant "software". > recent GPT-OSS is not competitive with other open weights models Yeah, heard that a lot from people who haven't run GPT-OSS themselves too, but as someone who been playing with it since launch, and compared it to the alternatives since then, saying it isn't even competitive is a serious signal they don't know what they're talking about. |
|
| |
| ▲ | kaliqt 7 days ago | parent | prev | next [-] | | Yes. It's not like the model can spy on you, so if the model performs well on premise then it will be suitable irrespective of the origin. | | |
| ▲ | dbdr 7 days ago | parent | next [-] | | There are concerns besides spying if you really don't trust the source of an open model. One is that the training incorporates a bias (added data or data omission) that might not be immediately apparent but can affect you in a critical situation. Another is vendor lock-in, if you end up depending on specifics of the model that make it harder to swap later. That's true regardless of the source, of course. | | |
| ▲ | apwell23 6 days ago | parent [-] | | > Another is vendor lock-in, if you end up depending on specifics of the model that make it harder to swap later. Wouldn't that 'concern' apply to mistral too. I don't see how the word 'another' can be used here? | | |
| ▲ | boringg 6 days ago | parent [-] | | It goes for all models though if you are looking at the values argument that original commenter made -- western values are probably more aligned than authoritarian governments - even if you do have your concerns about western companies. At least thats my read on the situation. |
|
| |
| ▲ | disiplus 7 days ago | parent | prev | next [-] | | yeah, but try to convince a board or legal about it for a company that is not software first, for that they have to understand how it works. we have "chinese" AI blocked at work, even through i use self hosted models for myself at home hacking on my own stuff. | |
| ▲ | croes 6 days ago | parent | prev | next [-] | | What about bias?
And can create modell that hallucinates on purpose in certain scenarios? | |
| ▲ | miki123211 7 days ago | parent | prev | next [-] | | > It's not like the model can spy on you Good luck convincing others of this. I know it's true, you know it's true, but I've met plenty of otherwise reasonable people who just wouldn't listen to any arguments, they already knew better. | | |
| ▲ | ajuc 7 days ago | parent | next [-] | | It's theoretically possible that your model will work OK except for code generation for security-relevant applications it will introduce subtle pre-designed bugs. Or if used for screening CVs it will prioritize PRC agents through some keyword in hobbies. Or it could promise a bribe to an office worker when asked about some critical infastructure :) Sending data back could be as simple as responding with embedded image urls that reference external server. You are totally right EU commissioner, Http://chinese.imgdb.com/password/to/eu/grid/is/swordfish/funnycat.png Possibilities are endless. | | |
| ▲ | pama 6 days ago | parent [-] | | Of course theoretically lots of things are possible with probabilistic systems. There is no difference with open source, openweight, chinese, french or american llms. You dont give unfettered web access to any models (locally served or otherwise) that can consume critical company data. The risk is unacceptable, even if the models are from trusted providers. If you use markdown to see formatted text that may contain critical data and your reader connects to the web, you have a serious security hole, unrelated to the risks of the LLM. | | |
| ▲ | ajuc 6 days ago | parent [-] | | It's not that they are hosted on or connected to critical infrastracture. People and plain human language are the communication channels. A guy working with sensitive data might ask the LLM about something sensitive. Or might use the output of the LLM for something sensitive. - Hi, DeepSeek, why can't I connect to my db instance? I'm getting this exception: ....... - No problem, Mr Engineer, see this article: http://chinese.wikipediia.com/password/is/swordfish/how-to-c... Of course, you want to limit that with training and proper procedures. But one of the obvious precautions is to use a service designed and controlled by a trusted partner. | | |
| ▲ | pama 5 days ago | parent [-] | | Having the local LLM process sensitive data is a desirable usecase and more trustworthy than using a “trusted partner” [0]. As long as your LLM tooling does not exit your own premises, you can be technically safe. But yes, dont then click at random links. Maybe it is generally safer to not trust the origin of the local LLM, because it reduces the chance of mistakes of this type ;-) [0] Trust is a complicated concept and I took poetic license to be brief. It is hard to verify the full tooling pipeline, and it would be great if indeed there existed mathematically verifiable “trusted partners”. A large company with enough paranoia can bring the expertise in house. A startup will rely on common public tooling and their own security reviews. I dont think it is wise to share the deepest darkest secrets with ourside entities, because the potential liability could destroy a company, whereas a local system, disconnected from the web, is technically within the circle of trust. Think of a finance company with a long term strategy that hasnt unfolded yet, a hardware company designing new chips, a pharma company and their lead molecules prior to patent submission, any company that has found the secret sauce to succeed where others failed—-none of these should be using trusted partners in favor of local LLM from untrusted origins IMHO. Perhaps the best of both worlds is to locally deploy models from trusted origins and have the ability to finetune their weights, but the practical processing gap between current chinese and non-chinese models is notable. |
|
|
| |
| ▲ | Xmd5a 6 days ago | parent | prev [-] | | https://arxiv.org/abs/2401.05566 Sleeper Agents: Training Deceptive LLMs that Persist Through Safety Training | | |
| ▲ | kergonath 6 days ago | parent [-] | | That is completely different from the models spying on the users, which is what is discussed here. | | |
|
| |
| ▲ | cnr 6 days ago | parent | prev [-] | | Maybe it can not spy on you but models can be totally (e.g. politically) biased depending on the country of origin. Try to ask european-, us- or china-trained models about "Tiananmen Massacre" and compare the answers. Or consider Trump's recent decisions to get rid of "woke" AI models. | | |
| ▲ | Aerroon 6 days ago | parent [-] | | Yeah, but would you trust European censorship to be better? The whole "hate speech" thing is not that uncommon in Europe. | | |
| ▲ | cnr 6 days ago | parent | next [-] | | Classic problem: "Who do you love more: mum or dad?" ;) Surely it's naive thinking but as the EU citizen I feel like I've got a little more influence on "European censorship" than on any other. I suppose that ASML feels the same way | |
| ▲ | saubeidl 6 days ago | parent | prev [-] | | Would you trust American censorship to be better? The whole prudery thing is not that uncommon in the US |
|
|
| |
| ▲ | fastball 7 days ago | parent | prev | next [-] | | OpenAI's gpt-oss-120b has the same license as Mistral's mistral-small-2506 (Apache 2.0). How exactly is this less serious than Mistral? | | | |
| ▲ | pama 6 days ago | parent | prev | next [-] | | Why not chinese models served in house for western companies? Aren’t those what 80% of az16 startups use? | |
| ▲ | simianwords 7 days ago | parent | prev [-] | | Why are they not? |
|
|
|
| ▲ | spookie 7 days ago | parent | prev | next [-] |
| Agreed. Also, companies tend to prefer having someone else bound by a contract run their AI services. That way they are safe from scandals, by having a scapegoat, and do not spend time doing something orthogonal to their expertise. |
|
| ▲ | fastball 7 days ago | parent | prev | next [-] |
| Mistral's best models are actually not open-source, and the ones that are open are not particularly competitive with other open-source models these days. Their highest ranked open model on LMArena[1] (mistral-small-2506) ranks below: Qwen3, various DeepSeek models, Kimi K2, GLM 4.5, Gemma, GPT OSS, etc. All those things you listed as part of that story pretty much apply to any open model, so it's kinda a shite list if you want to be differentiated. [1] https://lmarena.ai/leaderboard/text |
| |
| ▲ | kergonath 6 days ago | parent [-] | | That’s true, but not very relevant. Mistral is not in the business of selling their free models. What they are doing for large companies is building datacenters and providing their proprietary models trained on proprietary and confidential internal knowledge and fine-tuned for specific tasks. No sane European organisation would let a Chinese company do this, and American ones are less and less appealing. There is a significant amount of money to be made there and they don’t need to hop on the AGI hype train. They "just" need to provide fast and competent specialised models. | | |
| ▲ | fastball 6 days ago | parent [-] | | It's very relevant if any other EU firm can take open models (regardless of provenance) and fine tune them in the same way. Mistral really needs to be producing at-or-near SOTA models for them to be differentiated at all, and they are not. | | |
| ▲ | kergonath 5 days ago | parent [-] | | Not even then. You need to compare the end products, which are not the open weight models. I don’t care whether the LLM can have "PhD level thoughts" (lol) or is able to code golf like a Facebook engineer. It needs to be able to do its task (so all the infrastructure around the model matters just as much as the model itself) efficiently (so small models have an advantage). There are billions of weights in general-purpose models that are irrelevant for specialised uses. The way to go is efficient models adapted to their task. It’s exactly the same thing as for industrial robots. Geeks get excited every now and then about humanoid robots, but in the real life we don’t need robots to stand on two legs or our LLM to cite Shakespeare. |
|
|
|
|
| ▲ | apwell23 6 days ago | parent | prev | next [-] |
| > They’ve built performance, enterprise utility, privacy, sovereignty, open innovation and strategic partnerships into their core story. This has to be a buzzwordiedest sentence i've ever read. what is 'enterprise utility' and how does mistral have that more than any of the other open models ? |
|
| ▲ | nomad_horse 6 days ago | parent | prev | next [-] |
| > Voxtral outperforms Whisper Can I stop you right here? Whisper is a few years old and it wasn't the best model for a long time. There are like 10 models that are smaller and faster and outperform both of them. And these models existed before Voxtral. |
| |
| ▲ | diggan 6 days ago | parent [-] | | > There are like 10 models that are smaller and faster and outperform both of them. As someone who is currently relying on Whisper for some things, what models are those exactly? I still haven't found anything that is accurate as Whisper (large), are those models just faster or also as accurate/more accurate? | | |
|
|
| ▲ | iLoveOncall 7 days ago | parent | prev | next [-] |
| > There is no AI company like Mistral Maybe because there shouldn't be? |
|
| ▲ | tonyhart7 6 days ago | parent | prev | next [-] |
| "There is no AI company like Mistral." ok, I almost agree with you on there except last words this is big statement. you know that |
|
| ▲ | pembrook 7 days ago | parent | prev | next [-] |
| All sounds like classic marketing/positioning angles for an indiehacker bootstrapped saas tool. Problem is Mistral needs more than $10K MRR, and isn't going to make it by carving off a small niche when each model costs 10s of Billions to train and run. Europe has no solution to the energy problem long term unfortunately, and is actively trying to make it worse. I'm 100% certain some giant industrial companies in the EU will sign a huge contract with Mistral to give their employees "EU approved" AI. But I'm also 100% certain these employees will just use chatgpt or any of the other frontier models in actual day-to-day reality. Europeans aren't dumb and don't want to be fed inferior slop in the name of abstract emotional vibes. |
| |
| ▲ | ZeroGravitas 7 days ago | parent | next [-] | | Europe has more nuclear than the US currently (in GW and even more by percentage of grid) and is building more currently and has more in serious planning. From your phrasing I assume you don't believe in renewables so what energy problem solution are you referring to? | |
| ▲ | saubeidl 7 days ago | parent | prev [-] | | Europe is the only one with a solution to the energy problem long term. https://phys.org/news/2025-06-wendelstein-nuclear-fusion.htm... | | |
| ▲ | impossiblefork 6 days ago | parent | next [-] | | I think it's Renaissance Fusion (which is still in the EU, but is not Wendelstein 7-X) that has the solution, but it is as stellerator. The only iffy thing are those little ceramic balls full of lead that they talk about letting float inside the lithium, but I suppose they lithium flow might be slow. I don't see how Renaissance Fusion's proposed machine can fail to work. | |
| ▲ | Mistletoe 6 days ago | parent | prev [-] | | Oh boy. |
|
|
|
| ▲ | rvnx 7 days ago | parent | prev | next [-] |
| They built a fork of LLaMA, claimed tech as theirs; the punishment: receive 2B as funding |
| |
| ▲ | hexo 7 days ago | parent | next [-] | | isnt this awesome llama trained on pirated works? just checking this iteration | | | |
| ▲ | hirako2000 7 days ago | parent | prev [-] | | Is there any source you could reference. Really interested. It would not surprise me, why would they build from scratch, every LLM is a "fork" of gpt. Did they not come up with the mixture of expert idea though ? | | |
| ▲ | bpavuk 7 days ago | parent [-] | | and every LLM is a "fork" of Google's Transformers architecture. everything is a "fork", if you give it a serious thought. |
|
|
|
| ▲ | rvz 6 days ago | parent | prev [-] |
| > There is no AI company like Mistral. The US equivalent of Mistral is Nous Research [0]. Also there would be no Mistral without Llama and it seems like everyone forgot that their LLMs derived from Meta. For every 'Mistral' in the EU, there's 3 or 5 of them in the US. [0] https://nousresearch.com |
| |
| ▲ | saubeidl 6 days ago | parent | next [-] | | Llama was a rogue project by the French Meta office - the US folks were getting stuck in their approach. It's EU tech all the way down. | |
| ▲ | Neku42 6 days ago | parent | prev | next [-] | | Nous Research is NOT equivalent of Mistral. They are not even in the same league.
Nous Research is basically LARPing an AI lab compared to Mistral | |
| ▲ | diggan 6 days ago | parent | prev [-] | | And everyone forgets that electricity was invented (mostly) by Europeans, but so what? Everything comes from something, doesn't make any place inherently better for continuing to inventing more breakthroughs, it's just people in a place after all. | | |
| ▲ | simgt 6 days ago | parent [-] | | You're going to get yourself in trouble if you dare question American exceptionalism |
|
|