Remix.run Logo
esafak 19 hours ago

Less than a year behind the SOTA, faster, and cheaper. I think Mistral is mounting a good recovery. I would not use it yet since it is not the best along any dimension that matters to me (I'm not EU-bound) but it is catching up. I think its closed source competitors are Haiku 4.5 and Gemini 3 Pro Fast (TBA) and whatever ridiculously-named light model OpenAI offers today (GPT 5.1 Codex Max Extra High Fast?)

kevin061 17 hours ago | parent | next [-]

The OpenAI thing is named Garlic.

(Surely they won't release it like that, right..?)

esafak 16 hours ago | parent [-]

TIL: https://garlicmodel.com/

That looks like the next flagship rather than the fast distillation, but thanks for sharing.

kevin061 16 hours ago | parent [-]

Lol, someone vibecoded an entire website for OpenAI's model, that's some dedication.

BoorishBears 12 hours ago | parent | next [-]

People have been doing this for literally every anticipated model release, and I presume skimming some amount of legitimate interest since their sites end up being top indexed until the actual model is released.

Google should be punishing these sites but presumably it's too narrow of a problem for them to care.

kevin061 12 hours ago | parent [-]

Black SEO in the age of LLMs

dmix 10 hours ago | parent [-]

It would need outbound links to be SEO

Or at least a profit model. I don't see either on that page but maybe I'm missing something

ewoodrich 8 hours ago | parent [-]

Every link in the "Legal" tree is a dead end redirecting back to the home page... strange thing to put together without any acknowledgement, unless they spam it on LLM adjacent subreddits for clout/karma?

ttul 10 hours ago | parent | prev [-]

"GPT, please make me a website about OpenAI's 'Garlic' model."

YetAnotherNick 16 hours ago | parent | prev [-]

No this is comparable to Deepseek-v3.2 even on their highlight task, with significantly worse general ability. And it's priced 5x of that.

esafak 15 hours ago | parent [-]

It's open source; the price is up to the provider, and I do not see any on openrouter yet. ̶G̶i̶v̶e̶n̶ ̶t̶h̶a̶t̶ ̶d̶e̶v̶s̶t̶r̶a̶l̶ ̶i̶s̶ ̶m̶u̶c̶h̶ ̶s̶m̶a̶l̶l̶e̶r̶,̶ ̶I̶ ̶c̶a̶n̶ ̶n̶o̶t̶ ̶i̶m̶a̶g̶i̶n̶e̶ ̶i̶t̶ ̶w̶i̶l̶l̶ ̶b̶e̶ ̶m̶o̶r̶e̶ ̶e̶x̶p̶e̶n̶s̶i̶v̶e̶,̶ ̶l̶e̶t̶ ̶a̶l̶o̶n̶e̶ ̶5̶x̶.̶ ̶I̶f̶ ̶a̶n̶y̶t̶h̶i̶n̶g̶ ̶D̶e̶e̶p̶S̶e̶e̶k̶ ̶w̶i̶l̶l̶ ̶b̶e̶ ̶5̶x̶ ̶t̶h̶e̶ ̶c̶o̶s̶t̶.̶

edit: Mea culpa. I missed the active vs dense difference.

NitpickLawyer 14 hours ago | parent | next [-]

> Given that devstral is much smaller, I can not imagine it will be more expensive

Devstral 2 is 123B dense. Deepseek is 37B Active. It will be slower and more expensive to run inference on this than dsv3. Especially considering that dsv3.2 has some goodies that make inference at higher context be more effective than their previous gen.

syntaxing 12 hours ago | parent [-]

Devstral is purely nonthinking too it’s very possible it uses less models (I don’t know how DS 3.2 nonthinking compares). It’s interesting because Qwen pretty much proved hybrid models work worse than fully separate models.

aimanbenbaha 11 hours ago | parent | prev | next [-]

Deepseek v3.2 is that cheap because its attention mechanism is ridiculously efficient.

esafak 9 hours ago | parent [-]

Yeah, DeepSeek Sparse Attention. Section 2: https://arxiv.org/abs/2512.02556

6 hours ago | parent | prev [-]
[deleted]