Remix.run Logo
amunozo 7 hours ago

I want to believe it's gonna be good, but after trying GPT-5.5 even the most advanced Chinese models seem depressing.

r0b05 6 hours ago | parent | next [-]

This is a French model sir

spwa4 6 hours ago | parent [-]

Évidemment

Funny detail: Google AI (the one they use in search) can't spell évidemment correctly.

baq 6 hours ago | parent [-]

What's French for 'goblin'...?

ako 6 hours ago | parent | prev | next [-]

Then you’ll be happy to learn it’s not Chinese

dotancohen 6 hours ago | parent [-]

GP is stating that the second best in the field, the Chinese, is so far behind the best in the field, GPT 5.5, that it is not even worth testing anything else.

amunozo 6 hours ago | parent | next [-]

Thanks for the translation, I did not express it very clearly. Anything that I try is so much worse.

Ritewut 6 hours ago | parent | prev [-]

Is GPT 5.5 the best in the field? I think Opus is still better despite Anthropic's recent stumbling.

manishsharan 6 hours ago | parent | prev | next [-]

I am not following this obsession with SOTA and benchmark rankings

I have been using DeepSeek and GLMnmodels with OpenCode and Codex and Claudr side by side.

I have not found the Chinese models lacking. I enjoy for coding and like to maintain full control of my codebade and deeply care about the GOF patterns. So I am very stringent in terms of what I want the LLM to code and how to code.

So from my perspective, they are all about the same.

amunozo 6 hours ago | parent [-]

That I agree with, but for more complex autonomous changes the differences are considerable. However, it seems that most models will reach the saturation time in which they will be useful for almost everything and the difference will be in more and more niche and specialized tasks.

lava_pidgeon 6 hours ago | parent | prev [-]

Honestly I depends on the context which this performance matters. Mistral is quiet cheap