| ▲ | 9dev 20 hours ago | ||||||||||||||||||||||||||||||||||
I always wonder how people make qualitative statements like this. There are so many variables! Is it my prompt? The task? The specific model version? A good or bad branch out of the non-deterministic solution space? Like, do you run a proper experiment where you hand the same task to multiple models several times and compare the results? Not snark by the way, I’m asking in earnest how you pick one model over another. | |||||||||||||||||||||||||||||||||||
| ▲ | embedding-shape 20 hours ago | parent | next [-] | ||||||||||||||||||||||||||||||||||
> Like, do you run a proper experiment where you hand the same task to multiple models several times and compare the results? This is what I do. I have a little TUI that fires off Claude Code, Codex, Gemini, Qwen Coder and AMP in separate containers for most task I do (although I've started to use AMP less and less), and either returns the last message of what they replied and/or a git diff of what exactly they did. Then I compare them side by side. If all of them got something wrong, I update the prompt, fire them off again. Always starting from zero, and always include the full context of what you're doing with the first message, they're all non-interactive sessions. Sometimes I do 3x Codex instead of different agents, just to double-check that all of them would do the same thing. If they go off and do different things from each other, I know the initial prompt isn't specific/strict enough, and again iterate. | |||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||
| ▲ | energy123 17 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||||||||
I have sent the same prompt to GPT-5.2 Thinking and Gemini 3.0 Pro many times because I subscribe to both. GPT-5.2 Thinking (with extended thinking selected) is significantly better in my testing on software problems with 40k context. I attribute this to thinking time, with GPT-5.2 Thinking I can coax 5 minutes+ of thinking time but with Gemini 3.0 Pro it only gives me about 30 seconds. The main problem with the Plus sub in ChatGPT is you can't send more than 46k tokens in a single prompt, and attaching files doesn't help either because the VM blocks the model from accessing the attachments if there's ~46k tokens already in the context. | |||||||||||||||||||||||||||||||||||
| ▲ | enraged_camel 20 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||||||||
Last night I gave one of the flaky tests in our test suite to three different models, using the exact same prompt. Gemini 3 and Gemini 3 Flash identified the root cause and nailed the fix. GPT 5.1 Codex misdiagnosed the issue and attempted a weird fix despite my prompt saying “don’t write code, simply investigate.” I run these tests regularly, and Codex has not impressed me. Not even once. At best it’s on par, but most of the time it just fails miserably. Languages: JavaScript, Elixir, Python | |||||||||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||||||||