| ▲ | rafaelmn 3 hours ago | |||||||||||||||||||||||||
GPT is shit at writing code. It's not dumb - extra high thinking is really good at catching stuff - but it's like letting a smart junior into your codebase - ignore all the conventions, surrounding context, just slop all over the place to get it working. Claude is just a level above in terms of editing code. | ||||||||||||||||||||||||||
| ▲ | sho_hn 2 hours ago | parent | next [-] | |||||||||||||||||||||||||
Very different experience for me. Codex 5.3+ on xhigh are the only models I've tried so far that write reasonably decent C++ (domains: desktop GUI, robotics, game engine dev, embedded stuff, general systems engineering-type codebases), and idiomatic code in languages not well-represented in training data, e.g. QML. One thing I like is explicitly that it knows better when to stop, instead of brute-forcing a solution by spamming bespoke helpers everywhere no rational dev would write that way. Not always, no, and it takes investment in good prompting/guardrails/plans/explicit test recipes for sure. I'm still on average better at programming in context than Codex 5.4, even if slower. But in terms of "task complexity I can entrust to a model and not be completely disappointed and annoyed", it scores the best so far. Saves a lot on review/iteration overhead. It's annoying, too, because I don't much like OpenAI as a company. (Background: 25 years of C++ etc.) | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | Jcampuzano2 3 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||
Not my experience. GPT 5.4 walks all over Claude from what I've worked with and its Claude that is the one willing to just go do unnecessary stuff that was never asked for or implement the more hacky solutions to things without a care for maintainability/readability. But I do not use extra high thinking unless its for code review. I sit at GPT 5.4 high 95% of the time. | ||||||||||||||||||||||||||
| ▲ | zarzavat 3 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||
Yes, it's becoming clear that OpenAI kinda sucks at alignment. GPT-5 can pass all the benchmarks but it just doesn't "feel good" like Claude or Gemini. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | leobuskin 3 hours ago | parent | prev | next [-] | |||||||||||||||||||||||||
And as a bonus: GPT is slow. I’m doing a lot of RE (IDA Pro + MCP), even when 5.4 gives a little bit better guesses (rarely, but happens) - it takes x2-x4 longer. So, it’s just easier to reiterate with Opus | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||
| ▲ | whalesalad 3 hours ago | parent | prev [-] | |||||||||||||||||||||||||
This has been my experience. With very very rigid constraints it does ok, but without them it will optimize expediency and getting it done at the expense of integrating with the broader system. | ||||||||||||||||||||||||||
| ||||||||||||||||||||||||||