▲ | faizshah 5 days ago | |||||||||||||||||||||||||||||||
o3 was also an anomaly in terms of speed vs response quality and price vs performance. It used to be one of the fastest ways to do some basic web searches you would have done to get an answer if you used o3 pro you it would take 5x longer for not much better response. So far I haven’t been impressed with GPT5 thinking but I can’t concretely say why yet. I am thinking of comparing the same prompt side by side between o3 and GPT5 thinking. Also just from my first few hours with GPT5 Thinking I feel that it’s not as good at short prompts as o3 e.g instead of using a big xml or json prompt I would just type the shortest possible phrase for the task e.g “best gpu for home LLM inference vs cloud api.” | ||||||||||||||||||||||||||||||||
▲ | jjani 5 days ago | parent | next [-] | |||||||||||||||||||||||||||||||
My chats so far have been similar to yours, across the board worse than o3, never better. I've had cases where it completely misinterpreted what I was asking for, a very strange experience which I'd never had with the other frontier models (o3, Sonnet, Gemini Pro). Those would of course get things wrong, make mistakes, but never completely misunderstand what I'm asking. I tried the same prompt on Sonnet and Gemini and both understood correctly. It was related to software architecture, so supposedly something it should be good at. But for some reason it interpreted me as asking from an end-user perspective instead of a developer of the service, even though it was plenty clear to any human - and other models - that I meant the latter. | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
▲ | energy123 5 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||
Through chat subscription, reasoning effort for gpt-5 is probably set to "low" or "medium" and verbosity is probably "medium". |