| ▲ | simoncion 3 hours ago | ||||||||||||||||
> you are not able to produce a _single_ counter example for my challenge? I have. For both your original challenge and your updated one. Consider: 1) AFAICT, there's no way to tell what version of the model was used to produce the output in a ChatGPT share link. 2) You don't appear to believe my assertions that aphyr is almost certainly paying for and using the latest version of the LLMs available, and that he's faithfully reporting his interactions with the LLMs. 3) Because of #2, I expect that you won't believe me if I report that I've more-or-less reproduced father_phi's results about the cup that's sealed on the top and open on the bottom on the very latest only-available-for-pay ChatGPT model. 3a) You might attempt to check my report, but I'd be shocked if you'd consider a failure to reproduce my results to be a significant strike against ChatGPT. I'd think it's more likely that you'd either call me a liar, or tell me that I must have had some setting wrong somewhere. 3b) Even if you told me to share the ChatGPT chat that proved my assertion, #1 -combined with your demeanor throughout this conversation- tells me that you'd almost certainly claim that I was using an inferior version of the model and was lying to you. | |||||||||||||||||
| ▲ | simianwords 3 hours ago | parent [-] | ||||||||||||||||
Haha ok. So still no example? The GPT shared link shows a "thought for" which indicates using the latest thinking model. You may try that. What you can do is this: submit a prompt that clearly makes GPT hallucinate. You may secretly use a worse model. You may use a system prompt that deliberately gives wrong answers. But I'm going to assume you won't go that far. We can leave it to the public to decide whether this is a legitimate counter example or not and whether it can really be reproduced. Shall we try that? I'm guessing you won't but worth a shot! | |||||||||||||||||
| |||||||||||||||||