| ▲ | simianwords 2 hours ago | |||||||||||||||||||||||||||||||||||||||||||
I agree that they hallucinate sometimes. I agree they bullshit sometimes. But the extent is way overblown. They basically don't bullshit ever under the constraints of 1. 2-3 pages of text context 2. GPT-5.4 thinking I don't think the spirit of the original article (not your comments to be fair) captured this, hence the challenge. I believe we are on the same page here. | ||||||||||||||||||||||||||||||||||||||||||||
| ▲ | camgunz an hour ago | parent [-] | |||||||||||||||||||||||||||||||||||||||||||
> I don't think the spirit of the original article (not your comments to be fair) captured this, hence the challenge. I believe we are on the same page here. No. GPT-5 has a 40% hallucination rate [0] on SimpleQA [1] without web searching. The SimpleQA questions meet your criteria of "2-3 pages of text content. Unless 5.4 + web searching erases that (I bet it doesn't!) these are bullshit machines. | ||||||||||||||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||||||||||||||