▲ | enether 2 days ago | |
there was one post that detailed how those OpenAI models hallucinate and double down on thier mistakes by "lying" - it speculated on a bunch of interesting reasons why this may be the case recommended read - https://transluce.org/investigating-o3-truthfulness I wonder if this is what's causing it to do badly in these cases |