Remix.run Logo
keiferski 5 hours ago

Right now, 30 seconds ago, I asked ChatGPT to tell me about a book I found that was written in the 60s.

It made up the entire description. When I pointed this out, it apologized and then made up another description.

The idea that this is going to lead to superintelligence in a few years is absolutely nonsense.

i_think_so 3 hours ago | parent | next [-]

Is that because this book is obscure and no human has yet written a description that could be scraped?

hirvi74 5 hours ago | parent | prev [-]

The other day I asked Claude Opus 4.6 one of my favorite trivia pieces:

What plural English word for an animal shares no letters with its singular form? Collective nouns (flock, herd, school, etc.) don't count.

Claude responded with:

"The answer is geese -- the plural of cow."

Though, to be fair, in the next paragraph of the response, Claude stated the correct answer. So, it went off the rails a bit, but self-corrected at least. Nevertheless, I got a bit of a chuckle out of its confidence in its first answer.

I asked GPT 5.2 the same question and it nailed the answer flawlessly. I wouldn't extrapolate much about the model quality based on this answer, but I thought it was interesting still.

(For those curious, the answer is 'kine' (archaic plural for cow).

ileonichwiesz 2 hours ago | parent [-]

Of course it’s important to remember that the ability of an LLM to answer an obscure riddle like that has nothing to do with its reasoning abilities, but rather depends on whether the answer was included in its training dataset.