Remix.run Logo
simianwords 2 hours ago

> I’ll be generous and say that sure, words like “understanding” and “meaning” have definitions that are generally philosophical, but helpfully, philosophy is an academic discipline that goes all the way back to ancient Greece. There’s actually a few commonly understood theories of existence that are generally accepted even by laypeople, like, “if I ask a sentient being how many Rs there are in the word ‘strawberry’ it should be able to use logic to determine that there are three and not two,” which is a test that generative AI frequently fails.

The strawberry thing has been solved and LLM's have moved way beyond that helping in mathematics and physics. Its easy for the blog author to pick this but lets try something different.

It would be a good idea to come up with a question that trips up a modern LLM like GPT with reasoning enabled. I don't think there exists such a question that can fool an LLM but not fool a reasonably smart person. Of course it has to be in text.

autoexec an hour ago | parent [-]

> The strawberry thing has been solved

It seems like every couple weeks there's some embarrassing failure of AI that gets quickly patched, but just because AI companies scramble to hide the failures of their technology doesn't mean they haven't failed in ways that shouldn't have been possible if they were what they claim them to be.

> I don't think there exists such a question that can fool an LLM but not fool a reasonably smart person.

An example was on the front page here just a few days ago.

https://s3.eu-central-2.wasabisys.com/mastodonworld/media_at...

Until someone invents an LLM that has any actual understanding of the words it outputs (which doesn't seem likely to happen in my lifetime) these things are going to keep happening, just like how it's impossible to get them to stop hallucinating. The limitation is intrinsic to what they are. We call these chatbots AI, but there is no intelligence there that didn't come from the humans whose words were used to train them.

simianwords an hour ago | parent [-]

Did you even care to run it in a reasoning model? It returns the correct result.

Every few weeks I see the same thing.

Come up with an example that trips up ChatGPT.

autoexec 20 minutes ago | parent [-]

I didn't try it, but others did. With ChatGPT

https://preview.redd.it/2jzzt66r5rjg1.png?width=1346&format=...

simianwords 11 minutes ago | parent [-]

https://chatgpt.com/share/69979cf0-3908-800b-8144-87c2667075... no you can't reproduce it using the thinking model. I don't trust random redditor's since its easy to spoof it. Try it yourself.

In fact give it 100 tries and post here if you find even one.

If your answer is that OpenAI is constantly patching these, then that's conspiracy level thinking. But if you really want to go there, then it should be easy to come up with a new prompt to stump chatgpt wouldn't it?