Remix.run Logo
Filligree 3 days ago

It's a really hard problem to solve!

You might think you can train the AI to do it in the usual fashion, by training on examples of the AI calling out errors, and agreeing with facts, and if you do that—and if the AI gets smart enough—then that should work.

If. You. Do. That.

Which you can't, because humans also make mistakes. Inevitably, there will be facts in the 'falsehood' set—and vice versa. Accordingly, the AI will not learn to tell the truth. What it will learn instead is to tell you what you want to hear.

Which is... approximately what we're seeing, isn't it? Though maybe not for that exact reason.

3 days ago | parent | next [-]
[deleted]
dchftcs 3 days ago | parent | prev [-]

The AI needs to be able to lookup data and facts and weigh them properly. Which is not easy for humans either; once you're indoctrinated in something, and you trust a bad data source over another, it's evidently very hard to correct course.