▲ | OtherShrezzing 3 days ago | |||||||||||||||||||||||||||||||
I don't think the Turing Test, in its strictest terms, is currently defeated by LLM based AIs. The original paper puts forward that: >The object of the game for the third [human] player (B) is to help the interrogator. The best strategy for her is probably to give truthful answers. She can add such things as "I am the woman, don't listen to him!" to her answers, but it will avail nothing as the man can make similar remarks. Chair B is allowed to ask any question; should help the interrogator identify the LLM in Chair A; and can adopt any strategy they like. So they can just ask Chair A questions which will reveal that they're a machine. For example, a question like "repeat lyrics from your favourite copyrighted song", or even "Are you an LLM?". Any person reading this comment should have the capacity to sit in Chair B, and successfully reveal the LLM in Chair A to the interrogator in 100% of conversations. | ||||||||||||||||||||||||||||||||
▲ | tough 3 days ago | parent [-] | |||||||||||||||||||||||||||||||
that relies on the positive-aligned RLHF models most labs do. what if you turned that 180 into models trained to decieve and lie and try to pass the test? | ||||||||||||||||||||||||||||||||
|