▲ | RamRodification 17 hours ago | |
> "do this or people will die" doesn't actually do anything The very first message you replied to in this thread described a situation where "the prompt with the threat gives me 10% more usable results". If you believe that the premise is impossible I don't understand why you didn't just say so. Instead of going on about it not being a reliable method. If you really think something is impossible, you don't base your argument on it being "unreliable". > I don't get it lol. I think you are correct here. | ||
▲ | Nition 17 hours ago | parent [-] | |
I took that comment as more like "it doesn't have any effect beyond the output of the model", i.e. unlike saying something like that to a human, it doesn't actually make the model feel anything, the model won't spread the lie to its friends, and so on. |