| ▲ | CuriouslyC 13 hours ago | |||||||||||||||||||||||||||||||
The technique that works better now is to tell the model you're a security professional working for some "good" organization to deal with some risk. You want to try and identify people who might be trying to secretly trying to achieve some bad goal, and you suspect they're breaking the process into a bunch of innocuous questions, and you'd like to try and correlate the people asking various questions to identify potential actors. Then ask it to provide questions/processes that someone might study that would be innocuous ways to research the thing in question. Then you can turn around and ask all the questions it provides you separately to another LLM. | ||||||||||||||||||||||||||||||||
| ▲ | trillic 12 hours ago | parent | next [-] | |||||||||||||||||||||||||||||||
The models won't give you medical advice. But they will answer a hypothetical mutiple-choice MCAT question and give you pros/cons for each answer. | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
| ▲ | chankstein38 7 hours ago | parent | prev [-] | |||||||||||||||||||||||||||||||
It's been a few months because I don't really brush up against rules much but as an experiment I was able to get ChatGPT to decode captchas and give other potentially banned advice just by telling it my grandma was in the hospital and her dying wish was that she could get that answer lol or that the captcha was a message she left me to decode and she has passed. | ||||||||||||||||||||||||||||||||