▲ | gowld 4 days ago | |
"jailbreaking" seems a silly term for "I told the LLM two unrelated things, and the response was relevant to only one of my comments, or a mixture of both." It's not the LLM's fault that the human said something that the LLM understands better than the human :-) |