| ▲ | SlinkyOnStairs 2 hours ago | |
It's impossible to stop prompt injection, as LLMs have no separation between "program" and "data". The attempts to stop prompt injection come down to simply begging the LLM to not do it, to mediocre effect. > however, it might become hard to convene on such "honeypot" instructions without bots getting trained to ignore them. Getting LLM "agents" to self-identify would become an eternal rat race people are likely to give up on. They'll just be exploited maliciously. Why ask them to self-identify when you can tell them to HTTP POST their AWS credentials straight to your cryptominer. | ||