Remix.run Logo
rcxdude 7 hours ago

The complication is that it doesn't work reliably. You can train an LLM with special tokens for delimiting different kinds of information (and indeed most non-'raw' LLMs have this in some form or another now), but they don't exactly isolate the concepts rigorously. It'll still follow instructions in 'user input' sometimes, and more often if that input is designed to manipulate the LLM in the right way.