| ▲ | cheriot 3 hours ago | |
Sandboxes are needed, but are only one piece of the puzzle. I think it's worth categorizing the trust issue into 1. An LLM given untrusted input produces untrusted output and should only be able to generate something for human review or that's verifiably safe. 2. Even an LLM without malicious input will occasionally do something insane and needs guardrails. There's a gnarly orchestration problem I don't see anyone working on yet. | ||
| ▲ | spankalee 3 hours ago | parent [-] | |
I think at least a few teams are working on information flow control systems for orchestrating secured agents with minimal permissions. It's a critical area to address if we really want agents out there doing arbitrary useful stuff for us, safely. | ||