▲ | simonw 3 days ago | |||||||
Right - that's more or less the idea behind https://simonwillison.net/2023/Apr/25/dual-llm-pattern/ and the DeepMind CaMeL paper: https://simonwillison.net/2025/Apr/11/camel/ The challenge is that you have to implement really good taint tracking (as seen in old school Perl) - you need to make sure that the output of a model that was exposed to untrusted data never gets fed into some other model that has access potentially harmful tool calls. I think that is possible to build, but I haven't seen any convincing implementation of the pattern yet. Hopefully soon! | ||||||||
▲ | tptacek 3 days ago | parent [-] | |||||||
So, we've surfaced a disagreement, because I don't think you need something like taint tracking. I think the security boundary between an LLM context that takes untrusted data (from, e.g., tickets) and a sensitive context (that can, e.g., make database queries) is essentially no different than the boundary between the GET/POST args in a web app and a SQL query. It's not a trivial boundary, but it's one we have a very good handle on. | ||||||||
|