| ▲ | kloud 3 days ago | |
Also in the context of LLMs I think model weights themselves could be considered an untrusted input, because who knows what was in the training dataset. Even an innocent looking prompt could potentially trigger a harmful outcome. In that regard it reminds me of the CAP theorem, which also has three parts. However, in practice partitioning in distributed systems is given, so the choice is just between availability or consistency. So in the case of lethal trifecta it is either private data or external communication, but the leg between these two will always have some risk. | ||