▲ | rcxdude 2 days ago | |
This is all possible, but an absolutely terrible idea from a security point of view, while prompt injection attacks are still a thing, and there's little evidence they will stop being a thing soon. | ||
▲ | pradn 12 hours ago | parent [-] | |
We can work toward closing security gaps with new technology, yes. It is necessary for large-scale adoption of LLM tech. |