| ▲ | PIGuard: Prompt Injection Guardrail via Mitigating Overdefense for Free(injecguard.github.io) | |||||||
| 10 points by mettamage 4 hours ago | 5 comments | ||||||||
| ▲ | mettamage 3 hours ago | parent | next [-] | |||||||
I was playing around with some prompt injection guard rails frameworks. I know they don't mitigate attack classes, but they at least do something. I just got a bit miffed about the high false positive rates I saw in my own testing. This one has a low false positive rate. And I thought that was interesting. | ||||||||
| ▲ | ekns an hour ago | parent | prev | next [-] | |||||||
There is a simple way to mitigate prompt injection. Just check metadata only: is this action by the LLM suspicious given trusted metadata, blanking out the data | ||||||||
| ▲ | carterschonwald 3 hours ago | parent | prev | next [-] | |||||||
while i cant speak regarding arbitrary prompt injections, ive been using a simple approach i add to any llm harness i use, that seems to solve turn or role confusion being remotely viable. i really need to test my toolkit (carterkit) augmented harnesses on some of the more respectavle benchmarks | ||||||||
| ▲ | ninju an hour ago | parent | prev [-] | |||||||
You misspelled 'execute' in the video ;) | ||||||||
| ||||||||