Remix.run Logo
groby_b 8 hours ago

> We have all of the tools to prevent these agentic security vulnerabilities,

We do? What is the tool to prevent prompt injection?

alienbaby 6 hours ago | parent | next [-]

The best I've heard is rewriting prompts as summaries before forwarding them to the underlying ai, but has it's own obvious shortcomings, and it's still possible. If harder. To get injection to work

lacunary 8 hours ago | parent | prev | next [-]

more AI - 60% of the time an additional layer of AI works every time

losthobbies 7 hours ago | parent | prev [-]

Sanitise input and LLM output.

chasd00 7 hours ago | parent [-]

> Sanitise input

i don't think you understand what you're up against. There's no way to tell the difference between input that is ok and that is not. Even when you think you have it a different form of the same input bypasses everything.

"> The prompts were kept semantically parallel to known risk queries but reformatted exclusively through verse." - this a prompt injection attack via a known attack written as a poem.

https://news.ycombinator.com/item?id=45991738

losthobbies 7 hours ago | parent [-]

That’s amazing.

If you cannot control what’s being input, then you need to check what the LLM is returning.

Either that or put it in a sandbox

danaris 6 hours ago | parent [-]

Or...

don't give it access to your data/production systems.

"Not using LLMs" is a solved problem.

losthobbies 5 hours ago | parent [-]

Yea agreed. Or use RBAC