▲ | Show HN: Thought Forgery, a new technique for jailbreaking LLMs | |||||||
2 points by UltraZartrex 10 hours ago | 7 comments | ||||||||
Hi HN, I'm an independent security researcher and wanted to share a new vulnerability I've discovered. My account is too new to submit the direct link, so I'm making a text post instead. The technique is called "Thought Forgery" (CoT Injection). It works by forging the AI's internal monologue, which acts as a universal amplifier for other jailbreaks. I've confirmed it works on the latest models from Google, Anthropic, OpenAI, etc. I'd be happy to share the link to the full technical write-up on GitHub in the comments if anyone is interested. | ||||||||
▲ | ndgold 3 hours ago | parent | next [-] | |||||||
This is well known | ||||||||
| ||||||||
▲ | tjopies 9 hours ago | parent | prev | next [-] | |||||||
Please do post your write up this is interesting but pretty vague frankly | ||||||||
| ||||||||
▲ | alexander2002 9 hours ago | parent | prev | next [-] | |||||||
sure | ||||||||
| ||||||||
▲ | UltraZartrex 10 hours ago | parent | prev [-] | |||||||
[dead] |