Remix.run Logo
Show HN: Agent Action Guard – AI agent action safety
2 points by praneeth-v 7 hours ago

Your agents can perform harmful actions without barriers. You do not know that yet. HarmActionBench experiments allowed AI agents to use tools based on harmful instructions, and the results are shocking. Even latest popular AI models, including GPT and Claude, scored very low. They have no barriers in performing harmful actions.

HarmActionsEval proves AI is not yet reliable enough for critical projects. Agent Action Guard blocks harmful actions. GitHub: https://github.com/Pro-GenAI/Agent-Action-Guard

I would love to discuss about possible use cases in your projects, and future directions. It helps to expand the dataset, model, and benchmark. Please discuss at https://github.com/Pro-GenAI/Agent-Action-Guard/discussions/....