| ▲ | Swizec 6 hours ago | |
Because the operator told the computer not to do something so the computer decided to do it. This is a huge security flaw in these newfangled AI-driven systems. Imagine if this was a "launch nukes" agent instead of a "write code" agent. | ||
| ▲ | verdverm 6 hours ago | parent [-] | |
It's not interesting because this is what they do, all the time, and why you don't give them weapons or other important things. They aren't smart, they aren't rationale, they cannot reliably follow instructions, which is why we add more turtles to the stack. Sharing and reading agent thinking text is boring. I had one go off on e one time, worse than the clawd bot who wrote that nasty blog after being rejected on GitHub. Did I share that session? No, because it's boring. I have 100s of these failed sessions, they are only interesting in aggregate for evals, which is why is save them. | ||