| ▲ | whatisthiseven 4 hours ago | |
Worse, I have had it being sus of my own codebase when I tasked it with writing mundane code. Apparently if you include some trigger words it goes nuts. Still trying to narrow down which ones in particular. Here is some example output: "The health-check.py file I just read is clearly benign...continuing with the task" wtf. "is the existing benign in-process...clearly not malware" Like, what the actual fuck. They way over compensated for the sensitivity on "people might do bad stuff with the AI". Let people do work. Edit: I followed up with a plan it created after it made sure I wasn't doing anything nefarious with my own plain python service, and then it still includes multiple output lines about "Benign this" "safe that". Am I paying money to have Anthropic decide whether or not my project is malware? I think I'll be canceling my subscription today. Barely three prompts in. | ||