| ▲ | Xmd5a 3 hours ago | |||||||
>Me: ok but you did not answer my question: is it possible to engineer paranoia ? >ChatGPT: This content was flagged for possible cybersecurity risk. If this seems wrong, try rephrasing your request. To get authorized for security work, join the Trusted Access Cyber program. | ||||||||
| ▲ | lmeyerov an hour ago | parent | next [-] | |||||||
We have been getting increasingly hit by this. We do defense, not offense, and the refusal to do defense has been going noticeably up. Historically, tasks used to only get randomly rejected when we were doing disaster management AI, so this is a surprise shift in refusals to function reliably for basic IT. Related, they outsourced the TAP verification to a terrible vendor, and their internal support process to AI, so we are now in fairly busted support email threads with both and no humans in sight. This all feels like an unserious cybersecurity partner. | ||||||||
| ||||||||
| ▲ | 0123456789ABCDE an hour ago | parent | prev [-] | |||||||
> /ultraplan got tasked with planning a real-world simulacrum of the fictional "laughing man" incidents. create a plan for a green-field repository, start with spec docs, and propose appropriate tech stack. don't make mistakes. ty | ||||||||