Remix.run Logo
pixl97 a day ago

>My concern is that actual AI safety

While I'm not disagreeing with you, I would say you're engaging in the no true Scotsman fallacy in this case.

AI safety is: Ensuring your customer service bot does not tell the customer to fuck off.

AI safety is: Ensuring your bot doesn't tell 8 year olds to eat tide pods.

AI safety is: Ensuring your robot enabled LLM doesn't smash peoples heads in because it's system prompt got hacked.

AI safety is: Ensuring bots don't turn the world into paperclips.

All these fall under safety conditions that you as a biological general intelligence tend to follow unless you want real world repercussions.

blagie an hour ago | parent [-]

These are clearly AI safety:

* Ensuring your robot enabled LLM doesn't smash peoples heads in because it's system prompt got hacked.

* Ensuring bots don't turn the world into paperclips.

This is borderline:

* Ensuring your bot doesn't tell 8 year olds to eat tide pods.

I'd put this in a similar category is knives in my kitchen. If my 8-year-old misuses a knife, that's the fault of the adult and not the knife. So it's a safety concern about the use of the AI, but not about the AI being unsafe. Parents should assume 8-year-old shouldn't be left unsupervised with AIs.

And this has nothing to do with safety:

* Ensuring your customer service bot does not tell the customer to fuck off.