Remix.run Logo
bluGill 4 hours ago

My companies makes potentially dangerous things like lawn mowers. We have a long set of training on how to handle safety issues that gets very complex. Our rules about safety issues is "design it out, then guard it out, and finally warn it out" - that is an ordered list so we cannot go to the next step until we take the previous as far as we can. (and every once in a while we [or a competitor] realize something new and have to revisit everything we sell for that new idea)

Courts will see these things for a while, but there have been enough examples of this type of thing that all AI vendors needs to either have some protection in their system. They can still say "we didn't think of this variation, and here is why it is different from what we have done before", but they can't tell the courts we had no idea people would do stupid things with AI - it is now well known.

I expect this type of thing to play out over many years in court. However I expect that any AI system that doesn't have protection against the common abuses like this that people do will get the owners fined - with fines increasing until they are either taken offline (because the owners can't afford to run them), or the problem fixed so it doesn't happen in the majority of cases.