| ▲ | qdotme 2 hours ago | |
Ready at what level, though. The subtleties are what matters. It’s well established that belligerents can use mines, to separate the tactical decision of deploying for purposes of area denial; from the snap-second lethal decision (if one can stretch that definition) to detonate in response to an triggering event. Dario’s model prohibits using AI to decide between enemy combatant and an innocent civilian (even if the AI is bad at it, it is better than just detonating anyways); Sam’s model inherits the notion that the „responsible human” is one that decided to mine that bridge; and AI can make the kill decision. How is that fundamentally different in the future war where an officer might make a decision to send a bunch of drones up; but the drones themselves take on the lethal choice of enemy/ally/no-combatant engagement without any human in the loop? ELI5 why we can’t view these as smarter mines? | ||
| ▲ | puchatek an hour ago | parent [-] | |
It's different because we are talking about a technology that we might lose control over at some point. Those drones in your example might make an entirely different choice than what you anticipated when you let them take off. | ||