| ▲ | ang_cire 4 hours ago | |
> In general we trust people that we bring onto our team not to betray us and to respect general rules and policies and practices that benefit everyone. And yet we give people the least privileges necessary to do their jobs for a reason, and it is in fact partially so that if they turn malicious, their potential damage is limited. We also have logging of actions employees do, etc etc. So yes, in the general sense we do trust that employees are not outright and automatically malicious, but we do put *very broad* constraints on them to limit the risk they present. Just as we 'sandbox' employees via e.g. RBAC restrictions, we sandbox AI. | ||
| ▲ | SirMaster 4 hours ago | parent [-] | |
But if there is a policy in place to prevent some sort of modification, then performing an exploit or workaround to make the modification anyways is arguably understood and respected by most people. That seems to be the difference here, we should really be building AI systems that can be taught or that learn to respect things like that. If people are claiming that AI is so smart or smarter than the average person then it shouldn't be hard for it to handle this. Otherwise it seems people are being to generous in talking about how smart and capable AI systems truly are. | ||