But AI-enabled police would be able to probe actions, infer motives, and detect bad behavior better than humans could. In addition, AI systems could have fewer rights than humans, and could be designed to be more transparent than humans, making the police’s job easier.
Isn’t most of this after a crime has already been committed? Is that enough if it’s an existential risk? To handle this, would we want continuous monitoring of autonomous AIs, at which point aren’t we actually just taking their autonomy away?
Also, if we want to automate “detect bad behavior”, wouldn’t that require AI alignment, too? If we don’t fully automate it, then can we be confident that humans can keep up with everything they need to check themselves, given that AIs could work extremely fast? AIs might learn how much work humans can keep up with and then overwhelm them.
Furthermore, AIs may be able to learn new ways of hiding things from the police, so there could be gaps where the police are trying to catch up.
Cullen’s argument was “alignment may not be enough, even if you solve alignment you might still want to program your AI to follow the law because <reasons>.” So in my responses I’ve been assuming that we have solved alignment; I’m arguing that after solving alignment, AI-powered enforcement will probably be enough to handle the problems Cullen is talking about. Some quotes from Cullen’s comment (emphasis mine):
Reasons other than directly getting value alignment from law that you might want to program AI to follow the law
We will presumably want organizations with AI to be bound by law.
We don’t want to rely on the incentives of human principals to ensure their agents advance their goals in purely legal ways
Some responses to your comments:
if we want to automate “detect bad behavior”, wouldn’t that require AI alignment, too?
Yes, I’m assuming we’ve solved alignment here.
Isn’t most of this after a crime has already been committed?
Good enforcement is also a deterrent against crime (someone without any qualms about murder will still usually not murder because of the harsh penalties and chance of being caught).
Furthermore, AIs may be able to learn new ways of hiding things from the police, so there could be gaps where the police are trying to catch up.
Remember that the police are also AI-enabled, and can find new ways of detecting things. Even so, this is possible: but it’s also possible today, without AI: criminals presumably constantly find new ways of hiding things from the police.
Isn’t most of this after a crime has already been committed? Is that enough if it’s an existential risk? To handle this, would we want continuous monitoring of autonomous AIs, at which point aren’t we actually just taking their autonomy away?
Also, if we want to automate “detect bad behavior”, wouldn’t that require AI alignment, too? If we don’t fully automate it, then can we be confident that humans can keep up with everything they need to check themselves, given that AIs could work extremely fast? AIs might learn how much work humans can keep up with and then overwhelm them.
Furthermore, AIs may be able to learn new ways of hiding things from the police, so there could be gaps where the police are trying to catch up.
Cullen’s argument was “alignment may not be enough, even if you solve alignment you might still want to program your AI to follow the law because <reasons>.” So in my responses I’ve been assuming that we have solved alignment; I’m arguing that after solving alignment, AI-powered enforcement will probably be enough to handle the problems Cullen is talking about. Some quotes from Cullen’s comment (emphasis mine):
Some responses to your comments:
Yes, I’m assuming we’ve solved alignment here.
Good enforcement is also a deterrent against crime (someone without any qualms about murder will still usually not murder because of the harsh penalties and chance of being caught).
Remember that the police are also AI-enabled, and can find new ways of detecting things. Even so, this is possible: but it’s also possible today, without AI: criminals presumably constantly find new ways of hiding things from the police.