Abstract
AbstractThe field of AI safety aims to prevent increasingly capable artificially intelligent systems from causing humans harm. Research on moral alignment is widely thought to offer a promising safety strategy: if we can equip AI systems with appropriate ethical rules, according to this line of thought, they’ll be unlikely to disempower, destroy or otherwise seriously harm us. Deontological morality looks like a particularly attractive candidate for an alignment target, given its popularity, relative technical tractability and commitment to harm-avoidance principles. I argue that the connection between moral alignment and safe behavior is more tenuous than many have hoped. In general, AI systems can possess either of these properties in the absence of the other, and we should favor safety when the two conflict. In particular, advanced AI systems governed by standard versions of deontology need not be especially safe.
Publisher
Springer Science and Business Media LLC
Reference89 articles.
1. Aboodi, R., Borer, A., & Enoch, D. (2008). Deontology, individualism, and uncertainty: A reply to Jackson and Smith. Journal of Philosophy, 105, 259–272.
2. Alexander, L., & Moore, M. (2021). Deontological ethics. In E. N. Zalta (ed.), The Stanford Encyclopedia of Philosophy (Winter 2021 Edition), https://plato.stanford.edu/archives/win2021/entries/ethics-deontological/.
3. Algander, P. (2012). A defence of the asymmetry in population ethics. Res Publica, 18, 145–157.
4. Alm, D. (2009). Deontological restrictions and the good/bad asymmetry. Journal of Moral Philosophy, 6, 464–481.
5. Barrington, M. Filtered maximization.