Affiliation:
1. University of Pennsylvania , USA
2. Georgetown University , USA
Abstract
Abstract
Uses of artificial intelligence (AI) are growing around the world. What will influence AI adoption in the international security realm? Research on automation bias suggests that humans can often be overconfident in AI, whereas research on algorithm aversion shows that, as the stakes of a decision rise, humans become more cautious about trusting algorithms. We theorize about the relationship between background knowledge about AI, trust in AI, and how these interact with other factors to influence the probability of automation bias in the international security context. We test these in a preregistered task identification experiment across a representative sample of 9,000 adults in nine countries with varying levels of AI industries. The results strongly support the theory, especially concerning AI background knowledge. A version of the Dunning–Kruger effect appears to be at play, whereby those with the lowest level of experience with AI are slightly more likely to be algorithm-averse, then automation bias occurs at lower levels of knowledge before leveling off as a respondent’s AI background reaches the highest levels. Additional results show effects from the task’s difficulty, overall AI trust, and whether a human or AI decision aid is described as highly competent or less competent.
Funder
Air Force Office of Scientific Research
Publisher
Oxford University Press (OUP)
Reference74 articles.
1. Human Factors Engineering and Flight Deck Design;Abbott,2000
2. Human-AI Interactions in Public Sector Decision-Making: ‘Automation Bias’ and ‘Selective Adherence’ to Algorithmic Advice;Alon-Barkat;Journal of Public Administration Research and Theory,2023
3. Reliance, Trust and Heuristics in Judgemental Forecasting;Alvarado-Valencia;Computers in Human Behavior,2014
4. The Moral Machine Experiment;Awad;Nature,2018
5. Is Trust Important in Technology Adoption? A Policy Capturing Approach;Bahmanziari;Journal of Computer Information Systems,2016