Affiliation:
1. University of Central Florida, Orlando, Florida, USA
2. Georgia Tech Research Institute, Atlanta, Georgia, USA
3. Air Force Research Laboratory, Dayton, Ohio, USA
Abstract
Objective The present meta-analysis sought to determine significant factors that predict trust in artificial intelligence (AI). Such factors were divided into those relating to (a) the human trustor, (b) the AI trustee, and (c) the shared context of their interaction. Background There are many factors influencing trust in robots, automation, and technology in general, and there have been several meta-analytic attempts to understand the antecedents of trust in these areas. However, no targeted meta-analysis has been performed examining the antecedents of trust in AI. Method Data from 65 articles examined the three predicted categories, as well as the subcategories of human characteristics and abilities, AI performance and attributes, and contextual tasking. Lastly, four common uses for AI (i.e., chatbots, robots, automated vehicles, and nonembodied, plain algorithms) were examined as further potential moderating factors. Results Results showed that all of the examined categories were significant predictors of trust in AI as well as many individual antecedents such as AI reliability and anthropomorphism, among many others. Conclusion Overall, the results of this meta-analysis determined several factors that influence trust, including some that have no bearing on AI performance. Additionally, we highlight the areas where there is currently no empirical research. Application Findings from this analysis will allow designers to build systems that elicit higher or lower levels of trust, as they require.
Funder
Air Force Research Laboratory
Subject
Behavioral Neuroscience,Applied Psychology,Human Factors and Ergonomics
Cited by
105 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献