Abstract
AbstractThe prevalence of artificial intelligence (AI) tools has inspired social studies researchers, ethicists, and policymakers to seriously examine AI’s sociopolitical and ethical impacts. AI ethics literature provides guidance on which ethical principles to implement via AI governance; AI auditing literature, especially ethics-based auditing (EBA), suggests methods to verify if such principles are respected in AI model development and deployment. As much as EBA methods are abundant, I argue that most currently take a top-down and post-hoc approach to AI model development: Existing EBA methods mostly assume a preset of high-level, abstract principles that can be applied universally across contexts; meanwhile, current EBA is only conducted after the development or deployment of AI models. Taken together, these methods do not sufficiently capture the very developmental practices surrounding the constitution of AI models on a day-to-day basis. What goes on in an AI development space and the very developers whose hands write codes, assemble datasets, and design model architectures remain unobserved and, therefore, uncontested. I attempt to address this lack of documentation on AI developers’ day-to-day practices by conducting an ethnographic “AI lab study” (termed by Florian Jaton), demonstrating just how much context and empirical data can be excavated to support a whole-picture evaluation of AI models’ sociopolitical and ethical impacts. I then propose a new method to be added to the arsenal of EBA: Ethnographic audit trails (EATs), which take a bottom-up and in-progress approach to AI model development, capturing the previously unobservable developer practices.
Funder
Geneva Graduate Institute
Publisher
Springer Science and Business Media LLC
Reference63 articles.
1. Suchman, L., Gerst, D., Krämer, H.: ‘If you want to Understand the Big issues, you need to Understand the Everyday practices that constitute them.’ Lucy Suchman in Conversation with Dominik Gerst & Hannes Krämer. Forum Qualitative Sozialforschung / Forum: Qualitative Social Research. (2019). https://doi.org/10.17169/fqs-20.2.3252
2. Panch, T., Mattie, H., Celi, L.A.: The ‘Inconvenient truth’ about AI in Healthcare. Npj Digit. Med. 2(1), 1–3 (2019). https://doi.org/10.1038/s41746-019-0155-4
3. Richardson, R., Schultz, J., Crawford, K.: L Rev. Online. 94, 192–228 (2019). https://ssrn.com/abstract=3333423 Dirty Data, Bad Predictions: How Civil Rights Violations Impact Police Data, Predictive Policing Systems, and Justice. 94 N.Y.U.
4. Yong, E.: A Popular Algorithm Is No Better at Predicting Crimes Than Random People. The Atlantic (blog). (2018). https://www.theatlantic.com/technology/archive/2018/01/equivant-compas-algorithm/550646/ Accessed 20 June 2024
5. Dressel, J., Farid, H.: The Accuracy, Fairness, and limits of Predicting Recidivism. Sci. Adv. 4(1) (2018). https://doi.org/10.1126/sciadv.aao5580