Abstract
AbstractThis article discusses the potential sources and consequences of unfairness in artificial intelligence (AI) predictive tools used for anti-corruption efforts. Using the examples of three AI-based anti-corruption tools from Brazil—risk estimation of corrupt behaviour in public procurement, among public officials, and of female straw candidates in electoral contests—it illustrates how unfairness can emerge at the infrastructural, individual, and institutional levels. The article draws on interviews with law enforcement officials directly involved in the development of anti-corruption tools, as well as academic and grey literature, including official reports and dissertations on the tools used as examples. Potential sources of unfairness include problematic data, statistical learning issues, the personal values and beliefs of developers and users, and the governance and practices within the organisations in which these tools are created and deployed. The findings suggest that the tools analysed were trained using inputs from past anti-corruption procedures and practices and based on common sense assumptions about corruption, which are not necessarily free from unfair disproportionality and discrimination. In designing the ACTs, the developers did not reflect on the risks of unfairness, nor did they prioritise the use of specific technological solutions to identify and mitigate this type of problem. Although the tools analysed do not make automated decisions and only support human action, their algorithms are not open to external scrutiny.
Funder
H2020 European Research Council
Alma Mater Studiorum - Università di Bologna
Publisher
Springer Science and Business Media LLC
Reference73 articles.
1. Aarvik, P. (2019). Artificial Intelligence – a promising anti-corruption tool in development settings? Report, U4, 1–38. Retrieved May 16, 2023, from https://www.u4.no/publications/artificial-intelligence-a-promising-anti-corruption-tool-in-development-settings
2. Adam, I., & Fazekas, M. (2021). Are emerging technologies helping win the fight against corruption? A review of the state of evidence. Information Economics and Policy, 57(100950), 1–14. https://doi.org/10.1016/j.infoecopol.2021.100950
3. Alfrink, K., Keller, I., Kortuem, G., & Doorn, N. (2022). Contestable AI by design: Towards a framework. Minds and Machines. https://doi.org/10.1007/s11023-022-09611-z
4. Angwin, J., Larson, J., Mattu, S., & Kirchner, L. (2016). Machine bias: There’s software used across the country to predict future criminals. and it’s biased against blacks. ProPublica. Retrieved May 16, 2023, from https://www.propublica.org/article/machine-bias-risk-assessments-in-criminal-sentencing
5. Aradau, C., & Blanke, T. (2017). Politics of prediction: Security and the time/space of governmentality in the age of big data. European Journal of Social Theory, 20(3), 373–391. https://doi.org/10.1177/1368431016667623