Affiliation:
1. Massachusetts Institute of Technology, Cambridge , MA, USA
2. Microsoft Research, Redmond, WA, USA
Abstract
Machine learning (ML) has become a crucial component in software products, either as part of the user experience or used internally by software teams. Prior studies have explored how ML is affecting development team roles beyond data scientists, including user experience designers, program managers, developers and operations engineers. However, there has been little investigation of how team members in different roles on the team communicate about ML, in particular about the quality of models. We use the general term quality to look beyond technical issues of model evaluation, such as accuracy and overfitting, to any issue affecting whether a model is suitable for use, including ethical, engineering, operations, and legal considerations. What challenges do teams face in discussing the quality of ML models? What work practices mitigate those challenges? To address these questions, we conducted a mixed-methods study at a large software company, first interviewing15 employees in a variety of roles, then surveying 168 employees to broaden our understanding. We found several challenges, including a mismatch between user-focused and model-focused notions of performance, misunderstandings about the capabilities and limitations of evolving ML technology, and difficulties in understanding concerns beyond one's own role. We found several mitigation strategies, including the use of demos during discussions to keep the team customer-focused.
Publisher
Association for Computing Machinery (ACM)
Subject
Computer Networks and Communications,Human-Computer Interaction,Social Sciences (miscellaneous)
Reference63 articles.
1. [n.d.]. 2017 Kaggle ML & DS Survey. https://kaggle.com/kaggle/kaggle-survey-2017 Library Catalog: www.kaggle.com. [n.d.]. 2017 Kaggle ML & DS Survey. https://kaggle.com/kaggle/kaggle-survey-2017 Library Catalog: www.kaggle.com.
2. [n.d.]. How to Prevent Discriminatory Outcomes in Machine Learning. https://www.weforum.org/whitepapers/how-to-prevent-discriminatory-outcomes-in-machine-learning/ Library Catalog: www.weforum.org. [n.d.]. How to Prevent Discriminatory Outcomes in Machine Learning. https://www.weforum.org/whitepapers/how-to-prevent-discriminatory-outcomes-in-machine-learning/ Library Catalog: www.weforum.org.
3. 2014. 3 Data Careers Decoded and What It Means for You. https://blog.udacity.com/2014/12/data-analyst-vs-data-scientist-vs-data-engineer.html Library Catalog: blog.udacity.com Section: Career Guidance. 2014. 3 Data Careers Decoded and What It Means for You. https://blog.udacity.com/2014/12/data-analyst-vs-data-scientist-vs-data-engineer.html Library Catalog: blog.udacity.com Section: Career Guidance.
4. 2018. Tutorial: 21 fairness definitions and their politics. https://www.youtube.com/watch?v=jIXIuYdnyyk&ab_channel=ArvindNarayanan 2018. Tutorial: 21 fairness definitions and their politics. https://www.youtube.com/watch?v=jIXIuYdnyyk&ab_channel=ArvindNarayanan
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献