Abstract
AbstractDespite the accuracy of machine learning (ML) techniques in predicting invalid bug reports, as shown in earlier research, and the importance of early identification of invalid bug reports in software maintenance, the adoption of ML techniques for this task in industrial practice is yet to be investigated. In this study, we used a technology transfer model to guide the adoption of an ML technique at a company for the early identification of invalid bug reports. In the process, we also identify necessary conditions for adopting such techniques in practice. We followed a case study research approach with various design and analysis iterations for technology transfer activities. We collected data from bug repositories, through focus groups, a questionnaire, and a presentation and feedback session with an expert. As expected, we found that an ML technique can identify invalid bug reports with acceptable accuracy at an early stage. However, the technique’s accuracy drops over time in its operational use due to changes in the product, the used technologies, or the development organization. Such changes may require retraining the ML model. During validation, practitioners highlighted the need to understand the ML technique’s predictions to trust the predictions. We found that a visual (using a state-of-the-art ML interpretation framework) and descriptive explanation of the prediction increases the trustability of the technique compared to just presenting the results of the validity predictions. We conclude that trustability, integration with the existing toolchain, and maintaining the techniques’ accuracy over time are critical for increasing the likelihood of adoption.
Funder
ELLIIT
Blekinge Institute of Technology
Publisher
Springer Science and Business Media LLC
Reference65 articles.
1. Aktas EU, Yilmaz C (2020) Automated issue assignment: results and insights from an industrial case. Empirical Software Engineering 25(5):3544–3589
2. Bachmann A, Bernstein A (2009) Software process data quality and characteristics: a historical view on open and closed source projects. In: The joint international and annual ERCIM workshops on Principles of Software Evolution (IWPSE) and software evolution (Evol) workshops, pp 119–128
3. Bennin KE, Ali NB, Börstler J, Yu X (2020) Revisiting the impact of concept drift on just-in-time quality assurance. 20th International Conference on Software Quality. Reliability and Security, QRS, IEEE, pp 53–59
4. Bettenburg N, Premraj R, Zimmermann T, Kim S (2008) Duplicate bug reports considered harmful. . . really? In: International conference on software maintenance, IEEE, pp 337–345
5. Bhattacharya P, Neamtiu I, Shelton CR (2012) Automated, highly-accurate, bug assignment using machine learning and tossing graphs. Journal of Systems and Software 85(10):2275–2292