Abstract
DL frameworks are the basis of constructing all DL programs and models, and thus their bugs could lead to the unexpected behaviors of any DL program or model relying on them. Such a wide effect demonstrates the necessity and importance of guaranteeing DL frameworks’ quality. Understanding the characteristics of DL framework bugs is a fundamental step for this quality assurance task, facilitating designing effective bug detection and debugging approaches. Hence, in this work, we conduct the most large-scale study on 1,000 bugs from four popular and diverse DL frameworks (i.e., TensorFlow, PyTorch, MXNet, and DL4J). By analyzing the root causes and symptoms of DL framework bugs associated with five components decomposed from DL frameworks, as well as measuring test coverage achieved by three state-of-the-art testing techniques, we obtain 12 major findings for the comprehensive understanding of DL framework bugs and the current status of existing DL framework testing practice, and then provide a series of actionable guidelines for better DL framework bug detection and debugging. Finally, based on the guidelines, we design and implement a prototype DL-framework testing tool, called
TenFuzz
, which is evaluated to be effective and finds three unknown bugs on the latest TensorFlow framework in a preliminary study, indicating the significance of our guidelines.
Funder
National Natural Science Foundation of China
Publisher
Association for Computing Machinery (ACM)
Reference82 articles.
1. Accessed: April 4th 2022. Coverage.py. Retrieved from https://coverage.readthedocs.io/.
2. Accessed: April 4th 2022. Deeplearning4J. Retrieved from https://deeplearning4j.org/.
3. Accessed: April 4th 2022. Gcov. Retrieved from https://gcc.gnu.org/onlinedocs/gcc/Gcov.html.
4. Accessed: April 4th 2022. Gradle. Retrieved from https://gradle.org/.
5. Accessed: April 4th 2022. MXNet. Retrieved from https://mxnet.apache.org.
Cited by
3 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. Silent bugs in deep learning frameworks: an empirical study of Keras and TensorFlow;Empirical Software Engineering;2023-11-29
2. An Empirical Study on Bugs Inside PyTorch: A Replication Study;2023 IEEE International Conference on Software Maintenance and Evolution (ICSME);2023-10-01
3. Survey on Fuzzing Techniques in Deep Learning Libraries;2023 8th International Conference on Data Science in Cyberspace (DSC);2023-08-18