Abstract
Deep learning-based software defect prediction has been popular these days. Recently, the publishing of the CodeBERT model has made it possible to perform many software engineering tasks. We propose various CodeBERT models targeting software defect prediction, including CodeBERT-NT, CodeBERT-PS, CodeBERT-PK, and CodeBERT-PT. We perform empirical studies using such models in cross-version and cross-project software defect prediction to investigate if using a neural language model like CodeBERT could improve prediction performance. We also investigate the effects of different prediction patterns in software defect prediction using CodeBERT models. The empirical results are further discussed.
Subject
Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science
Cited by
28 articles.
订阅此论文施引文献
订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献
1. On the Value of Code Embedding and Imbalanced Learning Approaches for Software Defect Prediction;Proceedings of the 12th International Symposium on Information and Communication Technology;2023-12-07
2. The EarlyBIRD Catches the Bug: On Exploiting Early Layers of Encoder Models for More Efficient Code Classification;Proceedings of the 31st ACM Joint European Software Engineering Conference and Symposium on the Foundations of Software Engineering;2023-11-30
3. Method-Level Bug Severity Prediction using Source Code Metrics and LLMs;2023 IEEE 34th International Symposium on Software Reliability Engineering (ISSRE);2023-10-09
4. Software Vulnerability Detection using Large Language Models;2023 IEEE 34th International Symposium on Software Reliability Engineering Workshops (ISSREW);2023-10-09
5. Explaining Transformer-based Code Models: What Do They Learn? When They Do Not Work?;2023 IEEE 23rd International Working Conference on Source Code Analysis and Manipulation (SCAM);2023-10-02