咨询与建议

看过本文的还看了

相关文献

该作者的其他文献

文献详情 >Leveraging an Enhanced CodeBER... 收藏

Leveraging an Enhanced CodeBERT-Based Model for Multiclass Software Defect Prediction via Defect Classification

作     者:Hussain, Rida Ghafoor Yow, Kin-Choong Gori, Marco 

作者机构:Univ Florence Dept Informat Engn I-53100 Florence Italy Univ Regina Fac Engn & Appl Sci Regina SK S4S 0A2 Canada Univ Siena Dept Informat Engn I-53100 Siena Italy 

出 版 物:《IEEE ACCESS》 (IEEE Access)

年 卷 期:2025年第13卷

页      面:24383-24397页

核心收录:

基  金:Natural Sciences and Engineering Research Council of Canada (NSERC) [DDG-2024-00035] Cette recherche a ete financee par le Conseil de recherches en sciences naturelles et en genie du Canada (CRSNG) [DDG-2024-00035] 

主  题:Software Codes Predictive models Computer bugs Source coding Deep learning Data models Transformers Semantics Measurement Software defect prediction CodeBERT defects GPT code snippets software reliability 

摘      要:Ensuring software reliability through early-stage defect prevention and prediction is crucial, particularly as software systems become increasingly complex. Automated testing has emerged as the most practical approach to achieving bug-free and efficient code. In this context, machine learning-driven methods, especially those leveraging natural language models, have gained significant traction for developing effective techniques. This paper introduces a novel framework for automating software defect prediction, focusing on eight specific defects: SIGFPE, NZEC, LOGICAL, SYNTAX, SIGSEGV, SIGABRT, SEMANTIC, and LINKER. Our research involves a specialized dataset comprising nine classes, including eight common programming errors and one error-free class. The goal is to enhance software testing and development processes by identifying defects within code snippets. The proposed framework utilizes a CodeBERT-based algorithm for defect prediction, optimizing model hyperparameters to achieve superior accuracy. Comparative analysis against established models such as RoBERTa, Microsoft CodeBERT, and GPT-2 demonstrates that our approach yields significant improvements in prediction performance, with accuracy gains of up to 20% and 7% respectively in binary and multi class experimentation. Empirical studies validate the effectiveness of neural language models like CodeBERT for software defect prediction, highlighting substantial advancements in software testing and development techniques. These findings underscore the potential benefits of incorporating advanced machine learning models into the software development lifecycle.

读者评论 与其他读者分享你的观点

用户名:未登录
我的评分