Software engineering workflows use version control systems to track changes and handle merge cases from multiple contributors. This has introduced challenges to testing because it is impractical to test whole codebase...
详细信息
Software engineering workflows use version control systems to track changes and handle merge cases from multiple contributors. This has introduced challenges to testing because it is impractical to test whole codebases to ensure each change is defect-free, and it is not enough to test changed files alone. Just-in-time software defect prediction (JIT-SDP) systems have been proposed to solve this by predicting the likelihood that a code change is defective. Numerous techniques have been studied to build such JIT software defect prediction models, but the power of pre-trained code transformer language models in this task has been underexplored. These models have achieved human-level performance in code understanding and software engineering tasks. Inspired by that, we modeled the problem of change defect prediction as a text classification task utilizing these pre-trained models. We have investigated this idea on a recently published dataset, ApacheJIT, consisting of 44k commits. We concatenated the changed lines in each commit as one string and augmented it with the commit message and static code metrics. Parameter-efficient fine-tuning was performed for 4 chosen pre-trained models, JavaBERT, CodeBERT, CodeT5, and CodeReviewer, with either partially frozen layers or low-rank adaptation (LoRA). Additionally, experiments with the Local, Sparse, and Global (LSG) attention variants were conducted to handle long commits efficiently, which reduces memory consumption. As far as the authors are aware, this is the first investigation into the abilities of pre-trained code models to detect defective changes in the ApacheJIT dataset. Our results show that proper fine-tuning improves the defect prediction performance of the chosen models in the F1 scores. CodeBERT and CodeReviewer achieved a 10% and 12% increase in the F1 score over the best baseline models, JITGNN and JITLine, when commit messages and code metrics are included. Our approach sheds more light on the abilities of l
The Internet of Things(IoT)is emerging as an innovative phenomenon concerned with the development of numerous vital *** the development of IoT devices,huge amounts of information,including users’private data,are *** ...
详细信息
The Internet of Things(IoT)is emerging as an innovative phenomenon concerned with the development of numerous vital *** the development of IoT devices,huge amounts of information,including users’private data,are *** systems face major security and data privacy challenges owing to their integral features such as scalability,resource constraints,and *** challenges are intensified by the fact that IoT technology frequently gathers and conveys complex data,creating an attractive opportunity for *** address these challenges,artificial intelligence(AI)techniques,such as machine learning(ML)and deep learning(DL),are utilized to build an intrusion detection system(IDS)that helps to secure IoT *** learning(FL)is a decentralized technique that can help to improve information privacy and performance by training the IDS on discrete linked *** delivers an effectual tool to defend user confidentiality,mainly in the field of IoT,where IoT devices often obtain privacy-sensitive personal *** study develops a Privacy-Enhanced Federated Learning for Intrusion Detection using the Chameleon Swarm Algorithm and Artificial Intelligence(PEFLID-CSAAI)*** main aim of the PEFLID-CSAAI method is to recognize the existence of attack behavior in IoT ***,the PEFLIDCSAAI technique involves data preprocessing using Z-score normalization to transformthe input data into a beneficial ***,the PEFLID-CSAAI method uses the Osprey Optimization Algorithm(OOA)for the feature selection(FS)*** the classification of intrusion detection attacks,the Self-Attentive Variational Autoencoder(SA-VAE)technique can be ***,the Chameleon Swarm Algorithm(CSA)is applied for the hyperparameter finetuning process that is involved in the SA-VAE model.A wide range of experiments were conducted to validate the execution of the PEFLID-CSAAI *** simulated outcomes demonstrated that the PEFLID-CSAAI
An imbalanced dataset often challenges machine learning, particularly classification methods. Underrepresented minority classes can result in biased and inaccurate models. The Synthetic Minority Over-Sampling Techniqu...
详细信息
An imbalanced dataset often challenges machine learning, particularly classification methods. Underrepresented minority classes can result in biased and inaccurate models. The Synthetic Minority Over-Sampling Technique (SMOTE) was developed to address the problem of imbalanced data. Over time, several weaknesses of the SMOTE method have been identified in generating synthetic minority class data, such as overlapping, noise, and small disjuncts. However, these studies generally focus on only one of SMOTE’s weaknesses: noise or overlapping. Therefore, this study addresses both issues simultaneously by tackling noise and overlapping in SMOTE-generated data. This study proposes a combined approach of filtering, clustering, and distance modification to reduce noise and overlapping produced by SMOTE. Filtering removes minority class data (noise) located in majority class regions, with the k-nn method applied for filtering. The use of Noise Reduction (NR), which removes data that is considered noise before applying SMOTE, has a positive impact in overcoming data imbalance. Clustering establishes decision boundaries by partitioning data into clusters, allowing SMOTE with modified distance metrics to generate minority class data within each cluster. This SMOTE clustering and distance modification approach aims to minimize overlap in synthetic minority data that could introduce noise. The proposed method is called “NR-Clustering SMOTE,” which has several stages in balancing data: (1) filtering by removing minority classes close to majority classes (data noise) using the k-nn method;(2) clustering data using K-means aims to establish decision boundaries by partitioning data into several clusters;(3) applying SMOTE oversampling with Manhattan distance within each cluster. Test results indicate that the proposed NR-Clustering SMOTE method achieves the best performance across all evaluation metrics for classification methods such as Random Forest, SVM, and Naїve Bayes, compared t
Acute Bilirubin Encephalopathy(ABE)is a significant threat to neonates and it leads to disability and high mortality *** and treating ABE promptly is important to prevent further complications and long-term *** studie...
详细信息
Acute Bilirubin Encephalopathy(ABE)is a significant threat to neonates and it leads to disability and high mortality *** and treating ABE promptly is important to prevent further complications and long-term *** studies have explored ABE ***,they often face limitations in classification due to reliance on a single modality of Magnetic Resonance Imaging(MRI).To tackle this problem,the authors propose a Tri-M2MT model for precise ABE detection by using tri-modality MRI *** scans include T1-weighted imaging(T1WI),T2-weighted imaging(T2WI),and apparent diffusion coefficient maps to get indepth ***,the tri-modality MRI scans are collected and preprocessesed by using an Advanced Gaussian Filter for noise reduction and Z-score normalisation for data *** Advanced Capsule Network was utilised to extract relevant features by using Snake Optimization Algorithm to select optimal features based on feature correlation with the aim of minimising complexity and enhancing detection ***,a multi-transformer approach was used for feature fusion and identify feature correlations ***,accurate ABE diagnosis is achieved through the utilisation of a SoftMax *** performance of the proposed Tri-M2MT model is evaluated across various metrics,including accuracy,specificity,sensitivity,F1-score,and ROC curve analysis,and the proposed methodology provides better performance compared to existing methodologies.
Deep learning models for computer vision applications specifically and for machine learning generally are now the state of the art. The growth of size and complexity of neural networks has made them more and more reli...
详细信息
Deep learning models for computer vision applications specifically and for machine learning generally are now the state of the art. The growth of size and complexity of neural networks has made them more and more reliable, yet in greater need of computational power and memory as is evident from the heavy reliance on graphical processing units and cloud computing for training them. As the complexity of deep neural networks increases, the need for fast processing neural networks in real-time embedded applications at the edge also increases and accelerating them using reconfigurable hardware suggests a solution. In this work, a convolutional neural network based on the inception net architecture is first optimized in software and then accelerated by taking advantage of field programmable gate array (FPGA) parallelism. Genetic algorithm augmented training is proposed and used on the neural network to produce an optimum model from the first training run without re-training iterations. Quantization of the network parameters is performed according to the weights of the network. The resulting neural network is then transformed into hardware by writing the register transfer level (RTL) code for FPGAs with exploitation of layer parallelism and a simple trial-and-error allocation of resources with the help of the roofline model. The approach is simple and easy to use as compared to many complex existing methods in literature and relies on trial and error to customize the FPGA design to the model needed to work on any computer vision or multimedia application deep learning model. Simulation and synthesis are performed. The results prove that the genetic algorithm reduces the number of back-propagation epochs in software and brings the network closer to the global optimum in terms of performance. Quantization to 16 bits also shows a reduction in network size by almost half with no performance drop. The synthesis of our design also shows that the Inception-based classifier is cap
Nowadays, with the growth of emerging technologies, increased attention has been paid to the classification of privacy-preserved medical data and development of various privacy-preserving models for the promotion of o...
详细信息
As a result of its aggressive nature and late identification at advanced stages, lung cancer is one of the leading causes of cancer-related deaths. Lung cancer early diagnosis is a serious and difficult challenge that...
详细信息
This work introduces an intrusion detection system (IDS) tailored for industrial internet of things (IIoT) environments based on an optimized convolutional neural network (CNN) model. The model is trained on a dataset...
详细信息
To detect the improper sitting posture of a person sitting on a chair,a posture detection system using machine learning classification has been proposed in this *** addressed problem correlates to the third Sustainabl...
详细信息
To detect the improper sitting posture of a person sitting on a chair,a posture detection system using machine learning classification has been proposed in this *** addressed problem correlates to the third Sustainable Development Goal(SDG),ensuring healthy lives and promoting well-being for all ages,as specified by the World Health Organization(WHO).An improper sitting position can be fatal if one sits for a long time in the wrong position,and it can be dangerous for ulcers and lower spine *** novel study includes a practical implementation of a cushion consisting of a grid of 3×3 force-sensitive resistors(FSR)embedded to read the pressure of the person sitting on ***,the Body Mass Index(BMI)has been included to increase the resilience of the system across individual physical variances and to identify the incorrect postures(backward,front,left,and right-leaning)based on the five machine learning algorithms:ensemble boosted trees,ensemble bagged trees,ensemble subspace K-Nearest Neighbors(KNN),ensemble subspace discriminant,and ensemble RUSBoosted *** proposed arrangement is novel as existing works have only provided simulations without practical implementation,whereas we have implemented the proposed design in *** results validate the proposed sensor placements,and the machine learning(ML)model reaches a maximum accuracy of 99.99%,which considerably outperforms the existing *** proposed concept is valuable as it makes it easier for people in workplaces or even at individual household levels to work for long periods without suffering from severe harmful effects from poor posture.
This study comprehensively analyzes the application of innovative deep learning (DL) and machine learning (ML) techniques in smart energy management systems (EMSs), with an emphasis on load forecasting, demand respons...
详细信息
暂无评论