Existing video-based human activity recognition (HAR) methods are susceptible to challenges such as lighting variations and occlusions in complex environments. Wearable sensors can effectively mitigate these issues. T...
详细信息
Ethiopia, known as the birthplace of coffee, relies on coffee exports as a major source of foreign currency. This research paper focuses on developing a hybrid feature mining technique to automatically classify Ethiop...
详细信息
Autism spectrum disorder (ASD) is a complex neurodevelopmental condition characterized by challenges in social interaction, communication difficulties, repetitive behaviors, and a range of strengths and differences in...
详细信息
Autism spectrum disorder (ASD) is a complex neurodevelopmental condition characterized by challenges in social interaction, communication difficulties, repetitive behaviors, and a range of strengths and differences in cognitive abilities. Early ASD diagnosis using machine learning and deep learning techniques is crucial for preventing its severity and long-term effects. The articles published in this area have only applied different machine learning algorithms, and a notable gap observed is the absence of an in-depth analysis in terms of hyperparameter tuning and the type of dataset used in this context. This study investigated predictive modeling for ASD traits by leveraging two distinct datasets: (i) a raw CSV dataset with tabular data and (ii) an image dataset with facial expression. This study aims to conduct an in-depth analysis of ASD trait prediction in adults and toddlers by doing hyper optimized and interpreting the result through explainable AI. In the CSV dataset, a comprehensive exploration of machine learning and deep learning algorithms, including decision trees, Naive Bayes, random forests, support vector machines (SVM), k-nearest neighbors (KNN), logistic regression, XGBoost, and ANN, was conducted. XGBoost emerged as the most effective machine learning algorithm, achieving an accuracy of 96.13%. The deep learning ANN model outperformed the traditional machine learning algorithms with an accuracy of 99%. Additionally, an ensemble model combining a decision tree, random forest, SVM, KNN, and logistic regression demonstrated superior performance, yielding an accuracy of 96.67%. The XGBoost model, utilized in hyperparameter optimization for CSV data, exhibited a substantial accuracy increase, reaching 98%. For the image dataset, advanced deep learning models, such as ResNet50, VGG16, Boosting, and Bagging, were employed. The bagging model outperformed the others, achieving an impressive accuracy of 99%. Subsequent hyperparameter optimization was conduct
Software defect prediction plays a critical role in software development and quality assurance processes. Effective defect prediction enables testers to accurately prioritize testing efforts and enhance defect detecti...
详细信息
Software defect prediction plays a critical role in software development and quality assurance processes. Effective defect prediction enables testers to accurately prioritize testing efforts and enhance defect detection efficiency. Additionally, this technology provides developers with a means to quickly identify errors, thereby improving software robustness and overall quality. However, current research in software defect prediction often faces challenges, such as relying on a single data source or failing to adequately account for the characteristics of multiple coexisting data sources. This approach may overlook the differences and potential value of various data sources, affecting the accuracy and generalization performance of prediction results. To address this issue, this study proposes a multivariate heterogeneous hybrid deep learning algorithm for defect prediction (DP-MHHDL). Initially, Abstract Syntax Tree (AST), Code Dependency Network (CDN), and code static quality metrics are extracted from source code files and used as inputs to ensure data diversity. Subsequently, for the three types of heterogeneous data, the study employs a graph convolutional network optimization model based on adjacency and spatial topologies, a Convolutional Neural Network-Bidirectional Long Short-Term Memory (CNN-BiLSTM) hybrid neural network model, and a TabNet model to extract data features. These features are then concatenated and processed through a fully connected neural network for defect prediction. Finally, the proposed framework is evaluated using ten promise defect repository projects, and performance is assessed with three metrics: F1, Area under the curve (AUC), and Matthews correlation coefficient (MCC). The experimental results demonstrate that the proposed algorithm outperforms existing methods, offering a novel solution for software defect prediction.
Detecting plagiarism in documents is a well-established task in natural language processing (NLP). Broadly, plagiarism detection is categorized into two types (1) intrinsic: to check the whole document or all the pass...
详细信息
Detecting plagiarism in documents is a well-established task in natural language processing (NLP). Broadly, plagiarism detection is categorized into two types (1) intrinsic: to check the whole document or all the passages have been written by a single author;(2) extrinsic: where a suspicious document is compared with a given set of source documents to figure out sentences or phrases which appear in both documents. In the pursuit of advancing intrinsic plagiarism detection, this study addresses the critical challenge of intrinsic plagiarism detection in Urdu texts, a language with limited resources for comprehensive language models. Acknowledging the absence of sophisticated large language models (LLMs) tailored for Urdu language, this study explores the application of various machine learning, deep learning, and language models in a novel framework. A set of 43 stylometry features at six granularity levels was meticulously curated, capturing linguistic patterns indicative of plagiarism. The selected models include traditional machine learning approaches such as logistic regression, decision trees, SVM, KNN, Naive Bayes, gradient boosting and voting classifier, deep learning approaches: GRU, BiLSTM, CNN, LSTM, MLP, and large language models: BERT and GPT-2. This research systematically categorizes these features and evaluates their effectiveness, addressing the inherent challenges posed by the limited availability of Urdu-specific language models. Two distinct experiments were conducted to evaluate the impact of the proposed features on classification accuracy. In experiment one, the entire dataset was utilized for classification into intrinsic plagiarized and non-plagiarized documents. Experiment two categorized the dataset into three types based on topics: moral lessons, national celebrities, and national events. Both experiments are thoroughly evaluated through, a fivefold cross-validation analysis. The results show that the random forest classifier achieved an ex
The increasing use of cloud-based image storage and retrieval systems has made ensuring security and efficiency crucial. The security enhancement of image retrieval and image archival in cloud computing has received c...
详细信息
The increasing use of cloud-based image storage and retrieval systems has made ensuring security and efficiency crucial. The security enhancement of image retrieval and image archival in cloud computing has received considerable attention in transmitting data and ensuring data confidentiality among cloud servers and users. Various traditional image retrieval techniques regarding security have developed in recent years but they do not apply to large-scale environments. This paper introduces a new approach called Triple network-based adaptive grey wolf (TN-AGW) to address these challenges. The TN-AGW framework combines the adaptability of the Grey Wolf Optimization (GWO) algorithm with the resilience of Triple Network (TN) to enhance image retrieval in cloud servers while maintaining robust security measures. By using adaptive mechanisms, TN-AGW dynamically adjusts its parameters to improve the efficiency of image retrieval processes, reducing latency and utilization of resources. However, the image retrieval process is efficiently performed by a triple network and the parameters employed in the network are optimized by Adaptive Grey Wolf (AGW) optimization. Imputation of missing values, Min–Max normalization, and Z-score standardization processes are used to preprocess the images. The image extraction process is undertaken by a modified convolutional neural network (MCNN) approach. Moreover, input images are taken from datasets such as the Landsat 8 dataset and the Moderate Resolution Imaging Spectroradiometer (MODIS) dataset is employed for image retrieval. Further, the performance such as accuracy, precision, recall, specificity, F1-score, and false alarm rate (FAR) is evaluated, the value of accuracy reaches 98.1%, the precision of 97.2%, recall of 96.1%, and specificity of 917.2% respectively. Also, the convergence speed is enhanced in this TN-AGW approach. Therefore, the proposed TN-AGW approach achieves greater efficiency in image retrieving than other existing
Accurate class and early detection of mind tumour is of maximum significance to facilitate powerful treatment planning to improve affected person consequences. This work offers a novel technique makes use of convoluti...
详细信息
Breast cancer poses a significant global threat, highlighting the urgent need for early detection to reduce mortality rates. Researchers are working to minimize the occurrence of false positives and false negatives, t...
详细信息
In this paper, a new approach for mining image association rules is presented, which involves the fine-tuned CNN model, as well as the proposed FIAR and OFIAR algorithms. Initially, the image transactional database is...
详细信息
Research on sign language translation is ongoing with a high social inclusive goal of crossing the bridge between people with hearing disability using sign language as their basic way to communicate to others who do n...
详细信息
暂无评论