Detecting fake news in the digital era is challenging due to the proliferation of misinformation. One of the crucial is-sues in this domain is the inherent class imbalance, where genuine news articles significantly ou...
Detecting fake news in the digital era is challenging due to the proliferation of misinformation. One of the crucial is-sues in this domain is the inherent class imbalance, where genuine news articles significantly outnumber fake ones. This imbalance severely hampers the performance of machine and deep learning models in accurately identifying fake news. Consequently, there is a compelling need to address this problem effectively. In this study, we delve into fake news detection and tackle the critical issue of imbalanced data. We investigate the application of Easy Data Augmentation (EDA) techniques, including back-translation, random insertion, random deletion, and random swap to mitigate the adverse effects of imbalanced data. This study focuses on employing these techniques in conjunction with a deep learning framework, specifically a Bidirectional Long Short-Term Memory (BiLSTM) architecture. The results of the EDA techniques will be systematically compared to see their effectiveness and their impacts on model performance. This study reveals that various EDA techniques, when coupled with a BiLSTM architecture, yield significant improvements in fake news detection. Among the experiments, it shows that Random Insertion, with an impressive accuracy rate of 81.68%, a precision score of 89.38%, and an F1-Score of 87.77% emerges as the most promising technique. The study also highlights the exceptional potential of Back-translation stands out with an 87.16% recall performance.
Addressing multiple criteria and parameter issues in computer modelling presents a significant challenge. Several factors including data types, parameter behaviours, and purposes, must be taken into account to enhance...
Addressing multiple criteria and parameter issues in computer modelling presents a significant challenge. Several factors including data types, parameter behaviours, and purposes, must be taken into account to enhance computer modelling capability; particularly in evaluation cases. Through the utilization of a multi-criteria and method approach, a decision model was effectively developed to assess a case of environmental sustainability level of a building. One method operated in the study is the curve method for handling membership function form in realizing fuzzy logic. This innovative model demonstrates superior performance. It achieves an impressive accuracy rate of 96%, surpassing the previous model that employed a trapezoidal approach to describe fuzzy membership functions hy 1%.
Cancer, also known as malignant neoplasm, is a complex and potentially fatal disease characterized by uncontrolled and abnormal cell growth in the body. The main problems with microarray cancer studies are the high cu...
Cancer, also known as malignant neoplasm, is a complex and potentially fatal disease characterized by uncontrolled and abnormal cell growth in the body. The main problems with microarray cancer studies are the high curse of dimensionality and small sample size caused by redundant and irrelevant genes. To deal with the number of features that exceed the amount of data, this research purposed double filtering method Lasso-GA, Lasso is used to select the features based on feature correlation while Genetic Algorithm is used to optimize the most important features with accuracy traditional machine learning as its fitness function. The results show how effective the suggested method is; in breast cancer, the linear SVC model achieves excellent accuracy (0.93), precision (0.94), recall (0.94), and F1 score (0.94), while in lung cancer, the linear SVC, random forest, and logistic regression models perform well (accuracy: 0.95, precision: 0.92, recall: 1, F1 score: 0.95). Logistic regression is the most effective method for bladder cancer, with an accuracy of 0.82, precision of 0.77, recall of 1, and F1 score of 0.87.
Containerization has become a popular approach in application development in applications development and deployment, many benefits we can get such as improved scalability, portability, and resource efficiency. Contai...
Containerization has become a popular approach in application development in applications development and deployment, many benefits we can get such as improved scalability, portability, and resource efficiency. Container-based applications, utilizing technologies like Docker and Kubernetes, have transformed the packaging, deployment, and management of software from the desktop environment to the cloud platform. In this context, software metrics approach plays a good role in evaluating the characteristics and performance of container-based applications, ensuring that developers and operators are on the same page. This article explores the importance of software metrics in optimizing the software lifecycle of container-based applications, addressing the unique challenges they present, and highlighting the potential benefits of leveraging metrics to improve performance and efficiency. Our finding Performance Metrics and Availability Metrics is the most metrics that the most measure by applications owner, relevant studies and industry practices, this study aims to provide insights and recommendations to effectively measure and optimize region-based software systems.
Rapid progress in the field of artificial intelligence has created opportunities for extensive applications in software development. One area that receives attention is the evaluation of code quality using machine lea...
Rapid progress in the field of artificial intelligence has created opportunities for extensive applications in software development. One area that receives attention is the evaluation of code quality using machine learning techniques. In this investigation, we examined the possible application of machine learning to predict the likelihood of defects in computer code. We employ NASA archival data as case studies. Machine learning models employ neural network algorithms. Our exploration involves partitioning the dataset into training data and test data for performance evaluation. The findings indicate that the Neural Organize technique with resampling yields a high level of accuracy in predicting software defects. Our simulated neural network is capable of identifying intricate patterns in the data and providing precise measurements of the size and intensity of defects. These results have significant implications in the software business, enabling developers to promptly identify possible vulnerabilities and take preventive measures before product release.
This study examines the necessity of employing BERT2GPT for single-document summarization in the current age of escalating digital data. The primary focus of this work is on the abstractive technique, which tries to g...
This study examines the necessity of employing BERT2GPT for single-document summarization in the current age of escalating digital data. The primary focus of this work is on the abstractive technique, which tries to generate versatile summaries that surpass the primary sentences of the original content. This study employs two models, namely BERT and GPT -2, for the purpose of the summarization system. The paper introduces the BERT2GPT model, which merges the bidirectional characteristics of BERT with the generative powers of GPT. The findings indicate that the BERT2GPT method successfully catches significant information and linguistic nuances, hence enhancing the quality of the generated summaries. The corresponding average values for Rl, R2, and RL are 0.62, 0.56, and 0.60, respectively.
The advancements in technology during the 20th century resulted in the onset of the digital computer era. This study investigates the relative importance of earlier algorithms in comparison to more recent ones. Variou...
The advancements in technology during the 20th century resulted in the onset of the digital computer era. This study investigates the relative importance of earlier algorithms in comparison to more recent ones. Various research projects suggest improving vectorization by combining conventional and modern techniques, while others suggest optimizing it through algorithmic methods. This study primarily focuses on employing Bayesian optimization to optimize hyperparameters, hence improving the performance evaluation of TF-IDF FastText sentiment classification models. This study presented four models: the initial model utilized FastText with a preprocessing step that involved removing stopwords, the second model employed Fast-Text without any optimization using Bayesian optimization, the third model utilized FastText with optimization using Bayesian optimization, and the fourth model combined FastText with TF-IDF and was further optimized using Bayesian optimization. The Support Vector Machine (SVM) technique will be employed to evaluate all of the models. The findings suggest that the model’s performance stays unchanged when stopwords are eliminated (Precision, Recall, F1: 0.9007). The model demonstrates a slight enhancement through the utilization of Bayesian optimization, resulting in a 0.02% increase, reaching a final accuracy of 0.9019%. Bayesian optimization is frequently employed to improve performance. The performance of Bayesian optimization is affected by the magnitude of the learning rate, and models that do not utilize Bayesian optimization demonstrate inferior performance. Both the TF-IDF FasText model and FastText model yielded comparable outcomes, attaining an F1 score of 0.9019 after being tuned by Bayesian optimization.
Applications designed utilizing Microservices Architecture (MSA) provide the desirable trait of good maintainability. To ensure optimal maintainability, it is important to provide services that are suitable and adhere...
Applications designed utilizing Microservices Architecture (MSA) provide the desirable trait of good maintainability. To ensure optimal maintainability, it is important to provide services that are suitable and adhere to prescribed rules. Multiple aspects must be taken into account while designing services to ensure optimal maintainability. The objective of this study is to examine the elements that impact the capacity to sustain and improve maintainability in service design, ultimately resulting in an application that possesses strong maintainability. The Systematic Literature Review (SLR) will be utilized to identify variables and strategies for their enhancement, by examining pertinent publications on the subject. After examining 45 publications, the study discovered 8 elements and 14 solutions that can enhance the highlighted parameters throughout the services design process. The outcomes of this systematic literature review (SLR) are anticipated to give valuable insights to application developers, empowering them to generate service designs that exhibit commendable maintainability for the developed applications.
This study addresses the Wireless Sensor Network Planning Problem with Multiple Sources/Destinations (WSNPMSD), an optimization challenge focused on reducing the sensor count within a network topology for a specified ...
详细信息
ISBN:
(数字)9781737749769
ISBN:
(纸本)9798350371420
This study addresses the Wireless Sensor Network Planning Problem with Multiple Sources/Destinations (WSNPMSD), an optimization challenge focused on reducing the sensor count within a network topology for a specified area, considering numerous sources and destinations. We introduce a hybrid strategy for tackling WSNP-MSD, particularly effective for large-scale scenarios, combining a Biased Random-key Genetic Algorithm with a Local Branching Technique. This methodology is justified by the limitations exact methods may encounter when the number of variables increases. Through computational experiments, we demonstrate the superiority of our proposed method over conventional exact methods in managing large instances of the WSNP-MSD.
This study discusses the development of smart precision farming systems using big data and cloud-based intelligent decision support systems. Big data plays an important role in collecting, storing, and analyzing large...
This study discusses the development of smart precision farming systems using big data and cloud-based intelligent decision support systems. Big data plays an important role in collecting, storing, and analyzing large amounts of data from various sources related to agriculture, including data from weather stations, soil sensors, satellite imagery, crop yield records, pest and disease reports, and other sources. This study highlights the differences between smart farming and precision farming. This study describes key techniques and system architecture, including data collection, processing, analysis, and decision support components. Utilizing a cloud platform enables scalability and optimized performance, which lowers costs and makes it safer and easier to manage. The integration of big data and Alibaba cloud computing in smart precision farming can improve farming productivity by providing timely information and recommendations to farmers for better decision-making. Finally, the system produces smart precision farming, which provides cost-effective real-time monitoring and predictive analytics to increase agricultural production and sustainability.
暂无评论