Federated Learning is a novel distributed machine learning paradigm that leverages the computing power of numerous decentralized data sources for jointly training machine learning models while ensuring user privacy. I...
Federated Learning is a novel distributed machine learning paradigm that leverages the computing power of numerous decentralized data sources for jointly training machine learning models while ensuring user privacy. In the most commonly used cross-device scenarios, the client cluster typically cover a vast number of heterogeneous end devices. Due to physical limitations such as bandwidth, only a few clients can participate in each round of training. The core issue of the client selection is to determine an appropriate client set for each training round. However, existing selection algorithms, especially the widely adopted random selection, suffer from a number of issues that prevent them from achieving a good balance between training efficiency and speed. Therefore, we propose Scout, which utilizes the heterogeneity features of clients’ data and resources to jointly model the utility function, and enhances the utilization of correlation among clients and the diversity among selected clients to achieve better training efficiency and speed. Furthermore, Scout maintains the scalability and fairness. Our experiments demonstrate that in large-scale heterogeneous clients scenarios, Scout outperforms three baseline algorithms and the state-of-the-art dual-feature dimension algorithm Oort in evaluation metrics.
In the context of 5G, data volume is exploding, and data transmission and processing are facing challenges. In this paper, a data compression algorithm based on neural network is proposed to solve the problem of large...
In the context of 5G, data volume is exploding, and data transmission and processing are facing challenges. In this paper, a data compression algorithm based on neural network is proposed to solve the problem of large scale and high dimension of demand response data in the process of transmission and storage. By mapping demand response data to low dimensional space for representation, the algorithm effectively reduces the amount of data and improves the efficiency of data transmission and storage. The algorithm uses Long Short Term Memory network for data compression and restoration. The experimental results show that the data compression algorithm based on neural network has good compression effect and accuracy. At the same time, the algorithm has good versatility and adaptability, and can be applied to the compression and restoration of data sets of different types and sizes, providing an effective solution for efficient demand response dataprocessing.
Clustering is a method of grouping data based on similarities, and is an unsupervised technique for discovering patterns in data. In this research paper, various clustering algorithms such as k-Means, DBSCAN, Spectral...
Clustering is a method of grouping data based on similarities, and is an unsupervised technique for discovering patterns in data. In this research paper, various clustering algorithms such as k-Means, DBSCAN, Spectral Clustering, Gaussian Mixture, and Agglomerative Clustering are compared and evaluated on Amazon Prime Video Movies and TV Shows, Netflix Movies and TV Shows, and Disney+ Movies and Tv Shows datasets. The results of the study indicate that the k-Means algorithm performed well in clustering the data for all datasets, with an overall high level of performance. Additionally, the study provides valuable insights into the genre distribution of the data, and highlights the advantages and limitations of each clustering algorithm.
Process mining (PM) has established itself in recent years as a main method for visualizing and analyzing processes. However, the identification of knowledge has not been addressed adequately because PM aims solely at...
详细信息
The Internet of Things (loT) is a fast-growing field that involves connecting a wide variety of devices and sensors to the Internet to collect and share data. The vast amount of data generated by these devices require...
The Internet of Things (loT) is a fast-growing field that involves connecting a wide variety of devices and sensors to the Internet to collect and share data. The vast amount of data generated by these devices requires secure storage and processing, which is where cloud computing comes in. Cloud computing provides scalable, on-demand storage and processing resources that can handle the massive amount of data that loT devices produce. By integrating loT and cloud computing, organizations can effectively store, manage, and analyze the data generated by loT devices, leading to more efficient and effective operations. The main disadvantage is security, which is the most important issue nowadays The primary goal of this paper is to make data more secure and reduce latency when data is transported from loT devices to cloud computing, must using efficient lightweight cryptography algorithm suitable in loT environment. In this paper, we will focus on the PRINCE lightweight encryption algorithm. evaluating the proposed system using main parameters so that one can understand the security and performance aspects of the suggested system such as Execution time, and Entropy. The result showed in the last reading as the following execution time is 0.202994713Sec and Entropy is 7.97083942.
One of the most studied disorders in modern medicine is the brain tumor. A correct cancer type diagnosis enables the professionals to select the best course of action and save the patient's life. The significance ...
详细信息
One of the most studied disorders in modern medicine is the brain tumor. A correct cancer type diagnosis enables the professionals to select the best course of action and save the patient's life. The significance of an automated tumor classification system incorporating image processing should go without saying. The classification of different forms of brain tumors using magnetic resonance images (MRI) has been improved in this paper utilizing deep learning and a group of machine learning algorithms. The Brain Tumor MRI dataset can be divided into four groups, which include three types of brain cancer (Glioma, Meningioma, and Pituitary), as well as non-cancerous, or normal type, tumors. o extract detailed information from the MRIpictures, a VGG-19 convolutional neural network is created. Then, to classify among various cancer kinds, these filtered deep characteristics are put into multi-class ML classifiers. The outputs of each ML classifier are thencombined to improve performance using a weighted average ensemble of classifiers. A total of 7022 MRI scansfrom four classes make up the dataset for the system. The accuracy of the Brain Tumor MRI dataset has improved, with a a Meningioma class accuracy of 97.88%, a Pituitary class accuracy of 98.76%, Glioma class accuracy of 99.78%, a normal class accuracy of 99.00% and an overall accuracy of 98.87%.
The construction and operation of Metaverse virtual environments, e.g., 3D reconstruction and activity detection is an important supporting technology of computer vision. Recently, synthetic data has seen a surge in a...
The construction and operation of Metaverse virtual environments, e.g., 3D reconstruction and activity detection is an important supporting technology of computer vision. Recently, synthetic data has seen a surge in adoption for model training in computer vision. Prior research generally show a positive correlation between the volume of synthetic training data and inference accuracy. This paper focuses on the domain of activity detection, and explores how to improve the performance of such algorithms using synthetic data. In particular, we present an overview of the state-of-the-art in using domain randomization approaches for synthetic data generation. This paper presents initial inference accuracies of a model trained on initial attempts at domain randomized synthetic data (7.2%), compared to a model trained on real-world data (9.2%). The synthetic data, although performed worse, indicated promising trajectories for future work, approximately 2% away from the real-world result.
data cleansing approaches aim at revealing and reducing different types of outsourced errors. Such errors introduce a major issue as data cleansing often involves costly computations and time consumption. data cleansi...
详细信息
data cleansing approaches aim at revealing and reducing different types of outsourced errors. Such errors introduce a major issue as data cleansing often involves costly computations and time consumption. data cleansing is complicated as most of the errors within the obtained data emerge in different forms, such as typos, duplicates, noncompliance with business rules, outdated data, and missing values. In this paper, the Supervised dataset Cleaning Model (SDCM) is proposed in order to detect and reduce different types of outsourced errors that are stored in the data repository according to the supervised cleaning rules that are practiced from the previous cleaning processes. The findings indicates that the cleaning execution time can be reduced with this supervised model and the accuracy of the model is also increased when it covers the entire training rules and classified error types derived from the data warehouse. There are some expected future directions, which include: implementing full scale of SDCM and comparing the results obtained with different current methods. By comparing different data cleaning algorithms with SDCM, a knowledge gap is likely to arise in the search for further future improvements.
The proceedings contain 29 papers. The special focus in this conference is on Business Modeling and Software Design. The topics include: Preface;modeling Change in Business Processes;Development of a Ca...
ISBN:
(纸本)9783031367564
The proceedings contain 29 papers. The special focus in this conference is on Business Modeling and Software Design. The topics include: Preface;modeling Change in Business Processes;Development of a Capability Maturity Model for Organization-Wide Green IT Adoption;From Conceptual Specification to OO Specification;Architecting Agility: Unraveling the Impact of AI Capability on Organizational Change and Competitive Advantage;enterprise Architecture Artifacts’ Role in Improved Organizational Performance;A Semiotic Analysis of the Representativeness of BPMN Graphic Elements;A Systematic Approach to Derive User Stories and Gherkin Scenarios from BPMN Models;composing an Initial Domain-Specific Modeling Language Notation by Reusing Icons;validating Trust in Human Decisions to Improve Expert Models Based on Small data Sets;how is Affect Social Justice Tensions: A Case Study of Asylum Management;applying Augmented Reality in Tourism: Analyzing Relevance as It Concerns Consumer Satisfaction and Purchase Intention;realizing Appropriate Process Standardization – Basis for Effective Digital Transformation;comparing Sensor-Based computing and Predictive data Analytics for Usage in Context-Aware Applications;a Model of a Multi-sensor System for Detection and Tracking of Vehicles and Drones;a Web-Based Approach for Traceability in Rule-Based Business Information Systems;a Conceptual Model for the Selection of Methods for Software Engineering Process Improvement;Towards Log-Driven Monitoring of Technical Degradation: An ERP Perspective;A Development Example: From Conceptual Specification to OO Specification;managing database Trigger Design;a View on Vulnerabilities Within IoT Devices in the Smart Home Environment;on Tools for Practical and Effective Security Policy Management and Vulnerability Scanning;linking Computers to the Brain: Overview of Cybersecurity Threats and Possible Solutions;serious Game-Based Haptic Modeling - An Application-Oriented Approach for Sequentially Devel
Cardio Vascular Diseases are the major risk factors for human survival. Cardio Vascular Diseases are comprised with many forms such as abnormal heart rhythms, congenital heart disease, deep vein thrombosis, heart atta...
详细信息
Cardio Vascular Diseases are the major risk factors for human survival. Cardio Vascular Diseases are comprised with many forms such as abnormal heart rhythms, congenital heart disease, deep vein thrombosis, heart attack, heart failure etc. These diseases are identified using scanning devices and doctors can analyze the report by observing heart x-ray images. This research work is carried out using various intelligence algorithms for predicting and classifying the type of heart disease. It is observed that Support Vector Classifier predicts with an accuracy of 99%, Logistic Regression predicts with an accuracy of 85%, Multi Layer Perceptron prediction accuracy is 86.4%, Simple Logistic prediction accuracy is 85.8%, J48 prediction accuracy as 88.62% and Random Forest classification as 100% based on the training data.
暂无评论