With the increasing penetration of distributed energy resources at the grid edge, including renewable energy generation, flexible loads, and energy storage devices, accurately predicting consumer-level distributed gen...
详细信息
ISBN:
(纸本)9798350382570;9798350382563
With the increasing penetration of distributed energy resources at the grid edge, including renewable energy generation, flexible loads, and energy storage devices, accurately predicting consumer-level distributed generation and consumption has become crucial. However, conventional centralized processing and machine learning approaches are impractical because of data security and privacy issues. This paper proposes a load forecasting method based on federated learning with LSTM neuralnetwork which updates model parameters by having clients responsible for model training and a server responsible for model aggregation, allowing multiple clients to collaboratively train a machine learning model without sharing raw data. Simulation results show that this method can achieve accurate predictions while preserving privacy.
As the size of deep learning models gets larger and larger, training takes longer time and more resources, making fault tolerance more and more critical. Existing state-of-the-art methods like CheckFreq and Elastic Ho...
详细信息
As the size of deep learning models gets larger and larger, training takes longer time and more resources, making fault tolerance more and more critical. Existing state-of-the-art methods like CheckFreq and Elastic Horovod need to back up a copy of the model state (i.e., parameters and optimizer states) in memory, which is costly for large models and leads to non-trivial overhead. This article presents Swift, a novel recovery design for distributed deep neuralnetwork training that significantly reduces the failure recovery overhead without affecting training throughput and model accuracy. Instead of making an additional copy of the model state, Swift resolves the inconsistencies of the model state caused by the failure and exploits the replicas of the model state in data parallelism for failure recovery. We propose a logging-based approach when replicas are unavailable, which records intermediate data and replays the computation to recover the lost state upon a failure. The re-computation is distributed across multiple machines to accelerate failure recovery further. We also log intermediate data selectively, exploring the trade-off between recovery time and intermediate data storage overhead. Evaluations show that Swift significantly reduces the failure recovery time and achieves similar or better training throughput during failure-free execution compared to state-of-the-art methods without degrading final model accuracy. Swift can also achieve up to 1.16x speedup in total training time compared to state-of-the-art methods.
This research aims to investigate synchronization issues in coupled memristive neuralnetworks (CMNNs) using both the static and dynamic edge-event triggered control protocols. An interval parameter system is develope...
详细信息
This research aims to investigate synchronization issues in coupled memristive neuralnetworks (CMNNs) using both the static and dynamic edge-event triggered control protocols. An interval parameter system is developed by integrating the concept of Filippov solution with differential inclusion theory. Unlike existing work, the suggested edge-event triggered mechanisms don't require the constant information transfer among neighboring nodes, providing a more distributed control approach that reduces system resources since each node communicates asynchronously. Additionally the absence of Zeno behavior at any given moment supports the efficacy of the approach. To demonstrate its viability, a practical simulation example is presented.
The transformer-based deep neuralnetwork (DNN) models have shown considerable success across diverse tasks, prompting widespread adoption of distributed training methods such as data parallelism and pipeline parallel...
详细信息
The transformer-based deep neuralnetwork (DNN) models have shown considerable success across diverse tasks, prompting widespread adoption of distributed training methods such as data parallelism and pipeline parallelism. With the increasing parameter number, hybrid parallel training becomes imperative to scale training. The primary bottleneck in scaling remains the communication overhead. The communication scheduling technique, emphasizing the overlap of communication with computation, has demonstrated its benefits in scaling. However, most existing works focus on data parallelism, overlooking the nuances of hybrid parallel training. In this paper, we propose TriRace, an efficient communication scheduling framework for accelerating communications in hybrid parallel training of asynchronous pipeline parallelism and data parallelism. To achieve effective computation-communication overlap, TriRace introduces 3D communication scheduling, which adeptly leverages data dependencies between communication and computations, efficiently scheduling AllReduce communication, sparse communication, and peer-to-peer communication in hybrid parallel training. To avoid possible communication contentions, TriRace also incorporates a topology-aware runtime which optimizes the execution of communication operations by considering ongoing communication operations and real-time network status. We have implemented a prototype of TriRace based on PyTorch and Pipedream-2BW, and conducted comprehensive evaluations with three representative baselines. Experimental results show that TriRace achieves up to 1.07-1.45x speedup compared to the state-of-the-art pipeline parallelism training baseline Pipedream-2BW, and 1.24-1.81x speedup compared to the Megatron.
Collaborative Inference is a prospective paradigm for accelerating Deep neuralnetwork (DNN) inference by harnessing the computational resources of multiple devices. However, in highly lossy network environments, such...
详细信息
ISBN:
(纸本)9798350344868;9798350344851
Collaborative Inference is a prospective paradigm for accelerating Deep neuralnetwork (DNN) inference by harnessing the computational resources of multiple devices. However, in highly lossy network environments, such as those encountered in wireless communication systems, the transmission loss of intermediate feature maps between devices can result in significant degradation of co-inference accuracy. In this paper, we first conduct a comprehensive investigation into the impact of intermediate feature map loss in real-world wireless scenarios and provide an in-depth analysis of loss patterns under UDP transmission. Motivated by these observations, we introduce Robust Co-inference Framework (RCIF), a novel framework that employs a hierarchical mask strategy to selectively drop activations at two different scales of feature maps. This approach enhances the robustness of DNN co-inference in the presence of network losses. Our evaluation on a variety of datasets and network architectures demonstrates that RCIF significantly enhances the accuracy and robustness of distributed DNN co-inference under highly lossy network conditions. Specifically, our results show that RCIF can achieve up to a 659% increase in accuracy compared to the original model under particularly poor network conditions.
With the increased usage of edge devices having local computation capabilities, deep neuralnetwork (DNN) training in a network of edge devices becomes promising. Several recent works have proposed fully edge-based di...
详细信息
ISBN:
(纸本)9781665480468
With the increased usage of edge devices having local computation capabilities, deep neuralnetwork (DNN) training in a network of edge devices becomes promising. Several recent works have proposed fully edge-based distributed training systems for situations when the communication to cloud is unstable or intermittent. However, such distributed systems become vulnerable when there are untrusted devices that launch data and model poisoning attacks during training, deteriorating the accuracy of the DNN model. To handle this challenge, we propose a Trustworthy distributed system for Machine learning training in an edge device network (TrustMe). TrustMe realizes both data and model parallelisms. It detects the untrusted devices producing illegitimate outputs. Next, it reassigns the training tasks of the untrusted devices to other trusted devices in such a way that the reassignment and the training that is restarted after the reassignment require minimal time. Our container-based emulation and real device experiments demonstrate that TrustMe achieves up to 12% higher accuracy and 45% less training time compared to existing methods in the presence of untrusted devices.
The ever-growing modern smart grid with more distributed energy resources is providing efficient energy supply while facing several challenges that include harmonics induced among many. Previous and present literature...
详细信息
The ever-growing modern smart grid with more distributed energy resources is providing efficient energy supply while facing several challenges that include harmonics induced among many. Previous and present literature shows that various machine and deep learning models are superior and accurate as compared to the traditional and conventional signal processing techniques. Obtaining accurate results becomes extremely important especially the fact that harmonics are essentially nonlinear, nonparametric, and adaptive in nature. This paper proposes a novel forecasting model that aggregates two deep learning models: convolutional neuralnetwork (CNN) and long short term memory (LSTM) recurrent neuralnetwork (RNN) detect and forecast harmonics in a power system. CNN-LSTM hybrid forecasting model for harmonics in the power grid system has achieved significantly superior performance in collaborative data mining on spatiotemporal measurement data. Sample features are extracted using CNN before they are passed through LSTM for prediction. To show the superiority of the hybrid CNN-LSTM deep neural prediction network model, it is compared with CNN, LSTM and NARX (Non-Linear Autoregressive with External (Exogenous) Input). CNN-LSTM forecasting performance is superior as compared to the other four models. MSE and RMSE for CNN-LSTM are 0.00038 ([3.8 x 10] omicron (-4)) and 0.0000014917 (1.4917 x 10 omicron(-6)) respectively.
Hadoop is a big data processing system that enables the distributedprocessing of massive data sets across multiple computers using straightforward programming techniques. Hadoop has been extensively investigated in m...
详细信息
Hadoop is a big data processing system that enables the distributedprocessing of massive data sets across multiple computers using straightforward programming techniques. Hadoop has been extensively investigated in many attacks as a result of its growing significance in industry. A company may learn about the actions of invaders as well as the weaknesses of the Hadoop cluster by examining a significant quantity of data from the log file. In a Big Data setting, the goal of the paper is to generate an analytical classification for intrusion detection. In this study, Hadoop log files were examined based on assaults that were recorded in the log files. Prior to analysis, the log data is cleaned and improved using a Hadoop preprocessing tool. For feature extraction, the hybrid Improved Sparrow Search Algorithm with Mutual Information Maximization (H-ISSA-MIM). Then the CNN (Convolutional neuralnetwork) classifier will detect the intrusions. The implementation is performed using the MATLAB 2020a software. The performance metrics like accuracy, precision, F-score, recall, specificity, FPR, FNR are calculated for the proposed methodology and it is compared with the existing techniques like Decision Tree (DT), Principal Components Analysis (PCA)-K means, Long Short Time Memory (LSTM). The maximum value of accuracy finds out in the proposed method 98%.
The purpose of this paper is to analyze the frequent power quality (PQ) issues happening in distributed generation, the outcomes of the PQ harmonics, the methods used to assess the quantity of harmonic distortion whic...
详细信息
The purpose of this paper is to analyze the frequent power quality (PQ) issues happening in distributed generation, the outcomes of the PQ harmonics, the methods used to assess the quantity of harmonic distortion which occurs in the power system (PS), and, in the end, classification of these disturbances using recent advance artificial intelligent techniques like a neuralnetwork, fuzzy logic, and the genetic algorithm has been further stated. To protect the PS detection and classification of voltage (V) and current (I) issues are essential tasks and due to increasing interest in a distributed generation, it is becoming more popular. Most PQ disturbances are unstable and ephemeral especially in a distributed generation;therefore, the call for detection and classification of voltage and current disruptions are essential tasks to protect the PS. Many disturbances of (PQ are unpredictable and transient. By using wavelet transforms, expert systems, and artificial neuralnetworks, some intelligent system technologies control fault analysis precisely saying it can help to detect the fault locations. The most important part of the generalized classification system of PQ events is the extraction and classification of features for PQ event classification.
作者:
Strypsteen, ThomasBertrand, Alexander
STADIUS Center for Dynamical Systems Signal Processing and Data Analytics Leuven.AI - KU Leuven institute for AI Kasteelpark Arenberg 10 LeuvenB-3001 Belgium
In this paper, we describe a conceptual design methodology to design distributedneuralnetwork architectures that can perform efficient inference within sensor networks with communication bandwidth constraints. The d...
详细信息
暂无评论