distributed computing is known for its high efficiency of processing large amounts of data in parallel, at the expense of communication load between different servers. Coding was introduced to minimize the communicati...
详细信息
ISBN:
(纸本)9781728182988
distributed computing is known for its high efficiency of processing large amounts of data in parallel, at the expense of communication load between different servers. Coding was introduced to minimize the communication load by exploiting the repetitive computing, thus drawing great attention within the academia. Most existing works assume that all servers are identical in computational capability, which is inconsistent with practical scenarios. In this paper, we investigate a distributed computing system that consists of two types of servers, i.e., fast servers and slow servers. Due to the heterogeneous computational capabilities within the system, the overall computation time will be delayed by the slow servers, which is called the straggling effect. To this end, we develop a novel framework of coding-based distributed computing to alleviate the straggling effect. Specifically, for a given number of fast servers and slow servers with their corresponding computational capabilities, we aim to minimize the overall computation time by assigning different amounts of workloads to different servers. Further, we derive the information-theoretic lower hound of the communication load of the system, which is shown to be within a constant multiplicative gap to the achievable communication load by our scheme.
Esparza and Reiter have recently conducted a systematic comparative study of models of distributed computing consisting of a network of identical finite-state automata that cooperate to decide if the underlying graph ...
详细信息
ISBN:
(纸本)9781450385480
Esparza and Reiter have recently conducted a systematic comparative study of models of distributed computing consisting of a network of identical finite-state automata that cooperate to decide if the underlying graph of the network satisfies a given property. The study classifies models according to four criteria, and shows that twenty-four initially possible combinations collapse into seven equivalence classes with respect to their decision power, i.e. the properties that the automata of each class can decide. However, Esparza and Reiter only show (proper) inclusions between the classes, and so do not characterise their decision power. In this paper we do so for labelling properties, i.e. properties that depend only on the labels of the nodes, but not on the structure of the graph. In particular, majority (whether more nodes carry label a than b) is a labelling property. Our results show that only one of the seven equivalence classes identified by Esparza and Reiter can decide majority for arbitrary networks. We then study the expressive power of the classes on bounded-degree networks, and show that three classes can. In particular, we present an algorithm for majority that works for all bounded-degree networks under adversarial schedulers, i.e. even if the scheduler must only satisfy that every node makes a move infinitely often, and prove that no such algorithm can work for arbitrary networks.
We consider a MapReduce-type task running in a distributed computing model which consists of K edge computing nodes distributed across the edge of the network and a Master node that assists the edge nodes to compute o...
详细信息
ISBN:
(纸本)9781728159621
We consider a MapReduce-type task running in a distributed computing model which consists of K edge computing nodes distributed across the edge of the network and a Master node that assists the edge nodes to compute output functions. The Master node and the edge nodes, both equipped with some storage memories and computing capabilities, are connected through a multicast network. We define the communication time spent during the transmission for the sequential implementation (all nodes send symbols sequentially) and parallel implementation (the Master node can send symbols during the edge nodes' transmission), respectively. We propose a mixed coded distributed computing scheme that divides the system into two subsystems where the coded distributed computing (CDC) strategy proposed by Songze Li et al. is applied into the first subsystem and a novel master-aided CDC strategy is applied into the second subsystem. We prove that this scheme is optimal, i.e., achieves the minimum communication time for both the sequential and parallel implementation, and establish an optimal information-theoretic tradeoff between the overall communication time, computation load, and the Master node's storage capacity. It demonstrates that incorporating a Master node with storage and computing capabilities can further reduce the communication time. For the sequential implementation, we deduce the approximately optimal file allocation between the two subsystems, which shows that the Master node should map as many files as possible in order to achieve smaller communication time. For the parallel implementation, if the Master node's storage and computing capabilities are sufficiently large (not necessary to store and map all files), then the proposed scheme requires at most 1/2 of the minimum communication time of system without the help of the Master node.
The content and structure of the news text are relatively complex and cannot effectively capture the core content. Existing supervision models cannot achieve good results in areas with less annotated data. To this end...
详细信息
The content and structure of the news text are relatively complex and cannot effectively capture the core content. Existing supervision models cannot achieve good results in areas with less annotated data. To this end, we propose a new Chinese keyword extraction model with a distributed computing method. Precisely, we first fused the Bidirectional Encoder Representation from Transformers (BERT) and Conditional Random Fields (CRF) so that each word learns its relationship with the context while reducing errors;secondly, the adversarial training encourages the model to retain a small amount of annotations Sample knowledge to help extract keywords from unannotated samples;and because the model contains a large number of time-consuming components, it creatively uses distributed computing to save overall computing time. The results show that our model can steadily improve the performance of keyword phrase extraction in areas with insufficient labeled samples.
Solar forecasting is a crucial and cost-effective tool for better utilization of solar energy for smart environment design. Artificial intelligence (AI) technologies, such as machine learning (ML) and deep learning (D...
详细信息
Solar forecasting is a crucial and cost-effective tool for better utilization of solar energy for smart environment design. Artificial intelligence (AI) technologies, such as machine learning (ML) and deep learning (DL), have gained great popularity and been widely applied in solar forecasting in recent years. However, conventional AI -based forecasting methods suffer from high variability and stochasticity of solar irradiation, making unreliable predictions due to heterogeneous solar resources. Moreover, the training process of DL models is less flexible and requires immense data. Even for a well-trained model, it can still yield deteriorated performances on other datasets of varying data distributions. To tackle the deficiencies of AI forecasting models, we present a flexible distributed solar forecasting framework based on a novel spatial and temporal attention-based neural network (STANN) in conjunction with federated learning (FL) technique, considering multi-horizon forecasting scenario from 5-30 min. The STANN model consists of a feature extractor and a forecaster, which can be respectively trained on various local datasets for better localization, and updated to further improve forecasting accuracy through global parameter aggregation under the proposed framework without data gathering. We evaluate effectiveness of the proposed method by conducting extensive experiments on real-world datasets and compare it to other popular forecasting models. The results demonstrate that our approach outperforms the other benchmarks with higher forecasting accuracy for all forecast horizons and better generalization on various datasets, achieving the highest forecast skill of 28.83% at 30 min horizon and an improvement of 11.2% compared with the centralized, localized, and conventional FL training methods.
Training a large-scale model over a massive data set is an extremely computation and storage intensive task, e.g. training ResNet with hundreds of millions of parameters over the data set ImageNet with millions of ima...
详细信息
Training a large-scale model over a massive data set is an extremely computation and storage intensive task, e.g. training ResNet with hundreds of millions of parameters over the data set ImageNet with millions of images. As a result, there has been significant interest in developing distributed learning strategies that speed up the training of learning models. Due to the growing computational power of the ecosystem of billions of mobile and computing devices, many future distributed learning systems operate based on storing data locally and pushing computation to the network edge. Unlike traditional centralized machine learning environments, however, machine learning at the edge is characterized by significant challenges including (1) scalability due to severe constraints on communication bandwidth and other resources including storage and energy, (2) robustness to stragglers, and edge failures due to slow edge nodes, (3) models generalizing to non-i.i.d. and heterogeneous data. In this thesis, we focus on two important distributed learning frameworks: Federated Learning and distributed computing, with a shared goal in mind: how to provably address the critical challenges in such paradigms using novel techniques from distributed optimization, statistical learning theory, probability theory, and communication and coding theory to advance the state-of-the-art in efficiency, resiliency, and scalability. In the first part of the thesis, we devise three methods to mitigate communication cost, straggler resiliency and robustness to heterogeneous data in federated learning paradigms. Our main ideas are to employ model compression, adaptive device participation and distributionally robust minimax optimization, respectively for such challenges. We characterize provable improvements for the proposed algorithms in terms of convergence speed, expected runtime, and generalization gaps. Moving on to the second part, we consider important instances of distributed computing framew
As natural disasters become extensive, due to various environmental problems, such as the global warming, it is difficult for the disaster management systems to rapidly provide disaster prediction services, due to com...
详细信息
As natural disasters become extensive, due to various environmental problems, such as the global warming, it is difficult for the disaster management systems to rapidly provide disaster prediction services, due to complex natural phenomena. Digital twins can effectively provide the services using high-fidelity disaster models and real-time observational data with distributed computing schemes. However, the previous schemes take little account of the correlations between environmental data of disasters, such as landscapes and weather. This causes inaccurate computing load predictions resulting in unbalanced load partitioning, which increases the prediction service times of the disaster management agencies. In this paper, we propose a novel distributed computing framework to accelerate the prediction services through semantic analyses of correlations between the environmental data. The framework combines the data into disaster semantic data to represent the initial disaster states, such as the sizes of wildfire burn scars and fuel models. With the semantic data, the framework predicts computing loads using the convolutional neural network-based algorithm, partitions the simulation model into balanced sub-models, and allocates the sub-models into distributed computing nodes. As a result, the proposal shows up to 38.5% of the prediction time decreases, compared to the previous schemes.
In order to make it easier for drivers to find a parking slot, optimise the resources of urban parking slots, and alleviate the problem of parking slot shortage, a distributed parking allocation algorithm was proposed...
详细信息
In order to make it easier for drivers to find a parking slot, optimise the resources of urban parking slots, and alleviate the problem of parking slot shortage, a distributed parking allocation algorithm was proposed. The algorithm collects the parking requests of user, this parking requests including the current position coordinate information of the users and destination coordinate information, the algorithm allocates parking spaces to users by analysing the available state of parking spaces, then return the parking route planning to the client. Compared with the traditional algorithm, the distributed parking algorithm has a higher ability to withstand pressure and global search capability, and it can ensure the real-time and validity of the parking information, so it can reduce the problem of the parking space shortage and unavailable parking space. The simulation results show that this algorithm can find the solution set more quickly and accurately under the circumstance of high demand. It also has application value and practicality.
Multi-view clustering (MvC) is an emerging task in data mining. It aims at partitioning the data sampled from multiple views. Although a great deal of research has been done, this task remains to be very challenging. ...
详细信息
Multi-view clustering (MvC) is an emerging task in data mining. It aims at partitioning the data sampled from multiple views. Although a great deal of research has been done, this task remains to be very challenging. We found an important problem in performing the MvC task. MvC needs large amounts of computation. To address this problem, we propose a parallel MvC method in a distributed computing environment. The proposed method builds upon concept factorization with local manifold learning, denoted by parallel multi-view concept clustering (PMCC). Concept factorization learns a compressed representation for the data. Local manifold learning preserves the locally intrinsic geometrical structure in the data. The weight of each view is learned automatically and a cooperative normalized approach is proposed to better guide the learning of a consensus representation for all views. For the proposed PMCC architecture, the calculation of each part is independent. It is clear that our PMCC can be performed in a distributed computing environment. Experimental results using real-world datasets demonstrate the effectiveness of the proposed method.
Ensuring social security through the defense organization determines the creation of links between the army and society. Realizing the benefits of the Internet of Battle Things in the defense system can perfectly mone...
详细信息
Ensuring social security through the defense organization determines the creation of links between the army and society. Realizing the benefits of the Internet of Battle Things in the defense system can perfectly monetize intelligence and strengthen the armed forces. It establishes a network for strong connectivity in the army with good coordination between complex processes to effectively edge out the enemies. However, this new technology poses organizational and national security challenges that present both opportunities and obstacles. The current framework of the defense IoT network for sustainable society is not adequate enough to make actionable situational awareness decisions in order to infer the state of the battlefield while preserving the privacy of sensitive data. In this paper, we propose a distributed computing defence framework for sustainable society using the features of blockchain technology and federated learning. The proposed model presents an algorithm to meet the challenges of limited training data in order to obtain high accuracy and avoid a reason specific model. To evaluate the effectiveness of the proposed model, we prepare the dataset and investigate the performance of our framework in various scenarios. The result outcomes are promising in terms of accuracy and loss compared to baseline approach.
暂无评论