Semi-supervised learning (SSL) has been applied to many practical applications over the past few years. Recently, distributed graph-based semi-supervised learning (DGSSL) has shown to have good performance. Traditiona...
详细信息
ISBN:
(纸本)9781665450850
Semi-supervised learning (SSL) has been applied to many practical applications over the past few years. Recently, distributed graph-based semi-supervised learning (DGSSL) has shown to have good performance. Traditional DGSSL algorithms usually have the problem of the straggler effect that algorithm execution time is limited by the slowest node. To solve this problem, a novel coded DGSSL(CDGSSL) algorithm based on the Maximum Distance Separable (MDS) code is proposed in this paper. Specifically, the proposed algorithm is based on the Maximum Distance Separable (MDS) code. In general, the proposed coded distributed algorithm is straggler-tolerant. Moreover, we provide optimal parameters design for the proposed algorithm. The superiority of the proposed algorithm has been confirmed via experiments on Alibaba Cloud Elastic Compute Service.
Swarm Learning (SL) has been recently proposed for distributed learning, where a group of individual centers perform a synchronized training. Unlike traditional machine learning models that rely on a central server, s...
详细信息
ISBN:
(纸本)9798350383744;9798350383737
Swarm Learning (SL) has been recently proposed for distributed learning, where a group of individual centers perform a synchronized training. Unlike traditional machine learning models that rely on a central server, swarm learning distributes the learning process across multiple nodes. Each node independently processes data and contributes to the overall learning task. This collaboration allows the swarm to benefit from individual nodes' different data. Unlike federated learning, here model parameters are not handled by a central server but are randomly handled across each individual node. The intrinsic attention of swarm learning to data privacy makes it suitable for distributed healthcare analysis, where a clinical center wants to benefit from all the other ones in the swarm network. However, the benefit for a single center or for the whole network could vary depending on data distribution. In this paper, we want to analyze the performance of the swarm learning in a network with multiple nodes, where different data distribution scenarios are taken into account. This analysis will show the gain of the whole swarm network and a specific (reference) node, focusing on scenarios where this node has a different amount of data with respect to the other nodes. To perform a more analytical analysis, we introduce a new Key Performance Indicator (KPI) to measure such gain. We then applied this method using ICU data extracted from the MIMIC EHR database and discussed the results obtained by analyzing 5 nodes with different data distribution scenarios.
This paper examines the equilibrium between user transaction fees and miner profitability within proof-of-work-based blockchains, specifically focusing on Bitcoin. We analyze the dependency of mining profit on factors...
详细信息
ISBN:
(数字)9783031626388
ISBN:
(纸本)9783031626371;9783031626388
This paper examines the equilibrium between user transaction fees and miner profitability within proof-of-work-based blockchains, specifically focusing on Bitcoin. We analyze the dependency of mining profit on factors such as transaction fee adjustments and operational costs, particularly electricity. By applying a multidimensional profitability model and performing a sensitivity analysis, we evaluate the potential for profit maximization through operational cost reduction versus fee increases. Our model integrates variable electricity costs, market-driven Bitcoin prices, mining hardware efficiency, network hash rate, and transaction fee elasticity. We show that mining strategies aimed at reducing electricity expenses are far more profitable than pursuing transactions with higher fees.
In this paper, the parking problem of a swarm of mobile robots has been studied. The robots are deployed at the nodes of an infinite grid, which has a subset of prefixed nodes marked as parking nodes. Each parking nod...
详细信息
ISBN:
(数字)9783031505836
ISBN:
(纸本)9783031505829;9783031505836
In this paper, the parking problem of a swarm of mobile robots has been studied. The robots are deployed at the nodes of an infinite grid, which has a subset of prefixed nodes marked as parking nodes. Each parking node p(i) has a capacity of k(i) which is given as input and equals the maximum number of robots a parking node can accommodate. As a solution to the parking problem, robots need to partition themselves into groups so that each parking node contains a number of robots that are equal to the capacity of the node in the final configuration. It is assumed that the number of robots in the initial configuration represents the sum of the capacities of the parking nodes. The robots are assumed to be autonomous, anonymous, homogeneous, identical and oblivious. They operate under an asynchronous scheduler. They neither have any agreement on the coordinate axes nor do they agree on a common chirality. All the initial configurations for which the problem is unsolvable have been identified. A deterministic distributed algorithm has been proposed for the remaining configurations, ensuring the solvability of the problem.
Large Language Models (LLMs) have changed the way we access and interpret information, communicate with each other and even operate computer systems through autonomous code generation. Typically, these billion-paramet...
详细信息
ISBN:
(纸本)9798350374247;9798350374230
Large Language Models (LLMs) have changed the way we access and interpret information, communicate with each other and even operate computer systems through autonomous code generation. Typically, these billion-parameter models rely on cloud storage and execution due to their computational demands. In this paper, we challenge this status quo by proposing JARVIS, a distributed LLM framework that splits model layers across edge devices with limited compute resources, trading off computation for increased peer-level communication. JARVIS is robust to individual node failures, including recovery methods for lost layers via peer-level duplication. We evaluate JARVIS using Google's open-source Gemma LLM (2B parameters) deployed over 18 software-defined radios in the NSF Colosseum RF emulator. Our evaluation explores LLM performance degradation from node losses, providing insights into node prioritization in tactical environments. The JARVIS software code is released for community exploration and adoption.
The Event Horizon Telescope (EHT) recently used 10 petabyte-scale observation data to construct the first images of black holes and 100 terabyte-scale simulation data to constrain the plasma properties around supermas...
详细信息
ISBN:
(纸本)9798400704192
The Event Horizon Telescope (EHT) recently used 10 petabyte-scale observation data to construct the first images of black holes and 100 terabyte-scale simulation data to constrain the plasma properties around supermassive black holes. This work leveraged the Open Science Grid (OSG) high throughput resources provided by the Partnership to Advance Throughput computing (PATh). While EHT has successfully utilized PATh to create the most extensive black hole simulation library to date, the broad adoption of this resource for data processing has been slower. The sophisticated command-line-driven HTCondor environment creates barriers for less technical researchers, limiting PATh's reach and impact on the broader astronomy and science communities. In May of 2023, the Cyberinfrastructure Integration Research Center (CIRC) at Indiana University was awarded an NSF EAGER award to collaborate with EHT and PATh in implementing a targeted science gateway instance that integrates critical EHT application functionality to leverage OSG within the Apache Airavata framework. The project leverages modern state-of-the-art User Experience (UX) techniques and participatory design methods to lower the barrier to adopting OSG resources for researchers trying to discover the properties of black holes.
With the growing scale of big datasets, fitting novel statistical models on larger-than-memory datasets becomes correspondingly challenging. This document outlines the development and use of an API for large scale mod...
详细信息
The emergence of programmable data planes (PDPs) has paved the way for in-network computing (INC), a paradigm wherein networking devices actively participate in distributed computations. However, PDPs are still a nich...
详细信息
ISBN:
(数字)9798350352917
ISBN:
(纸本)9798350352924;9798350352917
The emergence of programmable data planes (PDPs) has paved the way for in-network computing (INC), a paradigm wherein networking devices actively participate in distributed computations. However, PDPs are still a niche technology, mostly available to network operators, and rely on packet-processing DSLs like P4. This necessitates great networking expertise from INC programmers to articulate computational tasks in networking terms and reason about their code. To lift this barrier to INC we propose a unified compute interface for the data plane. We introduce C/C++ extensions that allow INC to be expressed as kernel functions processing in-flight messages, and APIs for establishing INC-aware communication. We develop a compiler that translates kernels into P4, and thin runtimes that handle the required network plumbing, shielding INC programmers from low-level networking details. We evaluate our system using common INC applications from the literature.
As network bandwidth struggles to keep up with rapidly growing computing capabilities, the efficiency of collective communication has become a critical challenge for exa-scale distributed and parallel applications. Tr...
详细信息
ISBN:
(数字)9798350352917
ISBN:
(纸本)9798350352924;9798350352917
As network bandwidth struggles to keep up with rapidly growing computing capabilities, the efficiency of collective communication has become a critical challenge for exa-scale distributed and parallel applications. Traditional approaches directly utilize error-bounded lossy compression to accelerate collective computation operations, exposing unsatisfying performance due to the expensive decompression-operation-compression (DOC) workflow. To address this issue, we present a first-ever homomorphic compression-communication co-design, hZCCL, which enables operations to be performed directly on compressed data, saving the cost of time-consuming decompression and recompression. In addition to the co-design framework, we build a light-weight compressor, optimized specifically for multi-core CPU platforms. We also present a homomorphic compressor with a run-time heuristic to dynamically select efficient compression pipelines for reducing the cost of DOC handling. We evaluate hZCCL with up to 512 nodes and across five application datasets. The experimental results demonstrate that our homomorphic compressor achieves a CPU throughput of up to 379.08 GB/s, surpassing the conventional DOC workflow by up to 36.53x. Moreover, our hZCCL-accelerated collectives outperform two state-of-the-art baselines, delivering speedups of up to 2.12x and 6.77x compared to original MPI collectives in single-thread and multi-thread modes, respectively, while maintaining data accuracy.
This paper introduces a spark-based fast solution for privacy-preserving frequent pattern mining problems for big data. Spark Resilient distributed Dataset (RDD) framework has been used to implement the Mask approach,...
详细信息
ISBN:
(纸本)9783031640759;9783031640766
This paper introduces a spark-based fast solution for privacy-preserving frequent pattern mining problems for big data. Spark Resilient distributed Dataset (RDD) framework has been used to implement the Mask approach, which uses the probabilistic distortion method for maintaining data privacy while mining frequent patterns. The masking technique shows very promising results in terms of privacy and utility both. However, due to sequential nature limits the application to small or medium size data. The spark-based proposed technique introduces two-level parallelization i.e. data and algorithmic level which in turn paves a way to gain faster analytical results in a bounded amount of time while dealing with a large volume of datasets. This makes the application feasible for the current growth of data size. A number of experiments have been conducted to compare the performance of the proposed scheme with benchmark parallel approaches in terms of privacy, utility, and time complexity over real and simulated data sets. It has been observed that the proposed scheme preserves the privacy of sensitive data while maintaining utility within a real bound of time. Experiments show that the proposed Spark-based scheme i.e. S-Mask gains 16 times speedup on average over different benchmark data sets and maintains a desired ratio between privacy and utility of the data.
暂无评论