With the rapid expansion of the Internet of Things (IoT), the shift from cloud computing to Mobile Edge computing (MEC) has become necessary to address the low-latency requirements of real-time applications. Verifiabl...
详细信息
With the rapid expansion of the Internet of Things (IoT), the shift from cloud computing to Mobile Edge computing (MEC) has become necessary to address the low-latency requirements of real-time applications. Verifiable computation (VC) enables resource-limited clients to outsource their computation-intensive tasks to a powerful cloud while ensuring the correctness of the computation result. However, traditional VC schemes, originally designed for cloud computing, face challenges when applied to MEC environments, such as scalability issues, robustness, and efficiency concerns. To this end, we propose a verifiable distributed computation scheme for MEC, where computation tasks are distributed between a cloud server cluster (consisting of $n$n servers) and an edge server. The cloud handles most of the computation through parallel sub-tasks, while the edge server verifies intermediate results and performs minimal computation to recover the final outcome. Our scheme guarantees that the result can be recovered if at least $t$t servers, out of a total of $n$n servers in the cloud server cluster, perform their computations honestly. By leveraging batch verification and matrix-optimized polynomial evaluations, our scheme significantly enhances scalability, fault tolerance, and efficiency. The extensive analysis and simulations demonstrate that our proposed scheme is more feasible than existing solutions.
NESIM-RT is a specialized tool designed for simulating neuromorphic systems. In this new release we extend its capabilities to include state-of-the art models like the AdexLIF and Izhikevich, and to incorporate dynami...
详细信息
NESIM-RT is a specialized tool designed for simulating neuromorphic systems. In this new release we extend its capabilities to include state-of-the art models like the AdexLIF and Izhikevich, and to incorporate dynamic synaptic mechanisms such as Spike-Timing Dependent Plasticity (STDP). With these new features, researchers can now observe in real -time how different parameters influence these models and learning rules, thereby gaining deeper insights into neuronal function and network dynamics.
Straggler nodes are well-known bottlenecks of distributed matrix computations which induce reductions in computation/communication speeds. A common strategy for mitigating such stragglers is to incorporate Reed-Solomo...
详细信息
Straggler nodes are well-known bottlenecks of distributed matrix computations which induce reductions in computation/communication speeds. A common strategy for mitigating such stragglers is to incorporate Reed-Solomon based MDS (maximum distance separable) codes into the framework;this can achieve resilience against an optimal number of stragglers. However, these codes assign dense linear combinations of submatrices to the worker nodes. When the input matrices are sparse, these approaches increase the number of non-zero entries in the encoded matrices, which in turn adversely affects the worker computation time. In this work, we develop a distributed matrix computation approach where the assigned encoded submatrices are random linear combinations of a small number of submatrices. In addition to being well suited for sparse input matrices, our approach continues to have the optimal straggler resilience in a certain range of problem parameters. Moreover, compared to recent sparse matrix computation approaches, the search for a "good" set of random coefficients to promote numerical stability in our method is much more computationally efficient. We show that our approach can efficiently utilize partial computations done by slower worker nodes in a heterogeneous system which can enhance the overall computation speed. Numerical experiments conducted through Amazon Web Services (AWS) demonstrate up to 30% reduction in per worker node computation time and 100x faster encoding compared to the available methods.
In recent years, exciting sources of data have been modeled as knowledge graphs (KGs). This modeling represents both structural relationships and the entity-specific multi-modal data in KGs. In various data analytics ...
详细信息
In recent years, exciting sources of data have been modeled as knowledge graphs (KGs). This modeling represents both structural relationships and the entity-specific multi-modal data in KGs. In various data analytics pipelines and machine learning (ML), the task of semantic similarity estimation plays a significant role. Assigning similarity values to entity pairs is needed in recommendation systems, clustering, classification, entity matching/disambiguation and many others. Efficient and scalable frameworks are needed to handle the quadratic complexity of all-pair semantic similarity on Big Data KGs. Moreover, heterogeneous KGs demand multi-modal semantic similarity estimation to cover the versatile contents like categorical relations between classes or their attribute literals like strings, timestamps or numeric data. In this paper, we propose the SimE4KG framework as a resource providing generic open-source modules that perform semantic similarity estimation in multi-modal KGs. To justify the computational costs of similarity estimation, the SimE4KG generates reproducible, reusable and explainable results. The pipeline results are a native semantic RDF KG, including the experiment results, hyper-parameter setup and explanation of the results, like the most influential features. For fast and scalable execution in memory, we implemented the distributed approach using Apache Spark. The entire development of this framework is integrated into the holistic distributed Semantic ANalytics StAck (SANSA).
Principal component analysis (PCA) is a widely used technique in the field of machine learning and one of the main dimensionality reduction methods. PCA can convert high-dimensional data into lower-dimensional represe...
详细信息
ISBN:
(纸本)9789819777068;9789819777075
Principal component analysis (PCA) is a widely used technique in the field of machine learning and one of the main dimensionality reduction methods. PCA can convert high-dimensional data into lower-dimensional representations, thereby helping to extract important information from data, and is an effective tool for data analysis and pattern recognition. In this paper, we propose a scheme for executing PCA on encrypted data based on homomorphic encryption in a distributed environment. In a distributed setting, computing nodes do not need to perform matrix operations such as matrix multiplication or addition, which reduces the computational burden on the nodes. To ensure the security of the data, we employ the CKKS homomorphic encryption scheme, which allows for approximate calculations on real numbers, meeting the requirements of machine learning. Additionally, we horizontally partition the data to facilitate its application in distributed computing. Furthermore, we optimize the detailed computation of HPCA (Homomorphic PCA), yielding promising results on various datasets.
Pseudospheres are simplicial complexes defined in the late 1990s to model some aspects of distributed systems. Since then, combinatorial properties of pseudospheres combined with topological properties have been very ...
详细信息
MPI+X has been the de facto standard for distributed memory parallel programming. It is widely used primarily as an explicit two-sided communication model, which often leads to complex and error-prone code. Alternativ...
详细信息
ISBN:
(数字)9783031725678
ISBN:
(纸本)9783031725661;9783031725678
MPI+X has been the de facto standard for distributed memory parallel programming. It is widely used primarily as an explicit two-sided communication model, which often leads to complex and error-prone code. Alternatively, PGAS model utilizes efficient one-sided communication and more intuitive communication primitives. In this paper, we present a novel approach that integrates PGAS concepts into the OpenMP programming model, leveraging the LLVM compiler infrastructure and the GASNet-EX communication library. Our model addresses the complexity associated with traditional MPI+OpenMP programming models while ensuring excellent performance and scalability. We evaluate our approach using a set of micro-benchmarks and application kernels on two distinct platforms: Ookami from Stony Brook University and NERSC Perlmutter. The results demonstrate that DiOMP achieves superior bandwidth and lower latency compared to MPI+OpenMP, up to 25% higher bandwidth and down to 45% on latency. DiOMP offers a promising alternative to the traditionalMPI+OpenMP hybrid programmingmodel, towards providing a more productive and efficient way to develop high-performance parallel applications for distributed memory systems.
Researchers increasingly rely on using web-based systems for accessing and running scientific applications across distributed computing resources. However existing systems lack a number of important features, such as ...
详细信息
ISBN:
(纸本)9781467371483
Researchers increasingly rely on using web-based systems for accessing and running scientific applications across distributed computing resources. However existing systems lack a number of important features, such as publication and sharing of scientific applications as online services, decoupling of applications from computing resources and providing remote programmatic access. This paper presents Everest, a web-based platform for researchers supporting publication, execution and composition of applications running across distributed computing resources. Everest addresses the described challenges by relying on modern web technologies and cloud computing models. It follows the Platform as a Service (PaaS) cloud delivery model by providing all its functionality via remote web and programming interfaces. Any application added to Everest is automatically published both as a user-facing web form and a web service. Another distinct feature of Everest is the ability to attach external computing resources by any user and flexibly use these resources for running applications. The paper provides an overview of the platform's architecture and its main components, describes recent developments, presents results of experimental evaluation of the platform and discusses remaining challenges.
distributed machine learning at the network edge has emerged as a promising new paradigm. Various machine learning (ML) technologies will distill Artificial Intelligence (AI) from enormous mobile data to automate futu...
详细信息
distributed machine learning at the network edge has emerged as a promising new paradigm. Various machine learning (ML) technologies will distill Artificial Intelligence (AI) from enormous mobile data to automate future wireless networking and a wide range of Internet-of-Things (IoT) applications. In distributed edge learning, multiple edge devices train a common learning model collaboratively without sending their raw data to a central server, which not only helps to preserve data privacy but also reduces network traffic. However, distributed edge training and edge inference typically still require extensive communications among devices and servers connected by wireless links. As a result, the salient features of wireless networks, including interference and channels’ heterogeneity, time-variability, and unreliability, have significant impacts on the learning performance.
In mobile networks, network performance can be severely degraded by a signaling storm, a phenomenon characterized by control plane (C-plane) congestion due to excessive signaling messages. This paper identifies conges...
详细信息
ISBN:
(纸本)9798350369588;9798350369595
In mobile networks, network performance can be severely degraded by a signaling storm, a phenomenon characterized by control plane (C-plane) congestion due to excessive signaling messages. This paper identifies congestion of inter-site links and overload of a specific NF instance as factors leading to a signaling storm. To address these problems, we propose a combined placement and routing algorithm approach. Firstly, our joint placement and routing algorithms perform geographically distributed deployment of C-plane NF instances according to the fluctuating mobility patterns of each user equipment (UE). Secondly, our routing algorithm identifies an appropriate C-plane NF instance to manage each UE, considering both the utilization rate and the number of contexts possessed by each NF instance. Extensive simulation evaluations demonstrate that our proposed methods significantly reduce inter-site signaling messages compared to existing placement scenarios. Furthermore, the dispersion of C-plane NF instances' utilization rate is reduced, enhancing network performance and efficiency.
暂无评论