In this paper we first present briefly QADPZ, an open source platform for heterogeneous desktop grid computing, which enables users from a local network (organization-wide) or Internet (volunteer computing) to share t...
详细信息
ISBN:
(纸本)9789898111517
In this paper we first present briefly QADPZ, an open source platform for heterogeneous desktop grid computing, which enables users from a local network (organization-wide) or Internet (volunteer computing) to share their resources. Users of the system can submit compute-intensive applications to the system, which are then automatically scheduled for execution. The scheduling is made based on the hardware and software requirements of the application. Users can later monitor and control the execution of the applications. Each application consists of one or more tasks. Applications can be independent, when the composing tasks do not require any interaction, or parallel, when the tasks communicate with each other during the computation. QADPZ uses a master worker-model that is improved with some refined capabilities: push of work units, pipelining, sending more work-units at a time, adaptive number of workers, adaptive timeout interval for work units, and use of multithreading, to be presented further in this paper. These improvements are meant to increase the performance and efficiency of such applications.
For many years, data mining has been one of the most important tools in the field of research. The mere reason for the rapid growth in availability of raw data is due to cheap information-sensing devices that are read...
详细信息
ISBN:
(纸本)9781467397544
For many years, data mining has been one of the most important tools in the field of research. The mere reason for the rapid growth in availability of raw data is due to cheap information-sensing devices that are readily available. Big data is a term used to define data sets that are comparatively very big and complex like user database of Facebook, and require a lot of resources to process. MANETs stands for Mobile Ad-hoc Networks and is a comparatively new technology and studies have been in progress to make it a more reliable technology. Message passing interface is a basic portable message-passing interface designed to function on several computers at once. It is regularly integrated where there is the need of high speed computing. This paper describes the use of Message Passing Interface which is in preference to MapReduce in Mobile Ad-Hoc Networks to process large and complex data. Here each node in the given network can use partially reserved CPU cycles and the other nodes in the network are used to carry out processing of data in an affordable manner using various architectures of distributedcomputing. This allows smaller organizations to gather refined information in a much easier way and that too without the need of huge amounts of server to do the same.
This paper proposes a blockchain platform architecture for clinical trial and precision medicine and discusses various design aspects and provides some insights in the technology requirements and challenges. We identi...
详细信息
ISBN:
(纸本)9781538617915
This paper proposes a blockchain platform architecture for clinical trial and precision medicine and discusses various design aspects and provides some insights in the technology requirements and challenges. We identify 44 new system architecture components that are required to be built on top of traditional blockchain and discuss their technology challenges in our blockchain platform: (a) a new blockchain based general distributed and parallel computing paradigm component to devise and study parallelcomputing methodology for big data analytics, (b) blockchain application data management component for data integrity, big data integration, and integrating disparity of medical related data, (c) verifiable anonymous identity management component for identity privacy for both person and Internet of Things (IoT) devices and secure data access to make possible of the patient centric medicine, and (d) trust data sharing management component to enable a trust medical data ecosystem for collaborative research.
Purpose: This study aimed to establish a cloud-based radiotherapy consultation and collaboration system, then investigated the practicability of remote decision support for community radiotherapy centers using the ***...
详细信息
Purpose: This study aimed to establish a cloud-based radiotherapy consultation and collaboration system, then investigated the practicability of remote decision support for community radiotherapy centers using the *** and Materials: A cloud-based consultation and collaboration system for radiotherapy, OncoEvid-ance (R), was developed to provide remote services of LINAC modeling, simulation CT data import/export, target volume and organ-at-risk delineation, prescription, and treatment planning. The system was de-ployed on a hybrid cloud. A federate of public nodes, each corresponding to a medical institution, are managed by a central node where a group of consultants have registered. Users can access the system through network using computing devices. The system has been tested at three community radiother-apy centers. One accelerator was modeled. 12 consultants participated the remote radiotherapy decision support and 77 radiation treatment plans had been evaluated ***: All the passing rates of per-beam dose verification are > 94% and all the passing rates of com-posite beam dose verification are > 99%. The average downloading time for one set of simulation CT data for one patient from Internet was within 1 min under the cloud download bandwidth of 8 Mbps and local network bandwidth of 100 Mbps. The average response time for one consultant to contour target volumes and make prescription was about 24 h. And that for one consultant to design and optimize a IMRT treatment plan was about 36 h. 100% of the remote plans passed the dosimetric criteria and could be imported into the local TPS for further verification. Conclusion: The cloud-based consultation and collaboration system saved the travel time for consultants and provided high quality radiotherapy to patients in community centers. The under-staffed community radiotherapy centers could benefit from the remote system with lower cost and better treatment quality control.(c) 2022 Published by
parallelizing data clustering algorithms has attracted the interest of many researchers over the past few years. Many efficient parallel algorithms were proposed to build partitioning over a huge volume of data. The e...
详细信息
parallelizing data clustering algorithms has attracted the interest of many researchers over the past few years. Many efficient parallel algorithms were proposed to build partitioning over a huge volume of data. The effectiveness of these algorithms is attributed to the distribution of data among a cluster of nodes and to the parallel computation models. Although the effectiveness of parallel models to deal with increasing volume of data little work is done on the validation of big clusters. To deal with this issue, we propose a parallel and scalable model, referred to as S-DI (Scalable Dunn Index), to compute the Dunn Index measure for an internal validation of clustering results. Rather than computing the Dunn Index on a single machine in the clustering validation process, the new proposed measure is computed by distributing the partitioning among a cluster of nodes using a customized parallel model under Apache Spark framework. The proposed S-DI is also enhanced by a Sketch and Validate sampling technique which aims to approximate the Dunn Index value by using a small representative data-sample. Different experiments on simulated and real datasets showed a good scalability of our proposed measure and a reliable validation compared to other existing measures when handling large scale data.
distributed Model Checking ( DMC ) is based on several distributed algorithms, which are often complex and error prone. In this paper, we consider one fundamental aspect of DMC design: message passing communication, t...
详细信息
distributed Model Checking ( DMC ) is based on several distributed algorithms, which are often complex and error prone. In this paper, we consider one fundamental aspect of DMC design: message passing communication, the implementation of which presents hidden tradeoffs often dismissed in DMC related literature. We show that, due to such communication models, high level abstract DMC algorithms might face implicit pitfalls when implemented concretely. We illustrate our discussion with a generic distributed state space generation algorithm.
Anomaly detection aims to separate anomalous pixels from the background, and has become an important application of remotely sensed hyperspectral image processing. Anomaly detection methods based on low-rank and spars...
详细信息
Anomaly detection aims to separate anomalous pixels from the background, and has become an important application of remotely sensed hyperspectral image processing. Anomaly detection methods based on low-rank and sparse representation (LRASR) can accurately detect anomalous pixels. However, with the significant volume increase of hyperspectral image repositories, such techniques consume a significant amount of time (mainly due to the massive amount of matrix computations involved). In this paper, we propose a novel distributedparallel algorithm (DPA) by redesigning key operators of LRASR in terms of MapReduce model to accelerate LRASR on cloud computing architectures. Independent computation operators are explored and executed in parallel on Spark. Specifically, we reconstitute the hyperspectral images in an appropriate format for efficient DPA processing, design the optimized storage strategy, and develop a pre-merge mechanism to reduce data transmission. Besides, a repartitioning policy is also proposed to improve DPA's efficiency. Our experimental results demonstrate that the newly developed DPA achieves very high speedups when accelerating LRASR, in addition to maintaining similar accuracies. Moreover, our proposed DPA is shown to be scalable with the number of computing nodes and capable of processing big hyperspectral images involving massive amounts of data.
This article presents a multi-GPU adaptation of a specific Monte Carlo and classification based method for pricing American basket options, due to Picazo. The first part relates how to combine fine and coarse-grained ...
详细信息
ISBN:
(纸本)9781632662163
This article presents a multi-GPU adaptation of a specific Monte Carlo and classification based method for pricing American basket options, due to Picazo. The first part relates how to combine fine and coarse-grained parallelization to price American basket options. A dynamic strategy of kernel calibration is proposed. Doing so, our implementation on a reasonable size (18) GPU cluster achieves the pricing of a high dimensional (40) option in less than one hour against almost 8 as observed for runs we conducted in the past, using a 64-core cluster (composed of quad-core AMD Opteron 2356). In order to benefit from different GPU device types, we detail the dynamic strategy we have used to load balance GPU calculus which greatly improves the overall pricing time we obtained. An analysis of possible bottleneck effects demonstrates that there is a sequential bottleneck due to the training phase that relies upon the AdaBoost classification method, which prevents the implementation to be fully scalable, and so prevents to envision further decreasing pricing time down to handful of minutes. For this we propose to consider using Random Forests classification method: it is naturally dividable over a cluster, and available like AdaBoost as a black box from the popular Weka machine learning library. However our experimental tests will show that its use is costly.
暂无评论