Adiabatic quantum computation has been proposed as quantum parallel processing with adiabatic evolution by using a superposition state to solve combinatorial optimization problem, then it has been applied to many prob...
详细信息
ISBN:
(纸本)9780769548791
Adiabatic quantum computation has been proposed as quantum parallel processing with adiabatic evolution by using a superposition state to solve combinatorial optimization problem, then it has been applied to many problems like satisfiability problem. Among them, Deutsch and Deutsch-Jozsa problems have been tried to be solved by using adiabatic quantum computation. In this paper, we modify the adiabatic quantum computation and propose to solve Bernstein-Vazirani problem more efficiently by a method with higher observation probability.
Statistical performance tuning of a parallel Monte Carlo (MC) radiative transfer code for ocean color (OC) applications is presented. A low observed-to-peak performance ratio due to highly sparse computations is compe...
详细信息
ISBN:
(纸本)9780769548791
Statistical performance tuning of a parallel Monte Carlo (MC) radiative transfer code for ocean color (OC) applications is presented. A low observed-to-peak performance ratio due to highly sparse computations is compensated by online and offline tuning techniques based on a statistical indicator of products accuracy. Run-time adaptive control employs the accuracy indicator to set up two complementary tuning criteria: one general to MC computations and the other specific to OC applications. the same accuracy indicator is also used for pre-execution tuning of a threshold parameter. Numerical simulations of real case scenarios showed that the proposed methods consistently led to faster runs, while satisfying application accuracy requirements. Specifically, speed-ups range from 2.17 to 7.44 times when compared withthe un-optimized version of the MC code. the applied techniques are orthogonal to parallelization, so that the reported performance gains are further amplified by parallel speed-ups.
May-Happen-in-parallel (MHP) analysis is a very important and fundamental mechanism to facilitate concurrent program analysis, optimization and even concurrency bug detection. However, the inefficiency in its design a...
详细信息
ISBN:
(纸本)9780769548791
May-Happen-in-parallel (MHP) analysis is a very important and fundamental mechanism to facilitate concurrent program analysis, optimization and even concurrency bug detection. However, the inefficiency in its design and implementation keeps MHP analysis away from being practical and effective. In this paper, we investigate the state-of-art of iterative data flow based (IDFB) MHP analysis and propose a new design and corresponding systematic implementation. Specifically, we address the most severe efficiency problems in node process order of the work-list in the original approach, and resolve them in our design and implementation by using the concept of parallel level to avoid redundant node visits. Our intensive experimental study shows that the proposed design and implementation have a relative speed up of 29.02x compared withthe original implementation, moreover, it achieves a relative speed up of 10.00x comparing to the state-of-art of non-IDFB approach which is claimed to be more efficient than the original IDFB approach. Our design and implementation are capable of achieving an order of magnitude efficiency improvement comparing to both IDFB and non-IDFB approaches.
An exploit involving the greatest common divisor (GCD) of RSA moduli was recently discovered [1]. this paper presents a tool that can efficiently and completely compare a large number of 1024-bit RSA public keys, and ...
详细信息
ISBN:
(纸本)9780769548791
An exploit involving the greatest common divisor (GCD) of RSA moduli was recently discovered [1]. this paper presents a tool that can efficiently and completely compare a large number of 1024-bit RSA public keys, and identify any keys that are susceptible to this weakness. NVIDIA's graphics processing units (GPU) and the CUDA massively-parallel programming model are powerful tools that can be used to accelerate this tool. Our method using CUDA has a measured performance speedup of 27.5 compared to a sequential CPU implementation, making it a more practical method to compare large sets of keys. A computation for finding GCDs between 200,000 keys, i.e., approximately 20 billion comparisons, was completed in 113 minutes, the equivalent of approximately 2.9 million 1024-bit GCD comparisons per second.
A wide variety of optimization problems requires the combination of Bioinspired and parallelcomputing to address the complexity needed to get optimal solutions in reduced times. the multicore era allows the researche...
详细信息
the inherently limited processing power and battery lifetime of mobile phones hinder the possible execution of computationally intensive applications like content-based video analysis or 3D modeling. Offloading of com...
详细信息
We assume a wireless sensor and actuator network with nodes that can harvest energy from the environment, and an application deployed in this network, which is structured as a set of cooperating mobile components that...
详细信息
ISBN:
(纸本)9780769548791
We assume a wireless sensor and actuator network with nodes that can harvest energy from the environment, and an application deployed in this network, which is structured as a set of cooperating mobile components that can be placed on any node that provides the required sensor and actuator resources. We propose algorithms that take into account the energy consumption rate of agents as well as energy reserves and harvesting rate of nodes, and decide about the migration of agents in order to improve application availability. Initial evaluation results via simulation show that application availability can be greatly improved compared to having a static application placement.
the purpose of this paper is to design and implement a hybrid compiler that combines JOMP's directives with javar's annotations in order to obtain a more performing compiler. this is an original approach and c...
详细信息
ISBN:
(纸本)9780769548791
the purpose of this paper is to design and implement a hybrid compiler that combines JOMP's directives with javar's annotations in order to obtain a more performing compiler. this is an original approach and consists of pooling the advantages of those two compilers while fixing some of their issues. However the achievement of this aim is facing the issue of the difference of implementation of these two compilers because JOMP is implemented in Java while javar is implemented in C language. We propose to entirely re-implement javar in Java by using JavaCC. thereafter, we present the implementation of the hybrid compiler. In the experiments, we propose to parallelize the matrix sort program by using this hybrid compiler. the results of experiments and the mathematical demonstration lead us to state that dealing withthis hybrid compiler gives performances better or equal to the best one between javar and JOMP.
Bag-of-Tasks applications are often composed of a large number of independent tasks;hence, they can easily scale out. With public clouds, the (dynamic) expansion of resource capacity in private clouds is much facilita...
详细信息
ISBN:
(纸本)9780769548791
Bag-of-Tasks applications are often composed of a large number of independent tasks;hence, they can easily scale out. With public clouds, the (dynamic) expansion of resource capacity in private clouds is much facilitated. Clearly, cost efficiently running BoT applications in a multi-cloud environment is of great practical importance. In this paper, we investigate how efficiently multiple clouds can be exploited for running BoT applications and present a fully polynomial time randomized approximation scheme (FPRAS) as a novel task assignment algorithm for BoT applications. the resulting task assignment can be optimized in terms of cost, makespan or the tradeoff between them. the objective function incorporated into our algorithm is devised in the way the optimization objective is tunable based on user preference. Our task assignment decisions are made without any prior knowledge of the processing time of tasks, i.e., non-clairvoyant task assignment. We adopt a Monte Carlo sampling method to estimate unknown task running time. the experimental results shows our algorithm approximates the optimal solution with little overhead.
the proceedings contain 93 papers. the topics discussed include: building input adaptive parallelapplications: a case study of sparse grid interpolation;a lightweight middleware for developing P2P applications with c...
ISBN:
(纸本)9780769549149
the proceedings contain 93 papers. the topics discussed include: building input adaptive parallelapplications: a case study of sparse grid interpolation;a lightweight middleware for developing P2P applications with component and service-based principles;automatic scaling of complex-event processing applications in eucalyptus;DCA: dynamic challenging level adapter for real-time strategy games;a grasp-based heuristic for the project portfolio selection problem;allocation of hard real-time periodic tasks for reliability maximization in distributed systems;automatic clustering assessment through a social tagging system;a new concise representation method of generalized frequent itemsets;parameter sweep and optimization of loosely coupled simulations using the DAKOTA toolkit;implementation of a data-driven workflow management system;and a new approach for indexing powder diffraction data based on dichotomy method.
暂无评论