The most important geophysical problems are inverse gravimetry and magnetometry problems. Among them are the structural gravimetry and magnetometry problems of finding interfaces between layers with different densitie...
详细信息
ISBN:
(纸本)9786197105100
The most important geophysical problems are inverse gravimetry and magnetometry problems. Among them are the structural gravimetry and magnetometry problems of finding interfaces between layers with different densities or magnetizations using known gravitational or magnetic data [1], [2], [3]. The gravimetry and magnetometry problems are described by nonlinear integral Fredholm equations of the first kind;they are ill-posed problems. After the discretization of integral operators, the problems are reduced to systems of nonlinear equations with dense matrices. The real gravity and magnetic measurements are carried out over a large area producing large-scale grids. Processing of gravity and magnetic data is time consuming and requires a lot of memory. In this paper, for solving the structural inverse magnetometry problem in a multilayer medium, efficient stable parallel algorithms based on iteratively regularized gradient methods with variable weight factors are proposed. The algorithms were implemented numerically with using new computing technologies on the parallel computing system Uran at the Institute of Mathematics and Mechanics of the UB RAS. The structural magnetometry problem with "quasi-model" data was solved.
In a complete directed weighted graph there are jobs located at nodes of the graph. Job i has an associated processing time or handling time h(i), and the job must start within a prespecified time window [r(i), d(i)]....
详细信息
ISBN:
(纸本)0818691948
In a complete directed weighted graph there are jobs located at nodes of the graph. Job i has an associated processing time or handling time h(i), and the job must start within a prespecified time window [r(i), d(i)]. A vehicle can move on the arcs of the graph, at unit speed, and that has to execute the jobs within their respective time windows. We consider three different problems on the CREW PRAM. (1) Find the minimum cost routes between all pairs of nodes in a network. We give an O(log(3) n) time algorithm with n(4)/log(2) n processors. (2) Services all locations in minimum time. The general problem is NP-complete but O(n(2)) time algorithms are known for a special case;for this case we obtain an O(log(3) n) time parallel algorithm using n(4)/log(2) n processors and a linear time optimal parallel algorithm. (3) Minimize the sum of waiting times at all locations. The general problem is NP-complete but O(n(2)) time algorithm are known for a special case;for this case, we obtain an O(log(2) n) time algorithm with n(3)/log n processors and also a linear time optimal parallel algorithm.
In this article we consider using random mappings to solve sparse binary subset sums via collision search. A mapping is constructed that suits our purpose and two parallel algorithms are proposed based on known collis...
详细信息
ISBN:
(纸本)9781728119441
In this article we consider using random mappings to solve sparse binary subset sums via collision search. A mapping is constructed that suits our purpose and two parallel algorithms are proposed based on known collision-finding techniques. Following the applicability of binary subset sums, results of this paper are relevant to learning parities with noise, decoding random codes and related problems.
We present a new parallel algorithm for computing a maximum cardinality matching in a bipartite graph suitable for distributed memory computers. The presented algorithm is based on the PUSH-RELABEL. algorithm which is...
详细信息
We present a new parallel algorithm for computing a maximum cardinality matching in a bipartite graph suitable for distributed memory computers. The presented algorithm is based on the PUSH-RELABEL. algorithm which is known to be one of the fastest algorithms for the bipartite matching problem. Previous attempts at developing parallel implementations of it have focused on shared memory computers using only a limited number of processors. We first present a straightforward adaptation of these shared memory algorithms to distributed memory computers. However, this is not a viable approach as it requires too much communication. We then develop our new algorithm by modifying the previous approach through a sequence of steps with the main goal being to reduce the amount of communication and to increase load balance. The first goal is achieved by changing the algorithm so that many push and relabel operations can be performed locally between communication rounds and also by selecting augmenting paths that cross processor boundaries infrequently. To achieve good load balance, we limit the speed at which global relabelings traverse the graph. In several experiments on a large number of instances, we study weak and strong scalability of our algorithm using up to 128 processors. The algorithm can also be used to find epsilon-approximate matchings quickly. (C) 2011 Elsevier B.V. All rights reserved.
In this paper we design and analyse parallel algorithms with the goal to get exact bounds on their speed-ups on real machines. For this purpose we employ the BSP model [3] which is an extension of Valiant's BSP mo...
详细信息
ISBN:
(纸本)0818672552
In this paper we design and analyse parallel algorithms with the goal to get exact bounds on their speed-ups on real machines. For this purpose we employ the BSP model [3] which is an extension of Valiant's BSP model [13] and rewards blockwise communication. Further we use Valiant's notion of c-optimality. Intuitively a c-optimal parallel algorithm for p processors tends to speed-up p/c, where the communication time is asymptotically smaller than the computation time. We consider a basic problem in Image Processing, Connected Component Labeling for two and three dimensional images. Our algorithms are randomized and 2-optimal with high probability for a wide range of BSP parameters where the range becomes larger with growing input sizes. Our algorithms improve on previous results as they either need an asymptotically smaller amount of data to be communicated or fewer communication rounds. We further report on implementation work and experiments.
The Lovasz Local Lemma (LLL) is a cornerstone principle in the probabilistic method of combinatorics, and a seminal algorithm of Moser & Tardos (2010) provides an efficient randomized algorithm to implement it. Th...
详细信息
ISBN:
(纸本)9781611974782
The Lovasz Local Lemma (LLL) is a cornerstone principle in the probabilistic method of combinatorics, and a seminal algorithm of Moser & Tardos (2010) provides an efficient randomized algorithm to implement it. This algorithm can be parallelized to give an algorithm that uses polynomially many processors and runs in O(log(3) n) time, stemming from O(log n) adaptive computations of a maximal independent set (MIS). Chung et al. (2014) developed faster local and parallel algorithms, potentially running in time O(log(2) n), but these algorithms work under significantly more stringent conditions than the LLL. We give a new parallel algorithm that works under essentially the same conditions as the original algorithm of Moser & Tardos but uses only a single MIS computation, thus running in O(log(2) n) time. This conceptually new algorithm also gives a clean combinatorial description of a satisfying assignment which might be of independent interest. Our techniques extend to the deterministic LLL algorithm given by Chandrasekaran et al. (2013) leading to an NC-algorithm running in time O(log(2) n) as well. We also provide improved bounds on the runtimes of the sequential and parallel resampling-based algorithms originally developed by Moser & Tardos. Our bounds extend to any problem instance in which the tighter Shearer LLL criterion is satisfied. We also improve on the analysis of Kolipaka & Szegedy (2011) to give tighter concentration results.
Mean-payoff games (MPGs) have many applications, especially in the synthesis, analysis and verification of computer systems. Because of the size of these systems, there is a need to solve very large MPGs. Existing alg...
详细信息
ISBN:
(纸本)9783642041273
Mean-payoff games (MPGs) have many applications, especially in the synthesis, analysis and verification of computer systems. Because of the size of these systems, there is a need to solve very large MPGs. Existing algorithms for solving MPGs are sequential, hence limited by the power of a single computer. In this paper, we propose several parallel algorithms based on the sequential ones. We also evaluate and compare the parallel algorithms experimentally.
Let k be a positive integer, a subset Q of the set of vertices of a graph G is k-dependent in G if each vertex of Q has no more than k neighbours in Q. We present a parallel algorithm which computes a maximal k-depend...
详细信息
ISBN:
(纸本)9783540549451
Let k be a positive integer, a subset Q of the set of vertices of a graph G is k-dependent in G if each vertex of Q has no more than k neighbours in Q. We present a parallel algorithm which computes a maximal k-dependent set in a graph on n nodes in time O(log 4 n) on an EREW PRAM with O(n2) processors. In this way, we establish the membership of the problem of constructing a maximal k-dependent set in the class NC. Our algorithm can be easily adapted to compute a maximal k-dependent set in a graph of bounded valence in time O(log* n) using only O(n) EREW PRAM processors. Let f be a positive integer function defined on the set V of vertices of a graph G. A subset F of the set of edges of G is said to be an f-matching if every vertex v is-an-element-of V is adjacent to at most f(v) edges in F. We present the first Nc algorithm for constructing a maximal f-matching. For a graph on n nodes and m edges the algorithm runs in time O(log4 n) and uses O(n + m) EREW PRAM processors. For graphs of constantly bounded valence, we can construct a maximal f-matching in O(log* n) time on an EREW PRAM with O(n) processors.
Vehicle routing problems involve the navigation of one or more vehicles through a network of locations. Locations have associated handling times as well as time windows during which they are active. The arcs connectin...
详细信息
ISBN:
(纸本)0818680679
Vehicle routing problems involve the navigation of one or more vehicles through a network of locations. Locations have associated handling times as well as time windows during which they are active. The arcs connecting locations have time costs associated with them. In this paper, we consider two different problems in single vehicle routing. The first is to find least time cost routes between all pairs of nodes in a network for navigating vehicles;we call this the all pairs routing problem. We show that there is an O(log2 n) time parallel algorithm using a polynomial number of processors for this problem on a CREW PRAM. We next consider the problem in which a vehicle services all locations in a network. Here, locations can be passed through at any time but only serviced during their time window. The general problem is NP-complete under even fairly stringent restrictions but polynomial algorithms have been developed for some special cases. In particular, when the network is a line, there is no time cost in servicing a location, and all time windows are unbounded at either their lower or upper end, O(n2) algorithms have been developed. We show that under the same conditions, we can reduce this problem to the all pairs routing problem and therefore obtain an O(log2 n) time parallel algorithm on a CREW PRAM.
The objective of this research is to develop parallel algorithms for enabling fast and scalable analysis of large-scale high-throughput sequencing datasets. Genome of an organism consists of one or more long DNA seque...
详细信息
The objective of this research is to develop parallel algorithms for enabling fast and scalable analysis of large-scale high-throughput sequencing datasets. Genome of an organism consists of one or more long DNA sequences called chromosomes, each a sequence of bases. Depending on the organism, the length of the genome can vary from several thousand bases to several billion bases. Genome sequencing, which involves deciphering the sequence of bases of the genome, is an important tool in genomics research. Sequencing instruments widely deployed today can only read short DNA sequences. However, these instruments can read up to several billion such sequences at a time, and are used to sequence a large number of randomly generated short fragments from the genome. These fragments are a few hundred bases long and are commonly referred to as âreadsâ. This work specifically tackles three problems associated with high-throughput sequencing short read datasets: (1) parallel read error correction for large-scale genomics datasets, (2) Partitioning of large-scale high-throughput sequencing datasets, and (3) parallel compression of large-scale genomics datasets.
暂无评论