We present an approach to the numerical simulation of dynamically adaptive problems on recursively structured adaptive triangular grids. The intended application is the simulation of oceanic wave propagation (e.g., ts...
详细信息
We present an approach to the numerical simulation of dynamically adaptive problems on recursively structured adaptive triangular grids. The intended application is the simulation of oceanic wave propagation (e.g., tsunami simulation) based on the shallow water equations. For the required 2D dynamically adaptive discretization, we adopt a grid generation process based on recursive bisection of triangles along marked edges. The recursive grid generation may be described via a respective refinement tree, which is sequentialized according to a Sierpinski space-filling curve. This allows a storage scheme for the adaptive grid that requires only a minimal amount of memory. Moreover, the sequentialization and, hence, the locality properties induced by the space-filling curve are retained throughout adaptive refinement and coarsening of the grid. Conforming adaptive refinement and coarsening, as well as time-stepping techniques for time-dependent systems of partial differential equations, are implemented using an inherently cache-efficient processing scheme, which is based on the use of stacks and stream-like data structures and a traversal of the adaptively refined grid along the Sierpinski curve. We demonstrate the computational efficiency of the approach on the solution of a simplified version of the shallow water equations, for which we use a discontinuous Galerkin discretization. Special attention is paid to the memory efficiency of the implementation.
Consider the following abstract coin tossing problem: Given a set of n coins with unknown biases, find the most biased coin using a minimal number of coin tosses. This is a common abstraction of various exploration pr...
详细信息
ISBN:
(纸本)9781450369794
Consider the following abstract coin tossing problem: Given a set of n coins with unknown biases, find the most biased coin using a minimal number of coin tosses. This is a common abstraction of various exploration problems in theoretical computer science and machine learning and has been studied extensively over the years. In particular, algorithms with optimal sample complexity (number of coin tosses) have been known for this problem for quite some time. Motivated by applications to processing massive datasets, we study the space complexity of solving this problem with optimal number of coin tosses in the streaming model. In this model, the coins are arriving one by one and the algorithm is only allowed to store a limited number of coins at any point - any coin not present in the memory is lost and can no longer be tossed or compared to arriving coins. Prior algorithms for the coin tossing problem with optimal sample complexity are based on iterative elimination of coins which inherently require storing all the coins, leading to memory-inefficient streaming algorithms. We remedy this state-of-affairs by presenting a series of improved streaming algorithms for this problem: we start with a simple algorithm which require storing only O(log n) coins and then iteratively refine it further and further, leading to algorithms with O(log log (n)) memory, O(log*(n)) memory, and finally a one that only stores a single extra coin in memory - the same exact space needed to just store the best coin throughout the stream. Furthermore, we extend our algorithms to the problem of finding the k most biased coins as well as other exploration problems such as finding top-k elements using noisy comparisons or finding an e-best arm in stochastic multi-armed bandits, and obtain efficient streaming algorithms for these problems.
We present an approach to the numerical simulation of dynamically adaptive problems on recursively structured adaptive triangular grids. The intended application is the simulation of oceanic wave propagation (e.g., ts...
详细信息
We present an approach to the numerical simulation of dynamically adaptive problems on recursively structured adaptive triangular grids. The intended application is the simulation of oceanic wave propagation (e.g., tsunami simulation) based on the shallow water equations. For the required 2D dynamically adaptive discretization, we adopt a grid generation process based on recursive bisection of triangles along marked edges. The recursive grid generation may be described via a respective refinement tree, which is sequentialized according to a Sierpinski space-filling curve. This allows a storage scheme for the adaptive grid that requires only a minimal amount of memory. Moreover, the sequentialization and, hence, the locality properties induced by the space-filling curve are retained throughout adaptive refinement and coarsening of the grid. Conforming adaptive refinement and coarsening, as well as time-stepping techniques for time-dependent systems of partial differential equations, are implemented using an inherently cache-efficient processing scheme, which is based on the use of stacks and stream-like data structures and a traversal of the adaptively refined grid along the Sierpinski curve. We demonstrate the computational efficiency of the approach on the solution of a simplified version of the shallow water equations, for which we use a discontinuous Galerkin discretization. Special attention is paid to the memory efficiency of the implementation.
A coloring of a graph is an assignment of colors to vertices such that no two neighboring vertices have the same color. The need for memory-efficient coloring algorithms is motivated by their application in computing ...
详细信息
ISBN:
(纸本)9798350387117;9798350387124
A coloring of a graph is an assignment of colors to vertices such that no two neighboring vertices have the same color. The need for memory-efficient coloring algorithms is motivated by their application in computing clique partitions of graphs arising in quantum computations where the objective is to map a large set of Pauli strings into a compact set of unitaries. We present Picasso, a randomized memory-efficient iterative parallel graph coloring algorithm with theoretical sublinear space guarantees under practical assumptions. The parameters of our algorithm provide a trade-off between coloring quality and resource consumption. To assist the user, we also propose a machine learning model to predict the coloring algorithm's parameters considering these trade-offs. We provide a sequential and parallel implementation of the proposed algorithm. We perform an experimental evaluation on a 64-core AMD CPU equipped with 512 GB of memory and an Nvidia A100 GPU with 40GB of memory. For a small dataset where existing coloring algorithms can be executed within the 512 GB memory budget, we show up to 68x memory savings. On massive datasets, we demonstrate that GPU-accelerated Picasso can process inputs with 49.5x more Pauli strings (vertex set in our graph) and 2,478x more edges than state-of-the-art parallel approaches.
暂无评论