Data generation, collection, and processing is an important workload of modern computer architectures. Stream or high-intensity data flow applications are commonly employed in extracting and interpreting the informati...
详细信息
Cloud Computing has made possible flexible resources provisioning from an almost unlimited pool. This has created the opportunity to broaden the horizon of data that can be analyzed, allowing to support the so called ...
详细信息
Cryptanalysis of lattice-based cryptography is an important field in cryptography since lattice problems are among the most robust assumptions and have been used to construct a variety of cryptographic primitives. The...
详细信息
Cryptanalysis of lattice-based cryptography is an important field in cryptography since lattice problems are among the most robust assumptions and have been used to construct a variety of cryptographic primitives. The security estimation model for concrete parameters is one of the most important topics in lattice-based cryptography. In this research, we focus on the Gauss Sieve algorithm proposed by Micciancio and Voulgaris, a heuristic lattice sieving algorithm for the central lattice problem, shortest vector problem (SVP). We propose a technique of lifting computations in prime-cyclotomic ideals into that in cyclic ideals. Lifting makes rotations easier to compute and reduces the complexity of inner products from O(n(3)) to O(n(2)). We implemented the Gauss Sieve on multi-GPU systems using two layers of parallelism in our framework, and achieved up to 55 times speed of previous results of dimension 96. We were able to solve SVP on ideal lattice in dimension up to 130, which is the highest dimension SVP instance solved by sieve algorithm so far. As a result, we are able to provide a better estimate of the complexity of solving central lattice problem.
Summary form only given. Personal computing is going mobile and applications are changing to adapt to take advantage of new opportunities offered by permanent availability and connectivity. Mobile devices are a signif...
详细信息
Summary form only given. Personal computing is going mobile and applications are changing to adapt to take advantage of new opportunities offered by permanent availability and connectivity. Mobile devices are a significant departure from traditional computing. On one hand, they are very personal, always on, always connected. They promise to fulfill the promise of being the hub for our digital lives. On the other hand, they are much more constrained in terms of resources than desktops. Even though progress in their computing capabilities has been staggering, they continue to rely on battery power and are packaged in appealing packages that are a nightmare for thermal dissipation. In this talk I will present the challenges facing programmers for mobile devices driven by architectural and packaging constraints, as well as the changes in applications domains. I will give examples on how we used concurrency to improve performance and power efficiency, in a number of projects at Qualcomm Research, including the Zoomm parallel browser.
Adapting sequential legacy software to parallel environments can not only save time and money, but additionally avoids the loss of valuable domain knowledge hidden in existing code. A common parallelization approach i...
详细信息
ISBN:
(纸本)9783030614690;9783030614706
Adapting sequential legacy software to parallel environments can not only save time and money, but additionally avoids the loss of valuable domain knowledge hidden in existing code. A common parallelization approach is the use of standardized parallel design patterns, which allow making best use of parallel programming interfaces such as OpenMP. When such patterns cannot be implemented directly, it can be necessary to apply code transformations beforehand to suitably reshape the input program. In this paper, we describe how we used Abstract Execution, a semi-automatic program proving technique for second-order program properties, to formally prove the conditional correctness of the restructuring techniques CU Repositioning, Loop Splitting and Geometric Decomposition-for all input programs. The latter two techniques require an advanced modeling technique based on families of abstract location sets.
The paper considers the software approach to solving the Fredholm integral equation of the first kind by the Tikhonov regularization method, which uses the properties of modern architectures of computing systems as mu...
详细信息
Reductions constitute a frequent algorithmic pattern in high-performance and scientific computing. Sophisticated techniques are needed to ensure their correct and scalable concurrent execution on modern processors. Re...
详细信息
In 2018, the "Govorun"supercomputer was put into operation at the Laboratory of Information Technologies and is currently used to solve a wide range of tasks facing the Joint Institute for Nuclear Research (...
详细信息
In 2018, the "Govorun"supercomputer was put into operation at the Laboratory of Information Technologies and is currently used to solve a wide range of tasks facing the Joint Institute for Nuclear Research (JINR). The "Govorun"supercomputer is a heterogeneous computing environment that contains Intel CPUs of different types and NVIDIA Tesla V100 graphics accelerators. The heterogeneous structure of the supercomputer allows users to choose optimal computing architectures for solving their tasks. One of the features of modern scientific tasks, both in the field of theoretical studies and tasks related to experimental data processing, is the analysis of large amounts of data. To accelerate the processing of big arrays of data, a hierarchical hyper-converged data processing and storage system with a software-defined architecture was implemented on the "Govorun"supercomputer. According to the speed of accessing data, the system is divided into layers that are available for the user's choice. Each layer of the developed data storage system can be used both independently and as part of data processing workflows. It is noteworthy that a part of the cold storage is managed by the geographically distributed EOS file system, which allows one to connect the data processing and storage system implemented on the "Govorun"supercomputer to geographically distributed storages, the so-called DataLakes. The implemented hierarchical data processing and storage system provides the low time of data access and a data read/write speed of 300 Gb/s. The heterogeneous structure of the supercomputer and the implemented hierarchical data processing and storage system enables the cardinal speed-up of research underway at the Institute. The article describes some examples of using the resources of the "Govorun"supercomputer. The results of studies conducted by different scientific groups using the resources of the supercomputer were published in more than 70 world's leading scientific journals. C
DVM-system is designed for the development of parallel programs of scientific and technical computations in C-DVMH and Fortran-DVMH languages. These languages use a single parallel programming model (DVMH model) and a...
详细信息
Image processing promotes many of the technological advancements these days. The main aspect while performing image processing operations is the time taken to deal with the application of different routines on these i...
详细信息
暂无评论