This paper focuses on the efficient parallel implementation of systems of numerically intensive nature over loosely coupled multiprocessor architectures, These analytical models art of significant importance to many r...
详细信息
This paper focuses on the efficient parallel implementation of systems of numerically intensive nature over loosely coupled multiprocessor architectures, These analytical models art of significant importance to many real-time systems that have to meet severe time constants, A parallel computing engine (PCE) has been developed in this work for the efficient simplification and the near optimal scheduling of numerical models over the different cooperating processors of the parallel computer, First, the analytical system is efficiently coded in its general form, The model is then simplified by using any available information (e.g., constant parameters). A task graph representing the interconnections among the different components (or equations) is generated, The graph can then be compressed to control the computation/communication requirements, The task scheduler employs a graph-based iterative scheme, based on the simulated annealing algorithm, to map the vertices of the task graph onto a Multiple-Instruction-stream Multiple-Data-stream (MIMD) type of architecture. The algorithm uses a nonanalytical cost function that properly considers the computation capability of the processors, the network topology, the communication time, and congestion possibilities, Moreover, the proposed technique is simple, flexible, and computationally viable, The efficiency of the algorithm is demonstrated by two case studies with good results.
作者:
SCHEININE, ALParallel Computing Group
Center for Advanced Studies Research and Development in Sardinia via Nazario Sauro 10 I-09123 Cagliari Italy
An overview is given of parallel computing work being done at CRS4 (Centro di Ricerca, Sviluppo e Studi Superiori in Sardegna). parallel computation projects include: parallelization of a simulation of the interaction...
详细信息
An overview is given of parallel computing work being done at CRS4 (Centro di Ricerca, Sviluppo e Studi Superiori in Sardegna). parallel computation projects include: parallelization of a simulation of the interaction of high energy particles with matter (GEANT), domain decomposition for numerical solution of partial differential equations, seismic migration for oil prospecting, finite-element structural analysis, parallel molecular dynamics, a C++ library for distributed processing of specific functions, and real-time visualization of a computer simulation that runs as distributed processes.
Recent developments in microprocessor technology have led to performance on scalar applications exceeding traditional supercomputers. This suggests that coupling hundreds or even thousands of these ''killer-mi...
详细信息
Recent developments in microprocessor technology have led to performance on scalar applications exceeding traditional supercomputers. This suggests that coupling hundreds or even thousands of these ''killer-micros'' (all working on a single physical problem) may lead to performance on vector applications in excess of vector supercomputers. Also, future generation killer-micros are expected to have vector floating point units as well. The purpose of this paper is to present an overview of the parallel computing environment at Lawrence Liver-more National Laboratory. However, the perspective is necessarily quite narrow and most of the examples are taken from the author's implementation of a large-scale molecular dynamics code on the BBN-TC2000 at LLNL. parallelism is achieved through a geometric domain decomposition - each processor is assigned a distinct region of space and all atoms contained therein. As the atomic positions evolve, the processors must exchange ownership of specific atoms. This geometric domain decomposition proves to be quite general and we highlight its application to image processing and hydrodynamics simulations as well.
• parallelism need not be hard - much easier than traditional concurrent programming • parallel programming, like programming, is a team effort that requires many different skills and many different tools - coarse-lev...
详细信息
The lack of a portable programming interface for parallel computers has inhibited the development of applications for such systems, and thus restricted the exploitation of this technology. In this paper we discuss var...
详细信息
The lack of a portable programming interface for parallel computers has inhibited the development of applications for such systems, and thus restricted the exploitation of this technology. In this paper we discuss various efforts to provide a portable interface to parallel computers, and present the Edinburgh CHIMP (Common High-level Interface to Message-Passing), and PUL (parallel Utilities Library) activities in this context. The concepts behind these projects are described, with details of the range of parallel systems on which they are available. An important consequence of the portability offered by the CHIMP message passing layer is the ability for applications programmers to reuse entire programs, as well as code modules. PUL assists this by providing a set of reusable components for implementing parallel applications. Developed on top of CHIMP, the PUL utilities are themselves portable across a range of platforms. Finally, this paper summarises experience using CHIMP and PUL to implement parallel applications in collaboration with both industrial and academic groups.
The Strategic Defense Initiative (SDI) mid-course tracking problem is a very computationally intensive problem that must be solved in real time. It is generally believed that a real-time solution is only possible thro...
详细信息
ISBN:
(纸本)0818691581
The Strategic Defense Initiative (SDI) mid-course tracking problem is a very computationally intensive problem that must be solved in real time. It is generally believed that a real-time solution is only possible through massively parallel computing. The authors discuss the implementation of the TRC code, a mid-course tracking code, on a 1024 processor nCUBE/ten computer. A brief discussion of the modeling approach used in the TRC is provided, together with a discussion of how the TRC was restructured to make it run efficiently in parallel. The performance of the parallel version of the TRC over a wide range of problem sizes is discussed, and results obtained on a Cray X-MP single processor are compared to the nCUBE/ten results. These results demonstrate the applicability of massively parallel MIMD computing to the SDI mid-course tracking problem.
Experimental research plays an important role in parallel computing, as in this field scientific discovery often relies on experimental findings, which complement and validate theoretical models. However, parallel har...
详细信息
In this paper, we present a general survey on parallel computing. The main contents include parallel computer system which is the hardware platform of parallel computing, parallel algorithm which is the theoretical ba...
详细信息
In this paper, we present a general survey on parallel computing. The main contents include parallel computer system which is the hardware platform of parallel computing, parallel algorithm which is the theoretical base of parallel computing, parallel programming which is the software support of parallel computing. After that, we also introduce some parallel applications and enabling technologies. We argue that parallel computing research should form an integrated methodology of "architecture algorithm programming application". Only in this way, parallel computing research becomes continuous development and more realistic.
作者:
A. COSNUAUP. LECAONERA
Parallel Computing Subdivision B.P. 72 92322 Chatillon Cedex France
This paper describes research conducted at ONERA using the Transputer as building block for experimentation in the field of scientific multiprocessing. The first part describes our experimental systems based on T800–...
详细信息
This paper describes research conducted at ONERA using the Transputer as building block for experimentation in the field of scientific multiprocessing. The first part describes our experimental systems based on T800–20 Transputers and gives performance measurements of some kernel operations. The second part is devoted to the implementation of two simulation codes on a network of forty Transputers. The first concerns a finite element dynamic analysis of a large scale structural problem. The aim is to compute the dynamic response of a clamped cylindrical beam subjected to a time dependent force. The second code simulates the behaviour of particles in an accelerator. This chromaticity correction code, called PATRICIA [17], has been adapted to our multi-Transputer system. All these experiments were conducted using OCCAM2, the parallel programming language of the Transputer.
Between 1987 and 1994, more than 100 students in a broad range of disciplines worked as summer scholars at Edinburgh parallel computing Centre. Many of these students have since taken their parallel computing skills i...
详细信息
ISBN:
(纸本)9780818666056
Between 1987 and 1994, more than 100 students in a broad range of disciplines worked as summer scholars at Edinburgh parallel computing Centre. Many of these students have since taken their parallel computing skills into graduate work and industry, and over a quarter of EPCC's technical staff are alumni of the Programme. This report describes the evolution and present operation of the Summer Scholarship Programme, and its costs and benefits.< >
暂无评论