This paper describes a system software design for multiple instruction stream control in a massively parallel associative computing environment. The purpose of providing multiple instruction stream control is to incre...
详细信息
This paper describes a system software design for multiple instruction stream control in a massively parallel associative computing environment. The purpose of providing multiple instruction stream control is to increase throughput and reduce the amount of parallel slackness inherent in single instruction stream parallel programming constructs. The multiple associative computing (MASC) model is used to describe this technique and a brief introduction to the MASC model of parallel computation is presented. A simple parallel computing example is used to illustrate the techniques for multiple instruction stream control in a massively parallel runtime environment.
OpenMP has attracted widespread interest because it is an easy-to-use parallel programming model for shared memory multiprocessor systems. Implementation of a "cluster-enabled" OpenMP compiler is presented. ...
详细信息
ISBN:
(纸本)9780769519197
OpenMP has attracted widespread interest because it is an easy-to-use parallel programming model for shared memory multiprocessor systems. Implementation of a "cluster-enabled" OpenMP compiler is presented. Compiled programs are linked to the page-based software distributed-shared-memory system, SCASH, which runs on PC clusters. This allows OpenMP programs to be run transparently in a distributed memory environment. The compiler converts programs written for OpenMP into parallel programs using the SCASH static library, moving all shared global variables into SCASH shared address space at runtime. As data mapping has a great impact on the performance of OpenMP programs compiled for software distributed-shared-memory, extensions to OpenMP directives are defined for specifying data mapping and loop scheduling behavior, allowing data to be allocated to the node where it is to be processed. Experimental results of benchmark programs on PC clusters using both Myrinet and fast Ethernet are reported.
This paper describes a distributed-memory parallel implementation of a time-domain integral-equation (TDIE) based scheme for analyzing a broad class of EMC/EMI problems. The specific TDIE scheme used is the multilevel...
详细信息
This paper describes a distributed-memory parallel implementation of a time-domain integral-equation (TDIE) based scheme for analyzing a broad class of EMC/EMI problems. The specific TDIE scheme used is the multilevel plane-wave time-domain (PWTD) accelerated marching-on-in-time (MOT) solver previously used by our group only on serial computing platforms. The proposed parallel MOT-PWTD solver is applied to sample EMC/EMI problems and the efficiency of the algorithm is investigated via speedup analysis.
Students at the School of Engineering of the University of Buenos Aires are requested, in their senior years from a six year program, to produce some research related activities and to write a thesis. Thesis works are...
详细信息
Students at the School of Engineering of the University of Buenos Aires are requested, in their senior years from a six year program, to produce some research related activities and to write a thesis. Thesis works are regularly procrastinated because of reasons such as a highly unstructured (and thus unfamiliar) activity, lack of background on how to perform research, workplace pressure or lack of time. In addition, there is a social appreciation of research as a highly difficult task reserved only to Nobel Prize winners. Also research activities have been considered as being very different from the daily engineer's work. In one of our courses, students get the big picture of concurrent programming fundamentals and its associated caveats and then, they must cover a more real world approach. We decided to guide them in some information collecting tasks, asking students to use some of the approaches in real world problems and comparing them by developing figures of merit. In the process we found ourselves developing in our students skills which seems to be more transferable than detail level programming skills and more in line with modern software industry requirement.
We propose a set of extensions to the OpenMP programming model to express point-to-point synchronisation schemes. This is accomplished by defining, in the form of directives, precedence relations among the tasks that ...
详细信息
We propose a set of extensions to the OpenMP programming model to express point-to-point synchronisation schemes. This is accomplished by defining, in the form of directives, precedence relations among the tasks that are originated from OpenMP work-sharing constructs. The proposal is based on the definition of a name space that identifies the work parceled out by these work-sharing constructs. Then the programmer defines the precedence relations using this name space. This relieves the programmer from the burden of defining complex synchronization data structures and the insertion of explicit synchronization actions in the program that make the program difficult to understand and maintain. We briefly describe the main aspects of the runtime implementation required to support precedence relations in OpenMP. We focus on the evaluation of the proposal through its use two benchmarks: NAS LU and ASCI Seep3d
In this paper, bye present a new powerful method for parallel program representation called Data Driven Graph (DDG). DDG takes all advantages of classical Directed Acyclic graph (DAC) and adds much more. Simple defini...
详细信息
ISBN:
(纸本)0769505007
In this paper, bye present a new powerful method for parallel program representation called Data Driven Graph (DDG). DDG takes all advantages of classical Directed Acyclic graph (DAC) and adds much more. Simple definition, flexibility and ability to represent loops and dynamically created tasks. With DDG, scheduling becomes an efficient tool for increasing performance of parallel systems. DDG is not only a parallel program model, it also initiates a new parallel programming style, allows programmer to write a parallel program with minimal difficulty. We also present our parallel program development tool with support for DDG and scheduling.
Java's support for parallel and distributed processing makes the language attractive for metacomputing applications, such as parallel applications that run on geographically distributed (wide-area) systems. To obt...
详细信息
Java's support for parallel and distributed processing makes the language attractive for metacomputing applications, such as parallel applications that run on geographically distributed (wide-area) systems. To obtain actual experience with a Java-centric approach to metacomputing, we have built and used a highperformance wide-area Java system, called Manta, Manta implements the Java Remote Method Invocation (RMI) model using different communication protocols (active messages and TCP/IP) for different networks. The paper shows how wide-area parallel applications can be expressed and optimized using Java RMI, Also, it presents performance results of several applications on a wide-area system consisting of four Myrinet-based clusters connected by ATM WANs, We finally discuss alternative programming models, namely object replication, JavaSpaces, and MPI for Java, Copyright (C) 2000 John Wiley & Sons, Ltd.
Janus is a conceptual framework and C++ template library that provides a flexible and extensible collection of efficient data structures and algorithms for a broad class of data-parallel applications. In particular, f...
详细信息
Janus is a conceptual framework and C++ template library that provides a flexible and extensible collection of efficient data structures and algorithms for a broad class of data-parallel applications. In particular, finite difference methods, (adaptive) finite element methods, and data-parallel graph algorithms are supported. An outstanding advantage of providing a generic C++ framework is that it provides application-oriented abstractions that achieve high performance without relying on language extension or non-standard compiler technology. The C++ template mechanism allows to plug user-defined types into the Janus data structures and algorithms. Moreover, Janus components can easily be combined with standard software packages of this field.
暂无评论