Retrograde analysis is an efficient exhaustive search method. It is a powerful tool that can be used in solving problems where end states have known values but starting states do not. It has been widely used to solve ...
详细信息
Retrograde analysis is an efficient exhaustive search method. It is a powerful tool that can be used in solving problems where end states have known values but starting states do not. It has been widely used to solve mathematically-precise games such as chess endgames, and is potentially usable in energy-minimization problems. With increasing computing power, both in speed and storage capacity, retrograde analysis will become more and more useful. this paper looks at successful applications to games, the challenges ahead and the modifications that are required to utilize distributed hardware. the power and the usefulness of retrograde analysis are still limited by the computing resources one has access to. Today, the best sequential retrograde algorithms are capable of solving problems with about 10/sup 9/ states in a few hours on a standard personal computer. Bigger problems need more powerful computers, or take much longer to solve, or are simply out of the reach of today's technologies. Introducing parallelism to retrograde analysis is a natural way to attack the bigger problems. there are today three main architectures available for doing parallel retrograde analysis, namely symmetric multiprocessor (SMP) systems, high-speed network-based distributed systems and Internet-based distributed systems. In this paper, we discuss some of the key issues in doing parallel retrograde analysis on these different architectures. Technical challenges are addressed in detail, as well as some examples and proposals. these examples and proposals are drawn from various board games, but the ideas can be applied to other problem domains.
Heterogeneous parallel clusters of workstations are being used to solve many important computational problems. Scheduling parallelapplications an the best collection of machines in a heterogeneous computing environme...
详细信息
ISBN:
(纸本)0769507832;0769507840
Heterogeneous parallel clusters of workstations are being used to solve many important computational problems. Scheduling parallelapplications an the best collection of machines in a heterogeneous computing environment is a complex problem. Performance prediction is vital to good application performance in this environment since utilization of an ill-suited machine can slow the computation down significantly. the heterogeneity of the different pieces composing the parallel platform (network links, CPU, memory, and OS) makes it incredibly difficult to accurately predict performance. this paper addresses the problem of network performance prediction. Since communication speed is often the bottleneck for parallel application performance, network performance prediction is important to the overall performance prediction problem. A new methodology for characterizing network links and application's need for network resources is developed which makes use of Performance Surfaces[3]. Mathematical operations on the performance surfaces are introduced that calculate an application's affinity for a network configuration. these affinity measures can be used for the scheduling of parallelapplications.
A new RAID-x (redundant array of inexpensive disks at level x) architecture is presented for distributed I/O processing on a serverless cluster of computers. the RAID-x architecture is based on a new concept of orthog...
详细信息
ISBN:
(纸本)0769507832;0769507840
A new RAID-x (redundant array of inexpensive disks at level x) architecture is presented for distributed I/O processing on a serverless cluster of computers. the RAID-x architecture is based on a new concept of orthogonal striping and mirroring (OSM) across all distributed disks in the cluster. the primary advantages of this OSM approach lie in: (1) a significant improvement in parallel I/O bandwidth, (2) hiding disk mirroring overhead in the background, and (3) greatly enhanced scalability and reliability in cluster computing applications. All claimed advantages are substantiated with benchmark performance results on the Trojans cluster built at USC in 1999. throughout the paper, we discuss the issues of scalable I/O performance, enhanced system reliability, and striped checkpointing on distributed RAID-x in serverless cluster environment.
the international Workshop on Embedded/distributed HPC Systems and applications (EHPC) is a forum for the presentation and discussion of approaches, research findings, and experiences in the applications of High Perfo...
ISBN:
(纸本)354067442X
the international Workshop on Embedded/distributed HPC Systems and applications (EHPC) is a forum for the presentation and discussion of approaches, research findings, and experiences in the applications of High Performance Computing (HPC) technology for embedded/distributed systems. Of interest are boththe development of relevant technology (e.g.: hardware, middleware, tools) as well as the embedded HPC applications built using such technology.
the proceedings contain 45 papers. the topics discussed include: automatic configuration and run-time adaptation of distributedapplications;a distributed multi-storage resource architecture and 110 performance predic...
ISBN:
(纸本)0769507832
the proceedings contain 45 papers. the topics discussed include: automatic configuration and run-time adaptation of distributedapplications;a distributed multi-storage resource architecture and 110 performance prediction for scientific computing;Uintah: a massively parallel problem solving environment;an enabling framework for master-worker applications on the computational grid;a component based services architecture for building distributedapplications;incorporating job migration and network RAM to share cluster memory resources;using idle workstations to implement predictive prefetching;a monitoring sensor management system for grid environments;robust resource management for metacomputers;performance evaluation of a firewall-compliant Globus-based wide-area cluster system;synchronizing network probes to avoid measurement intrusiveness withthe network weather service;and an evaluation of alternative designs for a grid information service.
Cooperative applications are widely used, e.g. as parallel calculations or distributed information processing systems. Whereby such applications meet the users demand and offer a performance improvement, the susceptib...
详细信息
Cooperative applications are widely used, e.g. as parallel calculations or distributed information processing systems. Whereby such applications meet the users demand and offer a performance improvement, the susceptibility to faults of any used computer node is raised. Often a single fault may cause a complete application failure. On the other hand, we the redundancy in distributed systems can be utilized for fast fault detection and recovery. So, we followed an approach that is based on duplication of each application process to detect crashes and faulty functions of single computer nodes. We concentrate on two aspects of efficient fault-tolerance - fast fault detection and recovery without delaying the application progress significantly. the contribution of this work is first a new fault detecting protocol for duplicated processes. Secondly, we enhance a roll forward recovery scheme so that it is applicable to a set of cooperative processes in conformity to the protocol.
the proceedings contain 12 papers. the topics discussed include: WATIS2: design and application of an environment simulation system for test improvement of control software for automatic logistic systems;preliminary i...
ISBN:
(纸本)0769508375
the proceedings contain 12 papers. the topics discussed include: WATIS2: design and application of an environment simulation system for test improvement of control software for automatic logistic systems;preliminary investigations into distributed computing applications on a beowulf cluster;distributed simulation over loosely coupled domains;distributed engineering plant simuiatiod stimulation environment accreditation of accuracy;scheduling for very large virtual environments and networked games using visibility and priorities;exploiting temporal uncertainty in time warp simulations;dynamic grid-based multicast group assignment in data distribution management;a hybrid approach to data distribution management;state of the HLN RTI;adapting a supply-chain simulation for HLA;a generic rollback manager for optimistic HLA simulations;HLA-based simulation within an interactive engineering environment;and an approach for stream retrieval over HLA-RTI in distributed virtual environments.
this paper examines implementations of a multi-layer perceptron (MLP) on bus-based shared memory (SM) and on distributed memory (DM) multiprocessor systems. the goal has been to optimize HW and SW architectures in ord...
详细信息
this paper examines implementations of a multi-layer perceptron (MLP) on bus-based shared memory (SM) and on distributed memory (DM) multiprocessor systems. the goal has been to optimize HW and SW architectures in order to obtain the fastest response possible. Prototyping parallel MLP algorithms for up to 8 processing nodes withthe DM as well as SM memory was done using CSP-based TRANSIM tool. the results of prototyping MLPs of different sizes on various number of processing nodes demonstrate the feasible speedups, efficiency and time responses for the given CPU speed, link speed or bus bandwidth.
We describe MW - a software framework that allows users to quickly and easily parallelize scientific computations using the master-worker paradigm on the computational grid. MW provides both a 'top level' inte...
详细信息
We describe MW - a software framework that allows users to quickly and easily parallelize scientific computations using the master-worker paradigm on the computational grid. MW provides both a 'top level' interface to application software and a 'bottom level' interface to existing grid computing toolkits. Both interfaces are briefly described. We conclude with a case study, where the necessary Grid services are provided by the Condor high-throughput computing system, and the MW-enabled application code is used to solve a combinatorial optimization problem of unprecedented complexity.
暂无评论