Summary form only given. It has been a little less than ten years since modeling and simulation (M&S) hit the knee on the curve. During these past few years a great deal of marketing of the potential of M&S ha...
详细信息
Summary form only given. It has been a little less than ten years since modeling and simulation (M&S) hit the knee on the curve. During these past few years a great deal of marketing of the potential of M&S has occurred, which resulted in a significant influx of funding for research and development (R&D) projects, especially in the area of distributedsimulation. One of the significant experiments in this area-the DARPA Synthetic Theater of War program-officially ended, bringing to an end one of the more robust experiments in distributedsimulation. We are now in a time where most of the M&S funding is targeted at production programs, with much fewer dollars going into R&D or experimentation. Although it is good that major programs are capitalizing on previous R&D efforts, it would not be true to say that the necessary R&D has been completed to realize the vision of distributedsimulation. It would be true to say that we now have a much better understanding of the issues. What is needed now is a period of reflection on the vision, where we are, what was done right, what hasn't worked well, and where we should be headed for the next five to ten years.
Several scheduling algorithms have been proposed to determine the next event to be executed on a processor in a time warp parallel discrete event simulation. However none of them is specifically designed for simulatio...
详细信息
Several scheduling algorithms have been proposed to determine the next event to be executed on a processor in a time warp parallel discrete event simulation. However none of them is specifically designed for simulations where the execution time (or granularity) for different types of events has large variance. We present a grain sensitive scheduling algorithm which addresses this problem. In our solution, the scheduling decision depends on both timestamp and granularity values with the aim at giving higher priority to small grain events even if their timestamp is not the lowest one (i.e. the closest one to the commitment horizon of the simulation). This implicitly limits the optimism of the execution of large grain events that, if rolled back, would produce a large waste of CPU time. The algorithm is adaptive in that it relies on the dynamic recalculation of the length of a simulated time window within which the timestamp of any good candidate event for the scheduling falls in. If the window length is set to zero, then the algorithm behaves like the standard Lowest-Timestamp-First (LTF) scheduling algorithm. simulation results of a classical benchmark in several different configurations are reported for a performance comparison with LTF: these results demonstrate the effectiveness of our algorithm.
We introduce a new time warp system called ROSS: Rensselaer's Optimistic simulation System. ROSS is an extremely modular kernel that is capable of achieving event rates as high as 1,250,000 events per second when ...
详细信息
We introduce a new time warp system called ROSS: Rensselaer's Optimistic simulation System. ROSS is an extremely modular kernel that is capable of achieving event rates as high as 1,250,000 events per second when simulating a wireless telephone network model (PCS) on a quad processor PC server. In a head-to-head comparison, we observe that ROSS out performs the Georgia Tech Time Warp (GTW) system on the same computing platform by up to 180%. ROSS only requires a small constant amount of memory buffers greater than the amount needed by the sequential simulation for a constant number of processors. The driving force behind these high-performance and low memory utilization results is the coupling of an efficient pointer-based implementation framework, Fujimoto's (1989) fast GVT algorithm for shared memory multiprocessors, reverse computation and the introduction of kernel processes (KPs). KPs lower fossil collection overheads by aggregating processed event lists. This aspect allows fossil collection to be done with greater frequency, thus lowering the overall memory necessary to sustain stable, efficient parallel execution.
We examine various modeling and simulation applications of cluster computing using a Beowulf cluster. These applications are used to investigate the performance of our cluster in terms of computational speedup, scalab...
详细信息
We examine various modeling and simulation applications of cluster computing using a Beowulf cluster. These applications are used to investigate the performance of our cluster in terms of computational speedup, scalability, and communications. The applications include solution of linear systems by Jacobi iteration, distributed image generation, and the finite difference time domain solution of Maxwell's equations. It is observed that the computational load for these applications must be large compared to the communication overhead to take advantage of the speedup obtained using parallel computing. For the applications reviewed, this condition is increasingly satisfied as the problem size becomes larger or as higher resolution is required.
Wireless networking technologies and mobile cellular communication systems are growing at an ever faster rate, and this is likely to continue in the foreseeable future. Higher reliability, better coverage and services...
详细信息
Wireless networking technologies and mobile cellular communication systems are growing at an ever faster rate, and this is likely to continue in the foreseeable future. Higher reliability, better coverage and services, higher capacity, mobility management, and wireless multimedia are all parts of the potpourri. The evolution of new systems and improved designs will always depend on the ability to predict system performance using analytical or simulation methods. To date, mathematical analysis has brought some insight into the design of such systems, but analytical methods are often not general or detailed enough for evaluation and comparison of various proposed mobile and/or wireless systems and their services. distributedsimulation techniques have been investigated in a number of studies to decrease the execution times of PCS simulations. We study the load balancing problem for PCS wireless simulation systems, and focus upon static strategies in order to reduce the synchronization overhead of SWiMNet, a parallel PCS simulation testbed developed at UNT.
The introduction of Java in the mid nineties has revolutionised and revitalised interest in the construction of "supercomputers" from "off the shelf" components. Much of the research spawned by the...
详细信息
The introduction of Java in the mid nineties has revolutionised and revitalised interest in the construction of "supercomputers" from "off the shelf" components. Much of the research spawned by the introduction of Java is in the area of metacomputing, where projects that utilise the distributed capabilities of the World Wide Web are becoming more frequent. WebCom is one such system that implements a framework to harness these distributed resources. Although WebCom supports different models of computation, by incorporating them as distinct modular components, an implicitly parallel graph-based model is used. Visual programming tools for simulation steering are employed allowing the creation, submission and dynamic manipulation of programs.
We mainly study the parallelization aspects of the accelerated waveform relaxation algorithms for the transient simulation of semiconductor devices on paralleldistributed memory computers since these methods are comp...
详细信息
We mainly study the parallelization aspects of the accelerated waveform relaxation algorithms for the transient simulation of semiconductor devices on paralleldistributed memory computers since these methods are competitive with standard pointwise methods on serial machines, but are significantly faster on parallel computers. We propose an improved version of the quasi-minimal residual (IQMR) method by using the Lanczos process as a major component combining elements of numerical stability and parallel algorithm design, for solving the resulting sequence of time-varying sparse linear differential-algebraic initial-value problems (IVP) arising at each linearization step. For the Lanczos process stability is obtained by a coupled two-term procedure that generates Lanczos vectors scaled to unit length. The algorithm is derived such that all inner products and matrix-vector multiplications of a single iteration step are independent and communication time required for inner product can be overlapped efficiently with computation time. Therefore, the cost of global communication can be significantly reduced. Experimental results carried out on a Parsytec GC regarding a comparison with other accelerated approaches such as convolution SOR and waveform GMRES techniques on the waveform relaxation algorithm and pointwise methods are also described.
暂无评论