the subject of this paper is to show the very high power of asynchronism for iterative algorithms in the context of global computing, that is to say, with machines scattered all around the world. the question is wheth...
详细信息
ISBN:
(纸本)0769516262
the subject of this paper is to show the very high power of asynchronism for iterative algorithms in the context of global computing, that is to say, with machines scattered all around the world. the question is whether or not asynchronism helps to reduce the communication penalty and the overall computation time of a given parallel algorithm. the asynchronous programming model is applied to a given problem implemented with a multi-threaded environment and tested over two kinds of clusters of workstations;a homogeneous local cluster and a heterogeneous non-local one. the main features of this programming model are exhibited and the high efficiency and interest of such algorithms is pointed out.
A good worst-case performance and the availability of high-quality bounds on the worst-case execution time (WCET) of tasks are central for the construction of hard real-time computer systems for safety-critical applic...
详细信息
ISBN:
(纸本)9781479921133
A good worst-case performance and the availability of high-quality bounds on the worst-case execution time (WCET) of tasks are central for the construction of hard real-time computer systems for safety-critical applications. Timing-predictability of the whole software/hardware system is a necessary prerequisite to achieve this. We show that a predictable architecture and the tight and seamless integration of compilation and WCET analysis is beneficial to achieve the initial two goals of good worst-case performance and the availability of high-quality bounds on the WCET of computation tasks. Information generated by the compiler improves the WCET analysis. Detailed timing feedback from the WCET analysis helps the compiler to reduce the worst case execution time. the paper describes the interface and the interaction between the industrial strength WCET analysis tool and the compiler as developed in the EU FP7 T-CREST project, and demonstrates the cooperation of these tools with an illustrative example.
Bioinformatics could greatly benefit from increased computational resources delivered by highperformancecomputing. However, the decision-making about which is the best architecture to deliver good performance for a ...
详细信息
ISBN:
(纸本)9781509024537
Bioinformatics could greatly benefit from increased computational resources delivered by highperformancecomputing. However, the decision-making about which is the best architecture to deliver good performance for a set of Bioinformatics applications is a hard task. the traditional way is finding the architecture with a hightheoretical peak of performance, obtained with benchmark tests. But, this is not an assured way for this decision, because each application of Bioinformatics has different computational requirements, which frequently are much different from usual benchmarks. We developed a methodology that assists researchers, even when their specialty is not highperformancecomputing, to define the best computational infrastructure focused on their set of scientific application requirements. For this purpose, the methodology enables to define representative evaluation tests, including a model to define the correct benchmark, when the tests endorsed by the methodology could not be fully used. Further, a Gain Function allows a reliable decision-making based on the performances of a set of applications and architectures. It is also possible to consider the relative importance between applications and also between cost and performance.
this special issue presents new trends in computerarchitecture and in parallel and distributed systems. It is based on the best papers of the 24th International symposium on computerarchitecture and highperformance...
详细信息
this special issue presents new trends in computerarchitecture and in parallel and distributed systems. It is based on the best papers of the 24th International symposium on computerarchitecture and highperformancecomputing, which was held in New York, NY, USA on October 24-26, 2012 in the Columbia University. the authors were invited to provide extended versions of the papers presented in the conference, taking into account suggestions by the double-blinded peer review process and comments gathered during the conference.
Due to mobility, energy limitations, and unreliable wireless channels, applications running on mobile devices suffer from faults such as temporary disconnection and data loss. We, therefore, need a fault tolerance mec...
详细信息
ISBN:
(纸本)1595936734
Due to mobility, energy limitations, and unreliable wireless channels, applications running on mobile devices suffer from faults such as temporary disconnection and data loss. We, therefore, need a fault tolerance mechanism to guarantee their smooth working and performance. In this paper, we present a novel proxy-based uncoordinated checkpointing scheme with pessimistic message logging for efficient fault recovery in mobile Grid system. Simulation results show that this scheme is reliable, efficient and, at the sametime, consumes less network traffic.
We propose an architecture for a storage system of surveillance videos. Such systems have to handle massive amounts of incoming video streams and relatively few requests for replay. In such a system load (i.e., write ...
详细信息
ISBN:
(纸本)9781479984480
We propose an architecture for a storage system of surveillance videos. Such systems have to handle massive amounts of incoming video streams and relatively few requests for replay. In such a system load (i.e., write requests) scheduling is essential to guarantee performance. Large-scale data-storage system (LSDSS) is an emerging hosting facility for video-storage, which has a very high number of writes while most of the videos are never or rarely watched. We discuss the design and implementation of LSDSS and load scheduling in autonomous storage environments called datacenters in LSDSS. A datacenter (DC) is the basic concept in our LSDSS, which has the self-management system to store data efficiently. A LSDSS consists of many DCs organized in a hierarchy fashion, thereby decentralizing load scheduling tasks. Because DC has a simple design, load scheduling is particularly suited for implementation on a real-time video surveillance and allows to make scheduling decisions. We also discuss experimental results that clearly show the advantage of load scheduling over the widely known base load scheduling.
We present JolokiaC++, an annotation based compiler framework which generates high quality CUDA (Compute Unified Device architecture) code for GPUs. Our contributions include: (1) developing explicit and implicit anno...
详细信息
Traditional network design incorporates a failure-recovery model in order to allow calculation of problems independent of knowledge of the network tool layer this paper explores the possibilities of improving the calc...
详细信息
ISBN:
(纸本)0769516262
Traditional network design incorporates a failure-recovery model in order to allow calculation of problems independent of knowledge of the network tool layer this paper explores the possibilities of improving the calculation throughput by constructing a tool for the specific solution of problems which have an inherent ability to deal with partial calculation failure. Using a modified Genetic Algorithm as the client tool, the amount of information the network layer needs to have is brought to an extremely minimal level;this allows for a large scalability factor of the tool due to the reduction of network management tables.
Beowulf clusters, on face value, offer the potential of a viable cost effective alternative for the provision of highperformancecomputing. In this paper we compare the performance of Beowulf clusters built from comm...
详细信息
ISBN:
(纸本)0769516262
Beowulf clusters, on face value, offer the potential of a viable cost effective alternative for the provision of highperformancecomputing. In this paper we compare the performance of Beowulf clusters built from commodity "off the shelf" components in the support of major research and production codes, with current high-end hardware such as the IBM SP, Compaq AlphaServer SC and SGI Origin 3800. the results concentrate on the application area of computational chemistry. Benchmark data on six commodity-based systems (CS1-CS6) featuring Intel, AMD Athlon and Alpha CPU architectures coupled to traditional Beowulf interconnect, such as Myrinet and Ethernet, are presented. Furthermore, we provide performance data on systems utilising the Quadrics QSNet interconnect technology, and initial results from a prototype of the Cray Supercluster.
Many highperformancecomputing systems today exploit the availability and remarkable performance characteristics of stand alone server systems and the impressive price / performance ratio of commodity components. Sma...
详细信息
ISBN:
(纸本)1595936734
Many highperformancecomputing systems today exploit the availability and remarkable performance characteristics of stand alone server systems and the impressive price / performance ratio of commodity components. Small scale HPC systems, in the range from 16 to 64 processors, have enjoyed significant popularity and are an indispensable tool for the research community. Scaling up to hundreds and thousands of processors, however, has exposed operational issues, which include system availability and reliability. In this paper, we explore the impact of individual component reliability rates on the overall reliability of an HPC system. We derive a mathematical model for determining the failure rate of the system, the probability of failure of a job running on a subset of the system, and show how to design a reasonable queue structure to provide a reliable system over abroad job mix. We also explore the impact of reliability and queue structure on checkpoint intervals and recovery. Our results demonstrate that it is possible to design a reliable highperformancecomputing system with very good operational reliability characteristics from a collection of moderately reliable components. Copyright 2007 ACM.
暂无评论