A high-level understanding of how an application executes and which performance characteristics it exhibits is essential in many areas of high performance computing, such as application optimization, hardware developm...
详细信息
ISBN:
(纸本)9783642141218
A high-level understanding of how an application executes and which performance characteristics it exhibits is essential in many areas of high performance computing, such as application optimization, hardware development, and system procurement. Tools are needed to help users in uncovering the application characteristics, but current approaches are unsuitable to help develop a structured understanding of program execution akin to flow charts. Profiling tools are efficient in terms of overheads but their way of recording performance data discards temporal information. Tracing preserves all the temporal information but distilling the essential high level structures, such as initialization and iteration phases can be challenging and cumbersome. We present a technique that extends an existing profiling tool to capture event flow graphs of MPI applications. Event flow graphs try to strike a balance between the abundance of data contained in full traces and the concise information profiling tools can deliver with low overheads. We describe our technique for efficiently gathering an event flow graph for each process of an MPI application and for combining these graphs into a single application-level flow graph. We explore ways to reduce the complexity of the graphs by collapsing nodes in a step-by-step fashion and present techniques to explore flow graphs interactively.
Enormous amount of data is being generated at a tremendous rate by multiple sources, often this data exists in different formats thus making it quite difficult to process the data using traditional methods. The platfo...
详细信息
ISBN:
(纸本)9781509035199
Enormous amount of data is being generated at a tremendous rate by multiple sources, often this data exists in different formats thus making it quite difficult to process the data using traditional methods. The platforms used for processing this type of data rely on distributed architecture like Cloud computing, Hadoop etc. The processing of big data can be efficiently carried out by exploring the characteristics of underlying platforms. With the advent of efficient algorithms, software metrics and by identifying the relationship amongst these measures, system characteristics can be evaluated in order to improve the overall performance of the computing system. By focusing on these measures which play important role in determining the overall performance, service level agreements can also be revised. This paper presents a survey of different performance modeling techniques of big data applications. One of the key concepts in performance modeling is finding relevant parameters which accurately represent performance of big data platforms. These extracted relevant performances measures are mapped onto software qualify concepts which are then used for defining service level agreements.
Stream computing has shown to be an effective technique to decouple communication from computation in many application domains. It provides an efficient mitigation of bandwidth restrictions, by reducing the amount of ...
详细信息
Due to the high frequency in location updates and the expensive cost of continuous query processing, server computation capacity and wireless communication bandwidth are the two limiting factors for large-scale deploy...
详细信息
ISBN:
(纸本)9783540744672
Due to the high frequency in location updates and the expensive cost of continuous query processing, server computation capacity and wireless communication bandwidth are the two limiting factors for large-scale deployment of moving object database systems. Many techniques have been proposed to address the server bottleneck including one using distributed servers. To address both of the scalability factors, P2P computing has been considered. These schemes enable moving objects to participate as a peer in query processing to substantially reduce the demand on server computation, and wireless communications associated with location updates. Most of these techniques, however, assume an open-space environment. In this paper, we investigate a P2P computing technique for continuous kNN queries in a network environment. Since network distance is different from Euclidean distance, techniques designed specifically for an open space cannot be easily adapted for our environment. We present the details of the proposed technique, and discuss our simulation study. The performance results indicate that this technique can significantly reduce server workload and wireless communication costs.
The proposed reactive real-time programming system is a new approach to implement complex distributed heterogeneous real-time applications. It is based on the notion of distributed multi-agent systems. The whole contr...
详细信息
ISBN:
(纸本)9783540634409
The proposed reactive real-time programming system is a new approach to implement complex distributed heterogeneous real-time applications. It is based on the notion of distributed multi-agent systems. The whole control task is decomposed top down into small execution units, called agents which communicate by sending and executing contracts and are specified in a hardware independent language based on states and guarded commands. At compile time the agents are distributed to specified targets. PC's, micro-controllers, programmable logic controllers and even programmable logic devices are supported. The system automatically translates each agent to the particular code and realizes the communication including a bidding protocol between the agents either on the same processor or within a network. Due to a strictly cyclic processing of the agents exact response times can be guaranteed. Zero delay agents can be implemented in hardware.
To execute MPI applications reliably, fault tolerance mechanisms are needed. Message logging is a well known solution to provide fault tolerance for MPI applications. It as been proved that it can tolerate higher fail...
详细信息
ISBN:
(纸本)9783642038686
To execute MPI applications reliably, fault tolerance mechanisms are needed. Message logging is a well known solution to provide fault tolerance for MPI applications. It as been proved that it can tolerate higher failure rate than coordinated checkpointing. However pessimistic and causal message logging can induce high overhead on failure free execution. In this paper, we present O2P, a new optimistic message logging protocol, based oil active optimistic message logging. Contrary to existing optimistic message logging protocols that saves dependency information on reliable storage periodically, O2P logs dependency information as soon as possible to reduce the amount of data piggybacked on application messages. Thus it reduces the overhead of the protocol on failure free execution, making it more scalable and simplifying recovery. O2P is implemented as a module of the Open MPI library. Experiment's show that active message logging is promising to improve scalability and performance of optimistic message logging.
Containerization is a software development approach aimed at packaging an application together with all its dependencies and execution environment in a light-weight, self-contained unit, of which Docker has become the...
详细信息
ISBN:
(纸本)9781538626672
Containerization is a software development approach aimed at packaging an application together with all its dependencies and execution environment in a light-weight, self-contained unit, of which Docker has become the de-facto industry standard. By defining the specific Docker image architecture and building orders, dockerfile plays an important role in the Dockerbased containerization process. Understanding the evolution of dockerfile and which dockerfile architecture attributes enhance dockerfile quality and reduce image build latency can benefit the efficient processing of containerization. In this paper, we perform an empirical study on a large dataset of 2,840 projects to shed light on the impact of dockerfile evolutionary trajectories on quality and latency in the Docker-based containerization. Based on the six categories of dockerfile evolutionary trajectories we discovered, we build two regression models to explore the impact of dockerfile evolutionary trajectories and specific architecture attributes on dockerfile quality and image build latency, which derives a number of suggestions for practitioners.
The proceedings contain 25 papers. The topics discussed include: an evaluation of data-parallel compiler support for line-sweep applications;increasing and detecting memory address congruence;compiler-controlled cachi...
ISBN:
(纸本)0769516203
The proceedings contain 25 papers. The topics discussed include: an evaluation of data-parallel compiler support for line-sweep applications;increasing and detecting memory address congruence;compiler-controlled caching in superword register files for multimedia extension architectures;effective compilation support for variable instruction set architecture;a framework for parallelizing load/stores on embedded processors;workload design: selecting representative program-input pairs;dataflow frequency analysis based on whole program paths;quantifying instruction criticality;the role of computational science in energy efficiency and renewable energy;application transformations for energy and performance-aware device management;and integrating adaptive on-chip storage structures for reduced dynamic power.
Industrial deployment of academic real-time techniques still struggles to gain momentum due to the non-familiarity of the industry with schedulability analysis, as well as the lack of appropriate commercial tools. Mor...
详细信息
ISBN:
(纸本)0769509304
Industrial deployment of academic real-time techniques still struggles to gain momentum due to the non-familiarity of the industry with schedulability analysis, as well as the lack of appropriate commercial tools. Moreover, it is imperative that academia realises the extent of pessimism in the proposed techniques, which often makes them less attractive to systems developers. The possible trade-offs in tinting guarantees vs. reliability is one such key area which needs closer study and scrutiny. There is a need for less stringent guarantees in order to avoid costly overdesigns of systems. In this paper, we present a framework and simulation based methodology for reliability analysis of distributed real-time systems. We have developed a tool which is quite versatile and can accommodate varied task models, network topologies and scheduling paradigms. The tool is illustrated by a comprehensive case-study. Since our method is based on simulation, which is a standard practice in many industrial projects, we believe it will be more comprehensible and acceptable to the industry.
Big Data and its uses are widely used in many applications and fields;artificial information, medical care, business, and much more. Big Data sources are widely distributed and diverse. Therefore, it is essential to g...
详细信息
ISBN:
(纸本)9783031042164;9783031042157
Big Data and its uses are widely used in many applications and fields;artificial information, medical care, business, and much more. Big Data sources are widely distributed and diverse. Therefore, it is essential to guarantee that the data collected and processed is of the highest quality, to deal with this large volume of data from different sources with caution and attention. Consequently, the quality of Big Data must be fulfilled starting from the beginning;data collection. This paper provides a viewpoint on the key Big Data collection Quality Factors that need to be considered every time the data are captured, generated, or created. This study proposes a quality model that can help create and measure data collection methods and techniques. However, the quality model is still introductory and needs to be further investigated.
暂无评论