MapReduce includes three phases of map, shuffle, and reduce. Since the map phase is CPU-intensive and the shuffle phase is I/O-intensive, these phases can be conducted in parallel. This paper studies a joint schedulin...
详细信息
ISBN:
(纸本)9781509022809
MapReduce includes three phases of map, shuffle, and reduce. Since the map phase is CPU-intensive and the shuffle phase is I/O-intensive, these phases can be conducted in parallel. This paper studies a joint scheduling optimization of overlapping map and shuffle phases to minimize the average job makespan. Challenges come from the dependency relationship between map and shuffle phases, since the shuffle phase may wait to transfer the data emitted by the map phase. A new concept of the strong pair is introduced. Two jobs are defined as a strong pair, if the shuffle and map workloads of one job equal the map and shuffle workloads of the other job, respectively. We prove that, if the entire set of jobs can be decomposed to strong pairs of jobs, then the optimal schedule is to pairwisely execute jobs that can form a strong pair. Following the above intuition, several offline and online scheduling policies are proposed. They first group jobs according to job workloads, and then, execute jobs within each group through a pairwise manner. Real data-driven experiments validate the efficiency and effectiveness of the proposed policies.
The increasing scale of multi-core processors are likely to be randomly heterogeneous by design or because of diversity and flaws. The latter type of heterogeneity introduced by some unforeseen variable factors such a...
详细信息
ISBN:
(纸本)9781509040940
The increasing scale of multi-core processors are likely to be randomly heterogeneous by design or because of diversity and flaws. The latter type of heterogeneity introduced by some unforeseen variable factors such as the manufacturing process variation is especially challenging because of its unpredictability. In this environment, thread scheduler and global power manager must handle such randomly heterogeneous. Furthermore, these algorithms must supply high efficiency, scalability and low overhead because future multi-core processors may have a number of cores on a single die. This paper presents a variation-aware scheduling algorithm for application scheduling and power management. Thread switching and sampling among different cores in the multi-core processor introduce obvious overhead than previous many-core scheduling algorithms. Proposed scheme records the information of swapped thread of preferential core and uses tabu search-based randomly heterogeneous scheduling algorithm(TSR) to avoid the occurrence of repeated sampling and reduce the migration frequency and sampling frequency of a thread. The experimental results show that TSR algorithm has decreased 45.7% of thread migration and 42.2% of the sampling time as compared with local search algorithm. This paper regards the transcendental Hungarian offline scheduling algorithm as the baseline. ED2 of TSR only decrease by 8.58% as compared with that of Hungarian offline scheduling algorithm, but compared with the random search scheduling algorithm, ED2 of TSR decreased by 39.4%.
Cloud provides convenient and on demand network access for computing resources available over internet. Individuals and organizations can access the software and hardware such as network, storage, server and applicati...
详细信息
ISBN:
(纸本)9781467382878
Cloud provides convenient and on demand network access for computing resources available over internet. Individuals and organizations can access the software and hardware such as network, storage, server and applications which are located remotely easily with the help of Cloud Service. The tasks/jobs submitted to this cloud environment needs to be executed on time using the resources available so as to achieve proper resource utilization, efficiency and lesser makespan which in turn requires efficient task scheduling algorithm for proper task allocation. In this paper, we have introduced an Optimized Task scheduling Algorithm which adapts the advantages of various other existing algorithms according to the situation while considering the distribution and scalability characteristics of cloud resources.
In Grid computing environment application scheduling is very crucial because the resources are more heterogeneous, geographically distributed, complex and owned by different organizations, they are more prone to failu...
详细信息
ISBN:
(纸本)9781509016679
In Grid computing environment application scheduling is very crucial because the resources are more heterogeneous, geographically distributed, complex and owned by different organizations, they are more prone to failures. Generally, during application/job scheduling only performance factor of resources are considered. But if a node with high computational power also have high failure rate, then there is no such benefit of allocating task to that node because every time a failure occurs it needs recovery and in turn costs in term of time. Thus, failure increases make-span for the job and decreases system/node performance. So, it would be a great idea if we take into consideration failure rate and computational capacity of resources during scheduling. In this paper, to improve the system performance we have proposed a failure-aware scheduling algorithm by taking into consideration both performance and failure factors.
Cloud computing is now trending and more popular in these days for the computation and adopted by many companies like Google, Amazon, Microsoft etc., As the cloud size increases with increase in number of data center ...
详细信息
ISBN:
(纸本)9781509020850
Cloud computing is now trending and more popular in these days for the computation and adopted by many companies like Google, Amazon, Microsoft etc., As the cloud size increases with increase in number of data center power consumption over a data center increases. As number of request over the data center increase with increase in load and power consumption of the data center. So the requests need to be balanced in such manner which having more effective strategy for utilization of resources and reduction in power consumption. Request balancing in such manner without having knowledge of load over server maximize resource utilization but also increasing power consumption at server. So to overcome these issues in cloud Infrastructure as a service (IaaS), we have proposing a trust based scheduling algorithm using ant colony to minimize the load, OoS of the system. Proposed algorithm has proven to have better performance in term of load and reduced request failure as compared to previously proposed scheduling balancing algorithm for cloud IaaS.
Active replication requires deterministic execution in each replica in order to keep them consistent. Debugging and testing need deterministic execution in order to avoid data races and "Heisenbugs". Beside ...
详细信息
Active replication requires deterministic execution in each replica in order to keep them consistent. Debugging and testing need deterministic execution in order to avoid data races and "Heisenbugs". Beside input, multi-threading constitutes a major source of nondeterminism. Several deterministic scheduling algorithms exist that allow concurrent but deterministic executions. Yet, these algorithms seem to be very different. Some of them were even developed without knowing the others. In this paper, we present the novel and flexible Unified Deterministic scheduling algorithm (UDS) for weakly and fully deterministic systems. Compared to existing algorithms, UDS has a broader parameter set, allowing for many configurations that can be used to adapt to a given work load. For the first time, UDS defines reconfiguration of a deterministic scheduler at run-time. Further, we informally show that existing algorithms can be imitated by a particular configuration of UDS, demonstrating its importance.
This paper capitalizes on two emerging trends, i.e., the growing use of wireless at the edge of industrial control networks and the growing interest to integrate IP into said networks. This is facilitated by recent de...
详细信息
This paper capitalizes on two emerging trends, i.e., the growing use of wireless at the edge of industrial control networks and the growing interest to integrate IP into said networks. This is facilitated by recent design contributions from the IEEE and the IETF, where the former developed a highly efficient deterministic time-frequency scheduled medium access control protocol in the form of IEEE 802.15.4e timeslotted channel hopping (TSCH) and the latter IPv6 networking paradigms in the form of 6LoWPAN/ROLL, and scheduling approaches in the form of 6TiSCH. The focus of the present work is on advancing the state-of-the-art of deterministic 6TiSCH schedules toward more flexible but equally reliable distributed approaches. In addition, this paper aims to introduce the first implementation of 6TiSCH networks for factory automation environments: it outlines the challenges faced to overcome the scalability issues inherent to multihop dense low-power networks;the experimental results confirm that the naturally unreliable radio medium can support time-critical and reliable applications. These developments pave the way for wireless industry-grade monitoring approaches.
Task scheduling is one of the core steps to efficiently complete space missions with limited resources in small satellite clusters. This problem becomes more complicated when taking various uncertainties into consider...
详细信息
ISBN:
(纸本)9781467384155
Task scheduling is one of the core steps to efficiently complete space missions with limited resources in small satellite clusters. This problem becomes more complicated when taking various uncertainties into consideration. In this paper, a heuristic robust task scheduling algorithm (RTSA) that handles performance variations of small satellites and task failures has been proposed. An initial solution is obtained by grouping tasks into partial critical paths (PCPs). Then three heuristic strategies and an evaluation function are introduced to improve the schedule iteratively. During this process, the robustness of RTSA is achieved by adding slack time and duplicating critical tasks judiciously. By simulating the Small Satellite Cluster environment and building task failure model, experimental results show that RTSA provides robust and fault-tolerant schedule while minimizing the makespan.
Interconnection network is an inevitable component that is responsible to the system's communication capability. It affects the system-level performance as well as the physical and logical structure of the paralle...
详细信息
ISBN:
(纸本)9781509026562
Interconnection network is an inevitable component that is responsible to the system's communication capability. It affects the system-level performance as well as the physical and logical structure of the parallel system. Many studies are reported to enhance the interconnection network technology, however, we have to further discuss remaining issues for building large-scale systems. One of the most important issues is congestion management. In an interconnection network, packets are transferred simultaneously, and the packets interfere to each other on the network. Congestion arises as a result of the interference among packets. Its fast spreading speed degrades communication performance drastically and it continues for long time. Thus, we should appropriately control the network to suppress the congested situation for maintaining the maximum performance. Many studies address the problem and present effective methods, however, the maximal performance in an ideal situation is not sufficiently clarified. Solving the ideal performance is, in general, an NP-hard problem. This paper introduces particle swarm optimization (PSO) method to overcome the problem. In this paper, we first formalize the optimization problem suitable for the PSO method and present three PSO methods for avoiding local minima. We furthermore introduce some non-PSO methods for comparison. Our preliminary evaluation results reveal high potentials of the PSO method.
Workflow scheduling is recognized as well-known NP-complete problem in the perspective of cloud computing environment. Workflow applications always need high compute-intensive operations because of the presence of pre...
详细信息
ISBN:
(纸本)9781509020300
Workflow scheduling is recognized as well-known NP-complete problem in the perspective of cloud computing environment. Workflow applications always need high compute-intensive operations because of the presence of precedence-constrains. The scheduling objective is to map the workflow application to the VMs pool at available cloud datacenters such that the overall processing time (makespan) is to be minimized and average cloud utilization is maximized. In this paper, we propose a two phase workflow scheduling algorithm with a new priority scheme. It considers the ratio of average communication cost to the average computation cost of the task node as a part of prioritization process in the first phase. Prioritized tasks are mapped to suitable virtual machines in the second phase. Proposed algorithm is capable of scheduling large size workflows in heterogeneous multi-cloud environment. The proposed algorithm is simulated rigorously on standard scientific workflows and simulated results are compared with the existing dependent task scheduling algorithms as per the assumed cloud model. The results remarkably show that the proposed algorithm supercedes the existing algorithms in terms of makespan, speed-up, schedule length ratio and average cloud utilization.
暂无评论