A Critical Path File Location (CPFL) algorithm for data-aware multiworkflow scheduling on HPC clusters
نویسندگان
چکیده
منابع مشابه
Scheduling File Transfers for Data-Intensive Jobs on Heterogeneous Clusters
This paper addresses the problem of efficient collective scheduling of file transfers requested by a batch of tasks. Our work targets a heterogeneous collection of storage and compute clusters. The goal is to minimize the overall time to transfer files to their respective destination nodes. Two scheduling schemes are proposed and experimentally evaluated against an existing approach, the Insert...
متن کاملOn Job Scheduling for HPC-Clusters and the dynP Scheduler
Efficient job-scheduling strategies are important to improve the performance and usability of HPC-clusters. In this paper we evaluate job-scheduling strategies (FCFS, SJF, and LJF) used in the resource management system CCS (Computing Center Software). As input for our simulations we use two job sets that are generated from trace files of CCS. Based on the evaluation we introduce the dynP sched...
متن کاملE-BaTS: Energy-Aware Scheduling for Bag-of-Task Applications in HPC Clusters
High-Performance Computing (HPC) systems consume large amounts of energy. As the energy consumption predictions for HPC show increasing numbers, it is important to make users aware of the energy spent for the execution of their applications. Drawing from our experience with exposing cost and performance in public clouds, in this paper we present a generic mechanism to compute fast and accurate ...
متن کاملA GPU-Based Enhanced Genetic Algorithm for Power-Aware Task Scheduling Problem in HPC Cloud
In this paper, we consider power-aware task scheduling (PATS) in HPC clouds. Users request virtual machines (VMs) to execute their tasks. Each task is executed on one single VM, and requires a fixed number of cores (i.e., processors), computing power (million instructions per second MIPS) of each core, a fixed start time and non-preemption in a duration. Each physical machine has maximum capaci...
متن کاملScheduling MapReduce Jobs in HPC Clusters
MapReduce (MR) has become a de facto standard for largescale data analysis. Moreover, it has also attracted the attention of the HPC community due to its simplicity, efficiency and highly scalable parallel model. However, MR implementations present some issues that may complicate its execution in existing HPC clusters, specially concerning the job submission. While on MR there are no strict par...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Future Generation Computer Systems
سال: 2017
ISSN: 0167-739X
DOI: 10.1016/j.future.2017.04.025