DAS-3/Grid’5000 meeting: 4th December 20061 The KOALA Grid Scheduler over DAS-3 and Grid’5000 Processor and data co-allocation in grids Dick Epema, Alexandru.

Slides:



Advertisements
Similar presentations
National Institute of Advanced Industrial Science and Technology Advance Reservation-based Grid Co-allocation System Atsuko Takefusa, Hidemoto Nakada,
Advertisements

A Workflow Engine with Multi-Level Parallelism Supports Qifeng Huang and Yan Huang School of Computer Science Cardiff University
7 april SP3.1: High-Performance Distributed Computing The KOALA grid scheduler and the Ibis Java-centric grid middleware Dick Epema Catalin Dumitrescu,
Resource Management §A resource can be a logical, such as a shared file, or physical, such as a CPU (a node of the distributed system). One of the functions.
Introduction CSCI 444/544 Operating Systems Fall 2008.
GridRPC Sources / Credits: IRISA/IFSIC IRISA/INRIA Thierry Priol et. al papers.
Workshop on HPC in India Grid Middleware for High Performance Computing Sathish Vadhiyar Grid Applications Research Lab (GARL) Supercomputer Education.
June 1, Inter-Operating Grids through Delegated MatchMaking Alexandru Iosup, Dick Epema PDS Group, TU Delft, NL Todd Tannenbaum, Matt Farrellee,
Resource Management of Grid Computing
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
June 3, 2015 Synthetic Grid Workloads with Ibis, K OALA, and GrenchMark CoreGRID Integration Workshop, Pisa A. Iosup, D.H.J. Epema Jason Maassen, Rob van.
2 nd GADA Workshop / OTM 2005 Conferences Eduardo Huedo Rubén S. Montero Ignacio M. Llorente Advanced Computing Laboratory Center for.
Universität Dortmund Robotics Research Institute Information Technology Section Grid Metaschedulers An Overview and Up-to-date Solutions Christian.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
Inter-Operating Grids through Delegated MatchMaking Alexandru Iosup, Dick Epema, Hashim Mohamed,Mathieu Jan, Ozan Sonmez 3 rd Grid Initiative Summer School,
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
A Grid Resource Broker Supporting Advance Reservations and Benchmark- Based Resource Selection Erik Elmroth and Johan Tordsson Reporter : S.Y.Chen.
Computer Science Department 1 Load Balancing and Grid Computing David Finkel Computer Science Department Worcester Polytechnic Institute.
1 Introduction to Load Balancing: l Definition of Distributed systems. Collection of independent loosely coupled computing resources. l Load Balancing.
June 25, GrenchMark: A synthetic workload generator for Grids KOALA Workshop A. Iosup, H. Mohamed, D.H.J. Epema PDS Group, ST/EWI, TU Delft.
June 25, GrenchMark: Synthetic workloads for Grids First Demo at TU Delft A. Iosup, D.H.J. Epema PDS Group, ST/EWI, TU Delft.
June 28, Resource and Test Management in Grids Rapid Prototyping in e-Science VL-e Workshop, Amsterdam, NL Dick Epema, Catalin Dumitrescu, Hashim.
4 december, The Distributed ASCI Supercomputer The third generation Dick Epema (TUD) (with many slides from Henri Bal) Parallel and Distributed.
Workload Management Massimo Sgaravatto INFN Padova.
June 6, 2002D.H.J. Epema/PDS/TUD1 Processor Co-Allocation in Multicluster Systems DAS-2 Workshop Amsterdam June 6, 2002 Anca Bucur and Dick Epema Parallel.
Euro-Par 2008, Las Palmas, 27 August DGSim : Comparing Grid Resource Management Architectures Through Trace-Based Simulation Alexandru Iosup, Ozan.
Resource Management Reading: “A Resource Management Architecture for Metacomputing Systems”
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
Euro-Par 2007, Rennes, 29th August 1 The Characteristics and Performance of Groups of Jobs in Grids Alexandru Iosup, Mathieu Jan *, Ozan Sonmez and Dick.
Self-Organizing Agents for Grid Load Balancing Junwei Cao Fifth IEEE/ACM International Workshop on Grid Computing (GRID'04)
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
 What is an operating system? What is an operating system?  Where does the OS fit in? Where does the OS fit in?  Services provided by an OS Services.
Kento Aida, Tokyo Institute of Technology Grid Challenge - programming competition on the Grid - Kento Aida Tokyo Institute of Technology 22nd APAN Meeting.
Scheduling of Parallel Jobs In a Heterogeneous Multi-Site Environment By Gerald Sabin from Ohio State Reviewed by Shengchao Yu 02/2005.
DISTRIBUTED COMPUTING
WP9 Resource Management Current status and plans for future Juliusz Pukacki Krzysztof Kurowski Poznan Supercomputing.
Meta Scheduling Sathish Vadhiyar Sources/Credits/Taken from: Papers listed in “References” slide.
High Throughput Computing on P2P Networks Carlos Pérez Miguel
Young Suk Moon Chair: Dr. Hans-Peter Bischof Reader: Dr. Gregor von Laszewski Observer: Dr. Minseok Kwon 1.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
A Framework for Elastic Execution of Existing MPI Programs Aarthi Raveendran Tekin Bicer Gagan Agrawal 1.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
Scientific Workflow Scheduling in Computational Grids Report: Wei-Cheng Lee 8th Grid Computing Conference IEEE 2007 – Planning, Reservation,
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
1 Challenge the future KOALA-C: A Task Allocator for Integrated Multicluster and Multicloud Environments Presenter: Lipu Fei Authors: Lipu Fei, Bogdan.
Batch Scheduling at LeSC with Sun Grid Engine David McBride Systems Programmer London e-Science Centre Department of Computing, Imperial College.
Grid Workload Management Massimo Sgaravatto INFN Padova.
Evaluation of Agent Teamwork High Performance Distributed Computing Middleware. Solomon Lane Agent Teamwork Research Assistant October 2006 – March 2007.
BOF: Megajobs Gracie: Grid Resource Virtualization and Customization Infrastructure How to execute hundreds of thousands tasks concurrently on distributed.
Rassul Ayani 1 Performance of parallel and distributed systems  What is the purpose of measurement?  To evaluate a system (or an architecture)  To compare.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
International Symposium on Grid Computing (ISGC-07), Taipei - March 26-29, 2007 Of 16 1 A Novel Grid Resource Broker Cum Meta Scheduler - Asvija B System.
EGEE is a project funded by the European Union under contract IST WS-Based Advance Reservation and Co-allocation Architecture Proposal T.Ferrari,
Introduction to Grid Computing and its components.
Scheduling MPI Workflow Applications on Computing Grids Juemin Zhang, Waleed Meleis, and David Kaeli Electrical and Computer Engineering Department, Northeastern.
Author Utility-Based Scheduling for Bulk Data Transfers between Distributed Computing Facilities Xin Wang, Wei Tang, Raj Kettimuthu,
EGI Technical Forum Amsterdam, 16 September 2010 Sylvain Reynaud.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
Background Computer System Architectures Computer System Software.
INTRODUCTION TO GRID & CLOUD COMPUTING U. Jhashuva 1 Asst. Professor Dept. of CSE.
1 Performance Impact of Resource Provisioning on Workflows Gurmeet Singh, Carl Kesselman and Ewa Deelman Information Science Institute University of Southern.
CPU scheduling.  Single Process  one process at a time  Maximum CPU utilization obtained with multiprogramming  CPU idle :waiting time is wasted 2.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Workload Management Workpackage
Duncan MacMichael & Galen Deal CSS 534 – Autumn 2016
Network Requirements Javier Orellana
Resource and Test Management in Grids
Wide Area Workload Management Work Package DATAGRID project
Presentation transcript:

DAS-3/Grid’5000 meeting: 4th December The KOALA Grid Scheduler over DAS-3 and Grid’5000 Processor and data co-allocation in grids Dick Epema, Alexandru Iosup, Mathieu Jan, Hashim Mohamed, Ozan Sonmez Parallel and Distributed Systems Group

DAS-3/Grid’5000 meeting: 4th December Contents Our context: grid scheduling and co-allocation The design of the KOALA co-allocating scheduler Some performance results KOALA over Grid’5000 and DAS-3 Conclusion & future work

DAS-3/Grid’5000 meeting: 4th December Grid scheduling environment System Grid schedulers usually do not own resources themselves Grid schedulers have to interface to different local schedulers Sun Grid Engine (SGE 6.0 ) on DAS-2/DAS-3 OAR on Grid’5000 Workload Various kind of applications Various requirements

DAS-3/Grid’5000 meeting: 4th December Co-Allocation (1) In grids, jobs may use multiple types of resources in multiple sites: co-allocation or multi-site operation Without co-allocation, a grid is just a big load-sharing device Find suitable candidate system for running a job If the candidate is not suitable anymore, migrate multiple separate jobs grid

DAS-3/Grid’5000 meeting: 4th December Co-Allocation (2) With co-allocation Use available resources (e.g., processors) Access and/or process geographically spread data Application characteristics (e.g., simulation in one location, visualization in another) Problems More difficult resource-discovery process Need to coordinate allocations of local schedulers Slowdown due to wide-area communications single global job grid

DAS-3/Grid’5000 meeting: 4th December A model for co-allocation: schedulers global queue with grid scheduler LS local queues with local schedulers local jobs global job KOALA clusters LS load sharingco-allocation non-local job

DAS-3/Grid’5000 meeting: 4th December A model for co-allocation: job types fixed job flexible job non-fixed job scheduler decides on component placement scheduler decides on split up and placement job components same total job size job component placement fixed

DAS-3/Grid’5000 meeting: 4th December A model for co-allocation: policies Placement policies dictate where the components of a job go Placement policies for non-fixed jobs Load-aware: Worst Fit (WF) (balance load in clusters) Input-file-location-aware: Close-to-Files (CF) (reduce file-transfer times) Communication-aware: Cluster Minimization (CM) (reduce number of wide-area messages) Placement policies for flexible jobs: Communication- and queue time-aware: Flexible Cluster (CM + reduce queue wait time)Minimization (FCM)

DAS-3/Grid’5000 meeting: 4th December KOALA: a Co-Allocating grid scheduler Main goals 1.Processor co-allocation: non-fixed/flexible jobs 2.Data co-allocation: move large input files to the locations where the job components will run prior to execution 3.Load sharing: in the absence of co-allocation KOALA Run alongside local schedulers Scheduler independent from Globus Uses Globus components (e.g., RSL and GridFTP) For launching jobs uses its own mechanisms or Globus DUROC Has been deployed on the DAS2 in September 2005

DAS-3/Grid’5000 meeting: 4th December KOALA: the architecture PIP/NIP: information services RLS: replica location service CO: co-allocator PC: processor claimer RM: run monitor RL: runners listener DM: data manager Ri: runners SGE ?

DAS-3/Grid’5000 meeting: 4th December KOALA: the runners The KOALA runners are adaptation modules for different application types Set up communication Launch applications Current runners KRunner: default KOALA runner that co-allocates processors and that’s it DRunner: DUROC runner for co-allocated MPI applications IRunner: runner for applications using the Ibis Java library for grid applications

DAS-3/Grid’5000 meeting: 4th December KOALA: job flow with four phases new submission place job + _ placement queueclaiming queue + _ Phase 1: job placement Phase 2: file transfer Phase 3: claim processors Phase 4: launch job runners claim processors retry

DAS-3/Grid’5000 meeting: 4th December KOALA: job time line If advanced reservations are not supported, don’t claim processors immediately after placing, but wait until close to the estimated job start time So processors are left idle (processor gained time) Placing and claiming may have to be retried multiple times time job placement estimated start time claiming time estimated file-transfer time processor gained time processor wasted time job submission

DAS-3/Grid’5000 meeting: 4th December KOALA: performance results (1) With replication (3 copies of input files, 2, 4, or 6 GB) Offer a 30% co-allocation load during two hours Try to keep the background load between 30% and 40% time (s) utilization (%) 90 KOALA workload background load processor gained time processor wasted time 1x8 2x8 4x8 1x16 2x16 4x16 job size (number of components X component size) CF placement tries WF placement tries CF claiming tries WF claiming tries 20 See, e.g.: H.H. Mohamed and D.H.J. Epema, “An Evaluation of the Close-to-Files Processor and Data Co-Allocation Policy in Multiclusters,” IEEE Cluster number of tries CF

DAS-3/Grid’5000 meeting: 4th December KOALA: performance results (2) Communication-intensive applications Workload 1: low load Workload 2: high load Background load: 15-20% workload 1workload 2 average wait time (s) average execution time (s) average middleware overhead (s) number of job components workload 1workload 2 See: O. Sonmez, H.H. Mohamed, D.H.J. Epema, Communication-Aware Job-Placement Policies for the KOALA Grid Scheduler, 2 nd IEEE Int’l Conf. on e-Science and Grid Computing, dec

DAS-3/Grid’5000 meeting: 4th December Grid’5000 and DAS-3 interconnection: scheduling issues Preserve each system usage Characterize jobs (especially for Grid’5000) Usage policies Allow simultaneous use of both testbeds One more level of hierarchy in latencies Co-allocation of jobs Various type of applications: PSAs, GridRPC, etc DAS-3DAS-3

DAS-3/Grid’5000 meeting: 4th December KOALA over Grid’5000 and DAS-3 Goal: testing KOALA policies … … in a heterogeneous environment … with different workloads … with OAR reservation capabilities Grid’5000 from DAS-3 “Virtual” clusters inside KOALA Used whenever DAS-3 is overloaded How: deployment of DAS-3 environment on Grid’5000 DAS-3DAS-3

DAS-3/Grid’5000 meeting: 4th December KOALA over Grid’5000 and DAS-3: how DAS-3DAS-3 DAS-3DAS-3 Lyon Orsay Rennes DAS-3DAS-3 file-server OAR DAS-3DAS-3 DAS-3DAS-3 DAS-3DAS-3 …

DAS-3/Grid’5000 meeting: 4th December Using DAS-3 from Grid’5000 Authorize Grid’5000 users to submit jobs … via SGE directly, OARGrid or KOALA Usage policies? Deployment of environments on DAS-3 as in Grid’5000? When: during nights and week-end? Deployment at grid level KOALA submit kadeploy jobs DAS-3DAS-3

DAS-3/Grid’5000 meeting: 4th December Current progress Collected traces of Grid’5000 [done] OAR tables of 15 clusters OARGrid tables LDAP database Analysis in progress KOALA over Grid’5000 [in progress] KOALA communicate with OAR for its information service [done] GRAM interface to OAR “DAS-2” image on Grid’5000: Globus, KOALA, OAR DAS-3DAS-3

DAS-3/Grid’5000 meeting: 4th December Conclusion Use bandwidth and latency in job placements (lightpaths?) Deal with more application types (PSAs, …) A decentralized P2P KOALA Future work KOALA is a grid resource management system Support processor and data co-allocation Several job placement policies (WF, CF, CM, FCM)

DAS-3/Grid’5000 meeting: 4th December Information Publications see PDS publication database at Web site KOALA:

DAS-3/Grid’5000 meeting: 4th December Slowdown due to wide-area communications Co-allocated applications are less efficient due to the relatively slow wide-area communications Extension factor of a job service time on multicluster service time on single cluster Co-allocation is beneficial when the extension factor ≤ 1.20 Unlimited co-allocation is no good Communications libraries may be optimized for wide-area communication (>1 usually) See, e.g.: A.I.D. Bucur and D.H.J. Epema, “Trace-Based Simulations of Processor Co-Allocation Policies in Multiclusters,” HPDC 2003.