Introduction to Work Queue Applications CSE 40822 – Cloud Computing – Fall 2014 Prof. Douglas Thain.

Slides:



Advertisements
Similar presentations
Lecture 12: MapReduce: Simplified Data Processing on Large Clusters Xiaowei Yang (Duke University)
Advertisements

MAP REDUCE PROGRAMMING Dr G Sudha Sadasivam. Map - reduce sort/merge based distributed processing Best for batch- oriented processing Sort/merge is primitive.
Operating Systems Lecture 7.
1 Real-World Barriers to Scaling Up Scientific Applications Douglas Thain University of Notre Dame Trends in HPDC Workshop Vrije University, March 2012.
MapReduce Online Created by: Rajesh Gadipuuri Modified by: Ying Lu.
Experience with Adopting Clouds at Notre Dame Douglas Thain University of Notre Dame IEEE CloudCom, November 2010.
Introduction to Scalable Programming using Makeflow and Work Queue Dinesh Rajan and Mike Albrecht University of Notre Dame October 24 and November 7, 2012.
1 Condor Compatible Tools for Data Intensive Computing Douglas Thain University of Notre Dame Condor Week 2011.
1 Models and Frameworks for Data Intensive Cloud Computing Douglas Thain University of Notre Dame IDGA Cloud Computing 8 February 2011.
Chapter 3: Processes. 3.2 Silberschatz, Galvin and Gagne ©2005 Operating System Concepts Objectives Understand Process concept Process scheduling Creating.
Work Queue: A Scalable Master/Worker Framework Peter Bui June 29, 2010.
CMPT 300: Operating Systems I Ch 3: Processes Dr. Mohamed Hefeeda
Introduction to Makeflow Li Yu University of Notre Dame 1.
L22: SC Report, Map Reduce November 23, Map Reduce What is MapReduce? Example computing environment How it works Fault Tolerance Debugging Performance.
Building Scalable Elastic Applications using Makeflow Dinesh Rajan and Douglas Thain University of Notre Dame Tutorial at CCGrid, May Delft, Netherlands.
Building Scalable Scientific Applications using Makeflow Dinesh Rajan and Peter Sempolinski University of Notre Dame.
Building Scalable Applications on the Cloud with Makeflow and Work Queue Douglas Thain and Patrick Donnelly University of Notre Dame Science Cloud Summer.
Introduction to Makeflow and Work Queue CSE – Cloud Computing – Fall 2014 Prof. Douglas Thain.
© 2011 IBM Corporation 11 April 2011 IDS Architecture.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
Introduction to Parallel Programming MapReduce Except where otherwise noted all portions of this work are Copyright (c) 2007 Google and are licensed under.
Elastic Applications in the Cloud Dinesh Rajan University of Notre Dame CCL Workshop, June 2012.
Süleyman Fatih GİRİŞ CONTENT 1. Introduction 2. Programming Model 2.1 Example 2.2 More Examples 3. Implementation 3.1 ExecutionOverview 3.2.
Massively Parallel Ensemble Methods Using Work Queue Badi’ Abdul-Wahid Department of Computer Science University of Notre Dame CCL Workshop 2012.
SAN DIEGO SUPERCOMPUTER CENTER Working with Inca Reporters Jim Hayes Inca Workshop September 4-5, 2008.
MapReduce: Simplified Data Processing on Large Clusters Jeffrey Dean and Sanjay Ghemawat.
Toward a Common Model for Highly Concurrent Applications Douglas Thain University of Notre Dame MTAGS Workshop 17 November 2013.
Building Scalable Scientific Applications with Makeflow Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts University.
MapReduce How to painlessly process terabytes of data.
Building Scalable Scientific Applications using Makeflow Dinesh Rajan and Douglas Thain University of Notre Dame.
The Cooperative Computing Lab  We collaborate with people who have large scale computing problems in science, engineering, and other fields.  We operate.
Introduction to Scalable Programming using Work Queue Dinesh Rajan and Ben Tovar University of Notre Dame October 10, 2013.
1 Computational Abstractions: Strategies for Scaling Up Applications Douglas Thain University of Notre Dame Institute for Computational Economics University.
Introduction to Work Queue Applications Applied Cyberinfrastructure Concepts Course University of Arizona 2 October 2014 Douglas Thain and Nicholas Hazekamp.
Stuart Wakefield Imperial College London Evolution of BOSS, a tool for job submission and tracking W. Bacchi, G. Codispoti, C. Grandi, INFN Bologna D.
MapReduce Kristof Bamps Wouter Deroey. Outline Problem overview MapReduce o overview o implementation o refinements o conclusion.
Ganga A quick tutorial Asterios Katsifodimos Trainer, University of Cyprus Nicosia, Feb 16, 2009.
Enabling Grids for E-sciencE EGEE-III INFSO-RI Using DIANE for astrophysics applications Ladislav Hluchy, Viet Tran Institute of Informatics Slovak.
Grid Computing at Yahoo! Sameer Paranjpye Mahadev Konar Yahoo!
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Building Scalable Scientific Applications with Work Queue Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts.
Computer Science and Engineering Parallel and Distributed Processing CSE 8380 February Session 11.
The Fresh Breeze Memory Model Status: Linear Algebra and Plans Guang R. Gao Jack Dennis MIT CSAIL University of Delaware Funded in part by NSF HECURA Grant.
Introduction to Makeflow and Work Queue Prof. Douglas Thain, University of Notre Dame
Analyzing LHC Data on 10K Cores with Lobster and Work Queue Douglas Thain (on behalf of the Lobster Team)
By Jeff Dean & Sanjay Ghemawat Google Inc. OSDI 2004 Presented by : Mohit Deopujari.
PVM (Parallel Virtual Machine)‏ By : Vishal Prajapati Course CS683 Computer Architecture Prof. Moreshwar R Bhujade.
Introduction to Scalable Programming using Work Queue Dinesh Rajan and Mike Albrecht University of Notre Dame October 24 and November 7, 2012.
Building Scalable Elastic Applications using Work Queue Dinesh Rajan and Douglas Thain University of Notre Dame Tutorial at CCGrid, May Delft,
Demonstration of Scalable Scientific Applications Peter Sempolinski and Dinesh Rajan University of Notre Dame.
Building Scalable Scientific Applications with Work Queue Douglas Thain and Dinesh Rajan University of Notre Dame Applied Cyber Infrastructure Concepts.
Massively Parallel Molecular Dynamics Using Adaptive Weighted Ensemble Badi’ Abdul-Wahid PI: Jesús A. Izaguirre CCL Workshop 2013.
Introduction to Makeflow and Work Queue Nicholas Hazekamp and Ben Tovar University of Notre Dame XSEDE 15.
1 An unattended, fault-tolerant approach for the execution of distributed applications Manuel Rodríguez-Pascual, Rafael Mayo-García CIEMAT Madrid, Spain.
Instrumenting Badi Abdul-Wahid, RJ Nowling CSE Operating Systems Professor Striegel.
Lecture 3 – MapReduce: Implementation CSE 490h – Introduction to Distributed Computing, Spring 2009 Except as otherwise noted, the content of this presentation.
Introduction to Makeflow and Work Queue
Scaling Up Scientific Workflows with Makeflow
Introduction to Makeflow and Work Queue
Apache Spark Resilient Distributed Datasets: A Fault-Tolerant Abstraction for In-Memory Cluster Computing Aditya Waghaye October 3, 2016 CS848 – University.
Introduction to Makeflow and Work Queue
Introduction to Makeflow and Work Queue
Introduction to Makeflow and Work Queue
Ch 4. The Evolution of Analytic Scalability
Introduction to Makeflow and Work Queue with Containers
What’s New in Work Queue
Creating Custom Work Queue Applications
Overview of Workflows: Why Use Them?
MapReduce: Simplified Data Processing on Large Clusters
Presentation transcript:

Introduction to Work Queue Applications CSE – Cloud Computing – Fall 2014 Prof. Douglas Thain

Quick Recap

3 Makeflow = Make + Workflow Makeflow LocalCondorTorque Work Queue Provides portability across batch systems. Enable parallelism (but not too much!) Trickle out work to batch system. Fault tolerance at multiple scales. Data and resource management.

FutureGrid Torque Cluster Campus Condor Pool Public Cloud Provider Private Cluster Makefile Makeflow Local Files and Programs Makeflow + Work Queue W W W ssh WW W W torque_submit_workers W W W condor_submit_workers W W W Thousands of Workers in a Personal Cloud submit tasks

Today: How to write Work Queue applications directly, without using Makeflow.

Makeflow vs. Work Queue Makeflow – Directed Acyclic Graph programming model. – Static structure known in advance. – All communication through files on disk. Work Queue – Submit-Wait programming model. – Dynamic structure decided at run-time. – Communicate through buffers or files. – More detailed knowledge of how tasks ran.

Sandbox for Each Task 7 worker sim in.txtout.txt put sim.exe put in.txt exec sim.exe out.txt get out.txt X O(1000) workers running on clusters, clouds, and grids. Work Queue System Work Queue Library Work Queue Application C PythonPerl $$$ Master Process submit wait

8 Work Queue API #include “work_queue.h” queue = work_queue_create(); while( not done ) { while (more work ready) { task = work_queue_task_create(); // add some details to the task work_queue_submit(queue, task); } task = work_queue_wait(queue); // process the completed task }

Basic Queue Operations #include “work_queue.h” struct work_queue *queue; struct work_queue_task *task; // Creates a new queue listening on a port, use zero to pick any port. queue = work_queue_create( port ); // Submits a task into a queue. (non-blocking) work_queue_submit( queue, task ); // Waits for a task to complete, returns the complete task. task = work_queue_wait( queue, timeout ); // Returns true if there are no tasks left in the queue. work_queue_empty( queue ); // Returns true if the queue is hungry for more tasks. work_queue_hungry( queue );

Basic Task Operations #include “work_queue.h” struct work_queue_task *task; // Create a task that will run a given Unix command. task = work_queue_task_create( command ); // Indicate an input or output file needed by the task. work_queue_task_specify_file( task, name, remote_name, type, flags ); // Indicate an input buffer needed by the task. work_queue_task_specify_buffer( task, data, length, remote_name, flags); // Destroy the task object. work_queue_task_delete( task );

Run One Task in C #include “work_queue.h” struct work_queue *queue; struct work_queue_task *task; queue = work_queue_create( 0 ); work_queue_specify_name( “myproject” ); task = work_queue_task_create(“sim.exe –p 50 in.dat >out.txt”); /// Missing: Specify files needed by the task. work_queue_submit( queue, task ); while(!work_queue_empty(queue)) { task = work_queue_wait( queue, 60 ); if(task) work_queue_task_delete( task ); }

Run One Task in Perl use work_queue; $queue = work_queue_create( 0 ); work_queue_specify_name( “myproject” ); $task = work_queue_task_create(“sim.exe –p 50 in.dat >out.txt”); ### Missing: Specify files needed by the task. work_queue_submit( $queue, $task ); while(!work_queue_empty($queue)) { $task = work_queue_wait( $queue, 60 ); if($task) work_queue_task_delete( $task ); }

Run One Task in Python from work_queue import * queue = WorkQueue( port = 0 ) queue.specify_name( “myproject” ); task = Task(“sim.exe –p 50 in.dat >out.txt”) ### Missing: Specify files needed by the task. queue.submit( task ) While not queue.empty(): task = queue.wait(60)

C: Specify Files for a Task work_queue_task_specify_file( task,“in.dat”,”in.dat”, WORK_QUEUE_INPUT, WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( task,“calib.dat”,”calib.dat”, WORK_QUEUE_INPUT, WORK_QUEUE_CACHE ); work_queue_task_specify_file( task,“out.txt”,”out.txt”, WORK_QUEUE_OUTPUT, WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( task,“sim.exe”,”sim.exe”, WORK_QUEUE_INPUT, WORK_QUEUE_CACHE ); sim.exe in.dat calib.dat out.txt sim.exe in.dat –p 50 > out.txt

Perl: Specify Files for a Task sim.exe in.dat calib.dat out.txt sim.exe in.dat –p 50 > out.txt work_queue_task_specify_file( $task,“in.dat”,”in.dat”, $WORK_QUEUE_INPUT, $WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( $task,“calib.dat”,”calib.dat”, $WORK_QUEUE_INPUT, $WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( $task,“out.txt”,”out.txt”, $WORK_QUEUE_OUTPUT, $WORK_QUEUE_NOCACHE ); work_queue_task_specify_file( $task,“sim.exe”,”sim.exe”, $WORK_QUEUE_INPUT, $WORK_QUEUE_CACHE );

Python: Specify Files for a Task task.specify_file( “in.dat”, ”in.dat”, WORK_QUEUE_INPUT, cache = False ) task.specify_file( “calib.dat”, ”calib.dat”, WORK_QUEUE_INPUT, cache = False ) task.specify_file( “out.txt”, ”out.txt”, WORK_QUEUE_OUTPUT, cache = False ) task.specify_file( “sim.exe”, ”sim.exe”, WORK_QUEUE_INPUT, cache = True ) sim.exe in.dat calib.dat out.txt sim.exe in.dat –p 50 > out.txt

You must state all the files needed by the command.

Running a Work Queue Program gcc work_queue_example.c -o work_queue_example -I $HOME/cctools/include/cctools -L $HOME/cctools/lib -lwork_queue -ldttools -lm./work_queue_example Listening on port 8374 … In another window:./work_queue_worker studentXX.cse.nd.edu 8374

… for Python setenv PYTHONPATH ${PYTHONPATH}: (no line break) ${HOME}/cctools/lib/python2.6/site-package./work_queue_example.py Listening on port 8374 … In another window:./work_queue_worker studentXX.cse.nd.edu 8374

… for Perl setenv PERL5LIB ${PERL5LIB}: (no line break) ${HOME}/cctools/lib/perl5/site_perl./work_queue_example.py Listening on port 8374 … In another window:./work_queue_worker studentXX.cse.nd.edu 8374

Start Workers Everywhere Submit workers to Condor: condor_submit_workers studentXX.cse.nd.edu Submit workers to SGE: sge_submit_workers studentXX.cse.nd.edu Submit workers to Torque: torque_submit_workers studentXX.cse.nd.edu

Possible Programming Styles Bite off a Piece Run Until Convergence Scatter-Gather Heuristic Tree Search Hill Climbing

More Advanced Features

Use Project Names Worker work_queue_worker –N myproject Catalog connect to studentXX:9037 advertise “myproject” is at studentXX:9037 query Work Queue (port 9037) query work_queue_status

Specify Project Names in Work Queue Specify Project Name for Work Queue master: Python: Perl: C: q.specify_name (“myproject”) work_queue_specify_name ($q, “myproject”); work_queue_specify_name (q, “myproject”);

Start Workers with Project Names Start one worker: $ work_queue_worker -N myproject Start many workers: $ sge_submit_workers -N myproject 5 $ condor_submit_workers -N myproject 5 $ torque_submit_workers -N myproject 5

work_queue_status

Tag a Work Queue task You can specify a tag for a task – Batch together related tasks with a tag – Add a unique identifier as tag to quickly identify on completed. t.specify_tag(“iteration_1”) Python: work_queue_task_specify_tag($t, “iteration_1”); Perl: C: work_queue_task_specify_tag(t, “iteration_1”);

Cancel Work Queue task You can cancel any task anytime after it has been submitted to Work Queue as follows: q.cancel_by_taskid(q, cancel_taskid) q.cancel_by_tasktag(q, “redudant_task”) Python: work_queue_cancel_by_taskid($q, $cancel_taskid); work_queue_cancel_by_tasktag($q, “redudant_task”); Perl: work_queue_cancel_by_taskid(q, cancel_taskid); work_queue_cancel_by_tasktag(q, “redudant_task”); C:

Retry “slow” Work Queue tasks Running on large number of resources – High probability of stragglers – These stragglers slow down completion of your computation Work Queue has a “fast abort” feature to detect stragglers and migrate tasks to other available resources. – It keeps running average of the task execution times. – Activate fast abort by specifying a multiple of the average task execution time – Work Queue will abort any task executing beyond the specified multiple of the task execution time. – The aborted task will then be retried for execution on another available machine.

Activating fast abort in Work Queue #abort if task exceeds 2.5 * avg execution time q.activate_fast_abort (2.5) Python: work_queue_activate_fast_abort ($q, 2.5); Perl: work_queue_activate_fast_abort (q, 2.5); C:

Send intermediate buffer data as input file for Work Queue Task You may have data stored in a buffer (e.g., output of a computation) that will serve as input for a work queue task. To send buffer data as input to a task: buffer = “This is intermediate data” t.specify_buffer(buffer, input.txt, WORK_QUEUE_INPUT, cache=True) Python:

Send intermediate buffer data as input file for Work Queue Task const char * buffer = “This is intermediate buffer data”; int buff_len = strlen(buffer); work_queue_task_specify_buffer(t, buffer, buff_len, “input.txt”, WORK_QUEUE_CACHE); my $buffer = “This is intermediate buffer data” my $buff_len = length($buffer); work_queue_task_specify_buffer($t,$buffer,$buf_len, “input.txt”, $WORK_QUEUE_CACHE); Perl: C:

Work Queue Task Structure The Work Queue task structure contains information about the task and its execution. The following information is available in the task structure: Task command (command line arguments) Task output (stdout) Task return status (exit code of command) Task host (hostname and port on which it ran) Task submit time, Task completion time Task execution time (command) And more..

Accessing Work Queue Task structure # After ‘t’ has been retrieved through q.wait() print % t.output print % t.host Python: # After ‘t’ has been retrieved through work_queue_wait() print “$t->{output}\n”; print “$t->{host})\n”: Perl: // After ‘t’ has been retrieved through work_queue_wait() printf (“%s\n”, t->output); printf (“%s\n”, t->host): C:

Work Queue Statistics Work Queue collects statistics on tasks & workers during run time The statistics can be retrieved anytime during run time Global statistics – Total bytes sent, Total send time – Total bytes received, Total receive time – Total tasks dispatched, Total tasks complete – Total workers joined, Total workers removed Statistics on Tasks – Tasks running, Tasks complete, Tasks waiting Statistics on Workers – Workers ready, workers busy, workers cancelling And more…

Accessing Work Queue Statistics You can access the Work Queue statistics anytime during run time as follows: print q.stats print q.stats.tasks_running Python: my $work_queue_stats = work_queue_stats->new(); work_queue_get_stats ($q, $work_queue_stats); print “$work_queue_stats->{tasks_running}”; Perl:

Work Queue Statistics You can access the Work Queue statistics anytime during run time as follows: struct work_queue_stats *wq_stats; work_queue_get_stats (q, wq_stats); printf (“%d”, work_queue_stats->tasks_running); C:

Thinking Even BIGGER

Managing Your Workforce SGE Cluster Master A Master B Master C Condor Pool W W W W W W Submits new workers. Restarts failed workers. Removes unneeded workers. WQ Pool 100 work_queue_pool –T sge 100 WQ Pool 200 work_queue_pool –T condor 200

Multi-Slot Workers Master Worker work_queue_worker --cores 8 --memory core task 1 core task 1 core task 1 core task 1 core task 4 cores 512 MB specify_cores(4); specify_memory(512); Worker work_queue_worker (implies 1 task, 1 core)

Using Foremen Master T TTTTT T TTTTT W W W W W W W W W Approx X1000 at each fanout. Fore man $$$ work_queue_worker --foreman $MASTER $PORT Fore man $$$ California Chicago

Sample Applications of Work Queue

Replica Exchange T=10KT=20K T=30KT=40K Replica Exchange Work Queue Simplified Algorithm: Submit N short simulations at different temps. Wait for all to complete. Select two simulations to swap. Continue all of the simulations. Dinesh Rajan, Anthony Canino, Jesus A Izaguirre, and Douglas Thain, Converting A High Performance Application to an Elastic Cloud Application, Cloud Com 2011.

Genome Assembly Christopher Moretti, Andrew Thrasher, Li Yu, Michael Olson, Scott Emrich, and Douglas Thain, A Framework for Scalable Genome Assembly on Clusters, Clouds, and Grids, IEEE Transactions on Parallel and Distributed Systems, 2012 Using WQ, we could assemble a human genome in 2.5 hours on a collection of clusters, clouds, and grids with a speedup of 952X. SAND filter master SAND align master Celera Consensus W W W W W W W Sequence Data Modified Celera Assembler

Adaptive Weighted Ensemble 47 Proteins fold into a number of distinctive states, each of which affects its function in the organism. How common is each state? How does the protein transition between states? How common are those transitions?

AWE on Clusters, Clouds, and Grids 48

Work Queue provides the Submit-Wait programming model using a Master-Worker architecture.