Douglas Thain, John Bent, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, and Miron Livny WiND and Condor Projects 6 May 2003 Pipeline and Batch Sharing in.

Slides:



Advertisements
Similar presentations
Phillip Dickens, Department of Computer Science, University of Maine. In collaboration with Jeremy Logan, Postdoctoral Research Associate, ORNL. Improving.
Advertisements

1 Storage-Aware Caching: Revisiting Caching for Heterogeneous Systems Brian Forney Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Wisconsin Network Disks University.
Consorzio COMETA - Progetto PI2S2 UNIONE EUROPEA NEMO Monte Carlo Application on the Grid R. Calcagno for the NEMO Collaboration.
RAMADDA for Big Climate Data Don Murray NOAA/ESRL/PSD and CU-CIRES Boulder/Denver Big Data Meetup - June 18, 2014.
IBM Software Group ® Recommending Materialized Views and Indexes with the IBM DB2 Design Advisor (Automating Physical Database Design) Jarek Gryz.
Grids: Why and How (you might use them) J. Templon, NIKHEF VLV T Workshop NIKHEF 06 October 2003.
1 The Case for Versatile Storage System NetSysLab The University of British Columbia Samer Al-Kiswany, Abdullah Gharaibeh, Matei Ripeanu.
Using Application Structure to Handle Failures and Improve Performance in a Migratory File Service John Bent, Douglas Thain, Andrea Arpaci-Dusseau, Remzi.
©Brooks/Cole, 2003 Chapter 5 Computer Organization.
Reliable I/O on the Grid Douglas Thain and Miron Livny Condor Project University of Wisconsin.
Astrophysics, Biology, Climate, Combustion, Fusion, Nanoscience Working Group on Simulation-Driven Applications 10 CS, 10 Sim, 1 VR.
An Introduction to Grid Computing Research at Notre Dame Prof. Douglas Thain University of Notre Dame
The Condor Data Access Framework GridFTP / NeST Day 31 July 2001 Douglas Thain.
8-1 Outline  Overview of Physical Database Design  File Structures  Query Optimization  Index Selection  Additional Choices in Physical Database Design.
Chapter 8 Physical Database Design. McGraw-Hill/Irwin © 2004 The McGraw-Hill Companies, Inc. All rights reserved. Outline Overview of Physical Database.
NPACI: National Partnership for Advanced Computational Infrastructure August 17-21, 1998 NPACI Parallel Computing Institute 1 Cluster Archtectures and.
Layers and Views of a Computer System Operating System Services Program creation Program execution Access to I/O devices Controlled access to files System.
Ian M. Fisk Fermilab February 23, Global Schedule External Items ➨ gLite 3.0 is released for pre-production in mid-April ➨ gLite 3.0 is rolled onto.
Design and Implementation of a Single System Image Operating System for High Performance Computing on Clusters Christine MORIN PARIS project-team, IRISA/INRIA.
Ch 4. The Evolution of Analytic Scalability
Chapter 3 Operating Systems Introduction to CS 1 st Semester, 2015 Sanghyun Park.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Harnessing the Capacity of Computational.
Alok 1Northwestern University Access Patterns, Metadata, and Performance Alok Choudhary and Wei-Keng Liao Department of ECE,
Using Grid Computing in Parallel Electronic Circuit Simulation Marko Dimitrijević FACULTY OF ELECTRONIC ENGINEERING, UNIVERSITY OF NIŠ LABORATORY FOR ELECTRONIC.
Introduction and Overview Questions answered in this lecture: What is an operating system? How have operating systems evolved? Why study operating systems?
DOE BER Climate Modeling PI Meeting, Potomac, Maryland, May 12-14, 2014 Funding for this study was provided by the US Department of Energy, BER Program.
Active Disks: Programming Model, Algorithm and Evaluation Anurag Acharya, Mustafa Uysal, Joel Saltz.
Data Warehousing 1 Lecture-24 Need for Speed: Parallelism Virtual University of Pakistan Ahsan Abdullah Assoc. Prof. & Head Center for Agro-Informatics.
1 Evaluation and Optimization of Multicore Performance Bottlenecks in Supercomputing Applications Jeff Diamond 1, Martin Burtscher 2, John D. McCalpin.
Chapter 3 System Performance and Models. 2 Systems and Models The concept of modeling in the study of the dynamic behavior of simple system is be able.
Workshop on the Future of Scientific Workflows Break Out #2: Workflow System Design Moderators Chris Carothers (RPI), Doug Thain (ND)
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
STORK: Making Data Placement a First Class Citizen in the Grid Tevfik Kosar and Miron Livny University of Wisconsin-Madison March 25 th, 2004 Tokyo, Japan.
Douglas Thain, John Bent Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, Miron Livny Computer Sciences Department, UW-Madison Gathering at the Well: Creating.
LCG Phase 2 Planning Meeting - Friday July 30th, 2004 Jean-Yves Nief CC-IN2P3, Lyon An example of a data access model in a Tier 1.
1 Removing Sequential Bottlenecks in Analysis of Next-Generation Sequencing Data Yi Wang, Gagan Agrawal, Gulcin Ozer and Kun Huang The Ohio State University.
Data-Driven Batch Scheduling John Bent Ph.D. Oral Exam Department of Computer Sciences University of Wisconsin, Madison May 31, 2005.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Deconstructing Storage Arrays Timothy E. Denehy, John Bent, Florentina I. Popovici, Andrea C. Arpaci-Dusseau, Remzi H. Arpaci-Dusseau University of Wisconsin,
Framework for MDO Studies Amitay Isaacs Center for Aerospace System Design and Engineering IIT Bombay.
1 GCA Application in STAR GCA Collaboration Grand Challenge Architecture and its Interface to STAR Sasha Vaniachine presenting for the Grand Challenge.
Transparently Gathering Provenance with Provenance Aware Condor Christine Reilly and Jeffrey Naughton Department of Computer Sciences University of Wisconsin.
Tevfik Kosar Computer Sciences Department University of Wisconsin-Madison Managing and Scheduling Data.
Active Sampling for Accelerated Learning of Performance Models Piyush Shivam, Shivnath Babu, Jeff Chase Duke University.
Flexibility, Manageability and Performance in a Grid Storage Appliance John Bent, Venkateshwaran Venkataramani, Nick Leroy, Alain Roy, Joseph Stanley,
SAGA: Array Storage as a DB with Support for Structural Aggregations SSDBM 2014 June 30 th, Aalborg, Denmark 1 Yi Wang, Arnab Nandi, Gagan Agrawal The.
Storage Research Meets The Grid Remzi Arpaci-Dusseau.
CS 484 Designing Parallel Algorithms Designing a parallel algorithm is not easy. There is no recipe or magical ingredient Except creativity We can benefit.
EGEE-III INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks Abel Carrión Ignacio Blanquer Vicente Hernández.
John Bent Computer Sciences Department University of Wisconsin-Madison Explicit Control in a Batch-aware.
Chapter 8 Physical Database Design. Outline Overview of Physical Database Design Inputs of Physical Database Design File Structures Query Optimization.
OPERATING SYSTEMS CS 3530 Summer 2014 Systems and Models Chapter 03.
PDAC-10 Middleware Solutions for Data- Intensive (Scientific) Computing on Clouds Gagan Agrawal Ohio State University (Joint Work with Tekin Bicer, David.
Explicit Control in a Batch-aware Distributed File System John Bent Douglas Thain Andrea Arpaci-Dusseau Remzi Arpaci-Dusseau Miron Livny University of.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Douglas Thain, John Bent Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, Miron Livny Computer Sciences Department, UW-Madison Gathering at the Well: Creating.
NeST: Network Storage John Bent, Venkateshwaran V Miron Livny, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau.
COMP7500 Advanced Operating Systems I/O-Aware Load Balancing Techniques Dr. Xiao Qin Auburn University
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Research and Service Support Resources for EO data exploitation RSS Team, ESRIN, 23/01/2013 Requirements for a Federated Infrastructure.
Gwangsun Kim, Jiyun Jeong, John Kim
Parallel Programming By J. H. Wang May 2, 2017.
Migratory File Services for Batch-Pipelined Workloads
UNIVERSITY of WISCONSIN-MADISON Computer Sciences Department
Ch 4. The Evolution of Analytic Scalability
NeST: Network Storage Technologies
Performance And Scalability In Oracle9i And SQL Server 2000
Parallel Feature Identification and Elimination from a CFD Dataset
Accelerating Regular Path Queries using FPGA
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

Douglas Thain, John Bent, Andrea Arpaci-Dusseau, Remzi Arpaci-Dusseau, and Miron Livny WiND and Condor Projects 6 May 2003 Pipeline and Batch Sharing in Grid Workloads

Goals › Study diverse range of scientific apps  Measure CPU, memory and I/O demands › Understand relationships btwn apps  Focus is on I/O sharing

Batch-Pipelined workloads › Behavior of single applications has been well studied  sequential and parallel › But many apps are not run in isolation  End result is product of a group of apps  Commonly found in batch systems  Run 100s or 1000s of times › Key is sharing behavior btwn apps

Batch-Pipelined Sharing Pipeline Batch width Shared dataset Pipeline sharing Shared dataset

3 types of I/O › Endpoint: unique input and output › Pipeline: ephemeral data › Batch: shared input data

Outline › Goals and intro › Applications › Methodology › Results › Implications

Six (plus one) target scientific applications › BLAST - biology › IBIS - ecology › CMS - physics › Hartree-Fock - chemistry › Nautilus - molecular dynamics › AMANDA -astrophysics › - astronomy

Common characteristics › Diamond-shaped storage profile › Multi-level working sets  logical collection may be greater than that used by app › Significant data sharing › Commonly submitted in large batches

BLAST search string blastp matches genomic database BLAST searches for matching proteins and nucleotides in a genomic database. Has only a single executable and thus no pipeline sharing.

IBIS inputs analyze forecast climate data IBIS is a global-scale simulation of earth’s climate used to study effects of human activity (e.g. global warming). Only one app thus no pipeline sharing.

CMS configuration cmkin raw events geometry CMS is a two stage pipeline in which the first stage models accelerated particles and the second simulates the response of a detector. This is actually just the first half of a bigger pipeline. cmsim triggered events configuration

Hartree-Fock problem setup initial state HF is a three stage simulation of the non-relativistic interactions between atomic nuclei and electrons. Aside from the executable files, HF has no batch sharing. argos integral scf solutions

Nautilus initial state nautilus intermediate Nautilus is a three stage pipeline which solves Newton’s equation for each molecular particle in a three- dimensional space. The physics which govern molecular interactions is expressed in a shared dataset. The first stage is often repeated multiple times. bin2coord coordinates rasmol visualization physics

AMANDA inputs corsika raw events AMANDA is a four stage astrophysics pipeline designed to observe cosmic events such as gamma-ray bursts. The first stage simulates neutrino production and the creation of muon showers. The second transforms into a standard format and the third and fourth stages follow the muons’ paths through earth and ice. corama standard events mmc noisy events physics mmc triggered events ice tables geometry

work unit setiathome analysis is a single stage pipeline which downloads a work unit of radio telescope “noise” and analyzes it for any possible signs that would indicate extraterrestrial intelligent life. Has no batch data but does have pipeline data as it performs its own checkpointing.

Methodology › CPU behavior tracked with HW counters › Memory tracked with usage statistics › I/O behavior tracked with interposition  mmap was a little tricky › Data collection was easy.  Running the apps was challenge.

Resources Consumed Relatively modest. Max BW is 7 MB/s for HF.

I/O Mix Only IBIS has significant ratio of endpoint I/O.

Observations about individual applications › Modest buffer cache sizes sufficient  Max is AMANDA, needs 500 MB › Large proportion of random access  IBIS, CMS close to 100%, HF ~ 80% › Amdahl and Gray balances skewed  Drastically overprovisioned in terms of I/O bandwidth and memory capacity

Observations about workloads › These apps are NOT run in isolation  Submitted in batches of 100s to 1000s › Large degree of I/O sharing  Significant scalability implications

Scalability of batch width Storage center (1500 MB/s) Commodity disk (15 MB/s)

Batch elimination Storage center (1500 MB/s) Commodity disk (15 MB/s)

Pipeline elimination Storage center (1500 MB/s) Commodity disk (15 MB/s)

Endpoint only Storage center (1500 MB/s) Commodity disk (15 MB/s)

Conclusions › Grid applications do not run in isolation › Relationships btwn apps must be understood › Scalability depends on semantic information  Relationships between apps  Understanding different types of I/O

Questions? › For more information: Douglas Thain, John Bent, Andrea Arpaci- Dusseau, Remzi Arpaci-Dusseau and Miron Livny, Pipeline and Batch Sharing in Grid Workloads, in Proceedings of High Performance Distributed Computing (HPDC- 12). – –