LEAD-VGrADS Day 1 Notes.

Slides:



Advertisements
Similar presentations
TeraGrid's GRAM Auditing & Accounting, & its Integration with the LEAD Science Gateway Stuart Martin Computation Institute, University of Chicago & Argonne.
Advertisements

LEAD Portal: a TeraGrid Gateway and Application Service Architecture Marcus Christie and Suresh Marru Indiana University LEAD Project (
Legacy code support for commercial production Grids G.Terstyanszky, T. Kiss, T. Delaitre, S. Winter School of Informatics, University.
User Introduction to the TeraGrid 2007 SDSC NCAR TACC UC/ANL NCSA ORNL PU IU PSC.
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
Globus Toolkit 4 hands-on Gergely Sipos, Gábor Kecskeméti MTA SZTAKI
2 nd GADA Workshop / OTM 2005 Conferences Eduardo Huedo Rubén S. Montero Ignacio M. Llorente Advanced Computing Laboratory Center for.
GridFlow: Workflow Management for Grid Computing Kavita Shinde.
A Grid Resource Broker Supporting Advance Reservations and Benchmark- Based Resource Selection Erik Elmroth and Johan Tordsson Reporter : S.Y.Chen.
CONDOR DAGMan and Pegasus Selim Kalayci Florida International University 07/28/2009 Note: Slides are compiled from various TeraGrid Documentations.
DIRAC Web User Interface A.Casajus (Universitat de Barcelona) M.Sapunov (CPPM Marseille) On behalf of the LHCb DIRAC Team.
High Performance Louisiana State University - LONI HPC Enablement Workshop – LaTech University,
Dynamic Firewalls and Service Deployment Models for Grid Environments Gian Luca Volpato, Christian Grimm RRZN – Leibniz Universität Hannover Cracow Grid.
1. Introduction  The JavaScript Grid Portal is trying to find a way to access Grid through Web browser, while using Web 2.0 technologies  The portal.
GRAM: Software Provider Forum Stuart Martin Computational Institute, University of Chicago & Argonne National Lab TeraGrid 2007 Madison, WI.
OGCE Workflow Suite GopiKandaswamy Suresh Marru SrinathPerera ChathuraHerath Marlon Pierce TeraGrid 2008.
CoG Kit Overview Gregor von Laszewski Keith Jackson.
Young Suk Moon Chair: Dr. Hans-Peter Bischof Reader: Dr. Gregor von Laszewski Observer: Dr. Minseok Kwon 1.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Integrating HPC into the ATLAS Distributed Computing environment Doug Benjamin Duke University.
GRAM5 - A sustainable, scalable, reliable GRAM service Stuart Martin - UC/ANL.
1 Overview of the Application Hosting Environment Stefan Zasada University College London.
CSF4 Meta-Scheduler Name: Zhaohui Ding, Xiaohui Wei
Predicting Queue Waiting Time in Batch Controlled Systems Rich Wolski, Dan Nurmi, John Brevik, Graziano Obertelli Computer Science Department University.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Code Applications Tamas Kiss Centre for Parallel.
Tool Integration with Data and Computation Grid GWE - “Grid Wizard Enterprise”
 Apache Airavata Architecture Overview Shameera Rathnayaka Graduate Assistant Science Gateways Group Indiana University 07/27/2015.
Institute For Digital Research and Education Implementation of the UCLA Grid Using the Globus Toolkit Grid Center’s 2005 Community Workshop University.
Cactus/TIKSL/KDI/Portal Synch Day. Agenda n Main Goals:  Overview of Cactus, TIKSL, KDI, and Portal efforts  present plans for each project  make sure.
Grid Execution Management for Legacy Code Applications Grid Enabling Legacy Applications.
Distributed Computing With Triana A Short Course Matthew Shields, Ian Taylor & Ian Wang.
OGCE Components for Enhancing UltraScan Job Management. Suresh Marru,Raminder Singh, Marlon Pierce.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Tool Integration with Data and Computation Grid “Grid Wizard 2”
OGCE Workflow and LEAD Overview Suresh Marru, Marlon Pierce September 2009.
LEAD Project Discussion Presented by: Emma Buneci for CPS 296.2: Self-Managing Systems Source for many slides: Kelvin Droegemeier, Year 2 site visit presentation.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
LEAD Workflow Orchestration Lavanya Ramakrishnan Renaissance Computing Institute University of North Carolina – Chapel Hill Duke University North Carolina.
Lessons from LEAD/VGrADS Demo Yang-suk Kee, Carl Kesselman ISI/USC.
NAREGI PSE with ACS S.Kawata 1, H.Usami 2, M.Yamada 3, Y.Miyahara 3, Y.Hayase 4 1 Utsunomiya University 2 National Institute of Informatics 3 FUJITSU Limited.
Shaowen Wang 1, 2, Yan Liu 1, 2, Nancy Wilkins-Diehr 3, Stuart Martin 4,5 1. CyberInfrastructure and Geospatial Information Laboratory (CIGI) Department.
Joy Rathnayake Senior Architect – Virtusa Pvt. Ltd.
Workload Management Workpackage
GPIR GridPort Information Repository
Simulation Production System
Processes and threads.
WWW and HTTP King Fahd University of Petroleum & Minerals
SC’07 Demo Draft VGrADS Team June 2007.
(on behalf of the POOL team)
Example: Rapid Atmospheric Modeling System, ColoState U
I2G CrossBroker Enol Fernández UAB
Shaowen Wang1, 2, Yan Liu1, 2, Nancy Wilkins-Diehr3, Stuart Martin4,5
Grid Computing.
Deploying and Configuring SSIS Packages
Dirk Düllmann CERN Openlab storage workshop 17th March 2003
Building Grids with Condor
Notification Service May 19, 2006 Jon Atherton Mark Mara.
Globus Job Management. Globus Job Management Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
Liang Chen Advisor: Gagan Agrawal Computer Science & Engineering
Open Grid Computing Environments
Lecture Topics: 11/1 General Operating System Concepts Processes
GATES: A Grid-Based Middleware for Processing Distributed Data Streams
Ainsley Smith Tel: Ex
Overview of Workflows: Why Use Them?
OGCE Portal Software for Big Red and the TeraGrid
Grid Systems: What do we need from web service standards?
Grid Computing Software Interface
GGF10 Workflow Workshop Summary
Condor-G: An Update.
L. Glimcher, R. Jin, G. Agrawal Presented by: Leo Glimcher
Presentation transcript:

LEAD-VGrADS Day 1 Notes

Introduction & discussion Workflow discussion Interaction points Higher level resource requirements Where does bind happen? WCS is being developed right now, timeline not clear Does vgLaunch need to be retrofitted for a web service call? LEAD is BPEL service based – for FT, adaptation, etc GT4 toolkit is also web service based

LEAD is moving to GT4 WS-GRAM, notifications are in different format from LEAD workflow LEAD has focus on deployment and integration, very application focused LEAD might use more than one scheduler/resource solutions ISI (Ann/Ewa) are working with LEAD on the data side (DRS, RLS), domain side (THREDS) WCS handles application configuration such as namelist files

What do each of us need? Interfaces between LEAD and VGrADS? Performance Model? What in LEAD communicates with VGrADS? And what does it give to VGrADS? Interfaces between vgES and batch queue prediction What requirements does LEAD have and does it conflict with vgES philosophy? New features to implement in vgES inline with deadline for SC06

Short-term and long-term integration plan How do I use vgES? From LEAD workflow – concrete example of LEAD workflow? How to use batch queue from vgES? Batch queue waiting time aware Interface points are crucial Full ahead planning – resources might not be available at this time. In LEAD the first few tasks are light, last are intensive. So not much time lead time LEAD today will launch immediately. What if you start running before or after you expect to?

There is post-processing steps once we know what resources are needed Each workflow task – needs to be on a cluster, will need shared filesystem, good network connectivity

Practical resource constraints Where does and where can LEAD run today? (What are the real resource choices) WRF does not run on Itanium2 Have dependencies lot of libraries, binaries, compilers Today NCSA machines: copper, tungsten SDSC: tglogin IU teragrid PSC (luemix?) LEAD testbed (?) All GridFTP and GRAM enabled How about UC and UT machines? Possibilities in vgES UNC RTC supports GRAM but is Itanium2 Rice has another cluster. Can it be made Grid aware? Other sites? Credentials under which account – drlead account. vgES developers get access to drlead

We need to use LEAD workflow engine Scheduler makes assumption that queue prediction is accurate for first architecture interactions. What do we do when we are not accurate? vgES select and bind is tight Dan’s perl workflow doesn’t have notion of when to launch Having a deadline changes the way we look at this We need to submit job right at some time based on current wait queue times Prediction is used for which queue to submit, cannot use that to predict (selection happens apriori) Can’t guarantee when we can tell where to run resources What is the interface Scheduler is static right now. Is dynamic scalable?

vgLaunch might or not do WS-GRAM UNC/Lavanya will have to provide Suresh a standard interface to ask for resources and to submit jobs. Lavanya will turn around and do vgLaunch Data transfer is tightly coupled with computation. Can it be considered as a separate task? It is handled by myLEAD independently vgLaunch might or not do WS-GRAM How will we get the WS-GRAM notifications back through vgES? Application service is doing file staging (RFT didn’t seem to work right now) Notifications for files that are generated is based on job completion (this will need to change with file streaming) Data mining, visualization, etc Does vgMon help with some notification?

LEAD Data Management WE first registers the topic ID with myLEAD LEAD publishes events on the data status (task A has input data) myLEAD agent contacts DRS (which is managing the data) and tells it what to do with the data When a task gets a unique id (resolves to GridFTP) it resolves to the best possible input source

Everything has to happen through this. But this is changing Portal (user uses GUI to compose wf) App Service (AS) Everything has to happen through this. But this is changing Refer to Suresh’s diagram (in Lavanya’s Intro slides) LEAD Today The service (a wrapper) in LEAD need not be on the same cluster Right now LEAD submits sequential Is engine/application service smart enough to understand when resources are available How does LEAD deal with resources arrival times? The BPEL engine itself is just event based execution. Doesn’t have notion of time. The application service manages the state. How will an application service be called to launch into a queue before it runs (breaks the BPEL sequential call) LEAD does the PBS job submission What is the model for vgES launch?

After Lunch Demo Few slides from Suresh to understand data parts Service map of application service Workflow DAG What do you want to see the demo part? More about the interfaces