10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.

Slides:



Advertisements
Similar presentations
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Advertisements

The DOE Science Grid Computing and Data Infrastructure for Large-Scale Science William Johnston, Lawrence Berkeley National Lab Ray Bair, Pacific Northwest.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
The Globus Toolkit Gary Jackson. Introduction The Globus Toolkit is a product of the Globus Alliance ( It is middleware for developing.
Slides for Grid Computing: Techniques and Applications by Barry Wilkinson, Chapman & Hall/CRC press, © Chapter 1, pp For educational use only.
The LHC Computing Grid Project Tomi Kauppi Timo Larjo.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Data Grids: Globus vs SRB. Maturity SRB  Older code base  Widely accepted across multiple communities  Core components are tightly integrated Globus.
Workload Management Massimo Sgaravatto INFN Padova.
Status of Globus activities within INFN (update) Massimo Sgaravatto INFN Padova for the INFN Globus group
Globus Computing Infrustructure Software Globus Toolkit 11-2.
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Globus 4 Guy Warner NeSC Training.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Assessment of Core Services provided to USLHC by OSG.
QCDgrid Technology James Perry, George Beckett, Lorna Smith EPCC, The University Of Edinburgh.
1 TclGlobus ITR2003 Meeting Argonne National Laboratory May 10 th, 2004 Kent Blackburn LIGO-G E.
April 19, 2005Scott Koranda1 The LIGO Scientific Collaboration Data Grid Scott Koranda UW-Milwaukee On behalf of the LIGO Scientific Collaboration.
Open Science Grid Software Stack, Virtual Data Toolkit and Interoperability Activities D. Olson, LBNL for the OSG International.
Don Quijote Data Management for the ATLAS Automatic Production System Miguel Branco – CERN ATC
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
Patrick R Brady University of Wisconsin-Milwaukee
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
CoG Kit Overview Gregor von Laszewski Keith Jackson.
G Z LIGO Scientific Collaboration Grid Patrick Brady University of Wisconsin-Milwaukee LIGO Scientific Collaboration.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
INFSO-RI Enabling Grids for E-sciencE The US Federation Miron Livny Computer Sciences Department University of Wisconsin – Madison.
Miguel Branco CERN/University of Southampton Enabling provenance on large-scale e-Science applications.
VOX Project Status T. Levshina. Talk Overview VOX Status –Registration –Globus callouts/Plug-ins –LRAS –SAZ Collaboration with VOMS EDG team Preparation.
Through the development of advanced middleware, Grid computing has evolved to a mature technology in which scientists and researchers can leverage to gain.
QCDGrid Progress James Perry, Andrew Jackson, Stephen Booth, Lorna Smith EPCC, The University Of Edinburgh.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
GriPhyN Status and Project Plan Mike Wilde Mathematics and Computer Science Division Argonne National Laboratory.
Miron Livny Computer Sciences Department University of Wisconsin-Madison Welcome and Condor Project Overview.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
Grid Workload Management Massimo Sgaravatto INFN Padova.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
Quick Introduction to NorduGrid Oxana Smirnova 4 th Nordic LHC Workshop November 23, 2001, Stockholm.
Tools for collaboration How to share your duck tales…
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
The Open Science Grid OSG Ruth Pordes Fermilab. 2 What is OSG? A Consortium of people working together to Interface Farms and Storage to a Grid and Researchers.
Part 9: MyProxy Pragmatics This presentation and lab ends the GRIDS Center agenda Q: When do we convene again tomorrow?
The GriPhyN Planning Process All-Hands Meeting ISI 15 October 2001.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.
LIGO-G E LIGO Scientific Collaboration Data Grid Status Albert Lazzarini Caltech LIGO Laboratory Trillium Steering Committee Meeting 20 May 2004.
G Z LIGO's Physics at the Information Frontier Grant and OSG: Update Warren Anderson for Patrick Brady (PIF PI) OSG Executive Board Meeting Caltech.
Alain Roy Computer Sciences Department University of Wisconsin-Madison Condor & Middleware: NMI & VDT.
Easy Access to Grid infrastructures Dr. Harald Kornmayer (NEC Laboratories Europe) Dr. Mathias Stuempert (KIT-SCC, Karlsruhe) EGEE User Forum 2008 Clermont-Ferrand,
Contact: Junwei Cao SC2005, Seattle, WA, November 12-18, 2005 The authors gratefully acknowledge the support of the United States National.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Development of e-Science Application Portal on GAP WeiLong Ueng Academia Sinica Grid Computing
Scott Koranda, UWM & NCSA 14 January 2016www.griphyn.org Lightweight Data Replicator Scott Koranda University of Wisconsin-Milwaukee & National Center.
2005 GRIDS Community Workshop1 Learning From Cyberinfrastructure Initiatives Grid Research Integration Development & Support
GriPhyN Project Paul Avery, University of Florida, Ian Foster, University of Chicago NSF Grant ITR Research Objectives Significant Results Approach.
USGS GRID Exploratory Status Review Stuart Doescher Mike Neiers USGS/EDC May
LIGO-G W Use of Condor by the LIGO Scientific Collaboration Gregory Mendell, LIGO Hanford Observatory On behalf of the LIGO Scientific Collaboration.
Status of Globus activities Massimo Sgaravatto INFN Padova for the INFN Globus group
LIGO-G Z1 Using Condor for Large Scale Data Analysis within the LIGO Scientific Collaboration Duncan Brown California Institute of Technology.
Grid Workload Management (WP 1) Massimo Sgaravatto INFN Padova.
1 An update on the Open Science Grid for IHEPCCC Ruth Pordes, Fermilab.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 22 February 2006.
ACGT Architecture and Grid Infrastructure Juliusz Pukacki ‏ EGEE Conference Budapest, 4 October 2007.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
VO Experiences with Open Science Grid Storage OSG Storage Forum | Wednesday September 22, 2010 (10:30am)
What is OSG? (What does it have to do with Atlas T3s?) What is OSG? (What does it have to do with Atlas T3s?) Dan Fraser OSG Production Coordinator OSG.
gLite The EGEE Middleware Distribution
The DZero/PPDG D0/PPDG mission is to enable fully distributed computing for the experiment, by enhancing SAM as the distributed data handling system of.
Presentation transcript:

10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee

10/20/05 LIGO Scientific Collaboration 2 Grid Middleware & the Gap LIGO Data Grid leverages the best of grid middleware Globus Toolkit » GSI and X.509 certificates for authentication/authorization » GridFTP for moving data and files » Replica Location Service (RLS) for data discovery Condor » Condor for high throughput computing on clusters » Condor DAGman for workflow execution Virtual Data Toolkit (VDT) » pyGlobus » Pegasus » many, many more

10/20/05 LIGO Scientific Collaboration 3 Grid Middleware & the Gap Using these middleware building blocks LIGO has... developed tools for data discovery developed tools for managing complex workflows...but the building blocks are middle ware Alone the middleware does not put tools into the hands of LIGO scientists An important GAP is left between the middleware and what LIGO scientists need to work efficiently

10/20/05 LIGO Scientific Collaboration 4 Grid Middleware & the Gap Different science projects (LIGO, SDSS, CMS, ATLAS) can leverage much of the same grid middleware......yet they are distinct experiments and so there will always be gaps between the common middleware and tools LIGO scientists can use Data discovery a good example: middleware tool is Globus Replica Location Service (RLS) GT provides the “simple” command line tool globus-rls-cli

10/20/05 LIGO Scientific Collaboration 5 Grid Middleware & the Gap

10/20/05 LIGO Scientific Collaboration 6 Filling the Gap Filling this gap requires dedicated LIGO effort work with LIGO scientists to understand use case » how can I find data from a particular site for a particular interval of time in a particular format? understand the underlying middleware (RLS) » how does the API work, can we use LSC favorite glue languages, what is necessary for deployment on LSC computing resources? write the code... test the code... package the code... document the tool... deploy the tool... support the users!

10/20/05 LIGO Scientific Collaboration 7 Filling the Gap Bridging these gaps is not “set and forget” Tools must continually evolve as LIGO data analysis matures LIGO computing professionals must constantly incorporate scientist feedback, evolve, and release again

10/20/05 LIGO Scientific Collaboration 8 Filling the Gap Data discovery in LIGO using LSCdataFind

10/20/05 LIGO Scientific Collaboration 9 Filling the Gap Tools for constructing LIGO workflows (built on top of Condor DAGman):

10/20/05 LIGO Scientific Collaboration 10 Bigger Gaps in Middleware For some important LIGO tasks the grid middleware stack comes up short: TclGlobus » many useful LIGO tools pre-date LDG written using Tcl/Tk » fully incorporating these tools into LDG requires a TclGlobus LIGO Metadata Service » metadata for different science experiments is mostly unique » little investigation so far in larger grid community » LIGO has working prototype that (just barely) scales for S5 Prototyping, developing, and deploying grid middleware from within LIGO is sometimes necessary...but comes with a cost

10/20/05 LIGO Scientific Collaboration 11 Maturing the LIGO Data Grid Metadata service » LIGO specific (for at least 3 years) requires prototyping, developing, deploying, support Resource discovery service » requires wrapping tools like Globus MDS or Condor ClassAds and developing LIGO specific schema Monitoring services » requires investigation and sampling of a large number of possible solutions, then customization, deployment, and documentation Advanced workflow planning » requires integration of Globus Pegasus and VDS into LIGO GLUE toolset Brokering services » still active research area...requires LIGO use case investigation and feedback into grid middleware community

10/20/05 LIGO Scientific Collaboration 12 Maturing the LIGO Data Grid Authentication & Authorization (LIGO CA) » deployment underway, requires documentation and solid long term user support VO management » initial LIGO deployment ongoing, requires continued investigation of the VO privilege project led by US CMS & ATLAS LDAS » requires tight integration into LDG authorization model LDR » requires metadata scaling solutions, advanced replica services to be developed and deployed each of these requires filling gaps between the common middleware and LIGO scientists with wrapping, deploying, documentation, support...

10/20/05 LIGO Scientific Collaboration 13 Maintaining the LDG Fabric None of the LIGO Data Grid services will run without hardware & networking......ongoing administration of all LDG computing resources is critical......as is administration of LDG services at each LDG site and dedication to local support

10/20/05 LIGO Scientific Collaboration 14 Evolving the LIGO Data Grid LIGO Data Grid exists now and is enabling LIGO science......but realizing the full potential of the LIGO Data Grid and exposing the full science potential of LIGO requires a constant investment in LIGO computing professionals around the LIGO Data Grid to develop, deploy, document, and support the services and tools LIGO scientists need.