1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.

Slides:



Advertisements
Similar presentations
Andrew McNab - Manchester HEP - 2 May 2002 Testbed and Authorisation EU DataGrid Testbed 1 Job Lifecycle Software releases Authorisation at your site Grid/Web.
Advertisements

Author - Title- Date - n° 1 GDMP The European DataGrid Project Team
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
 Contributing >30% of throughput to ATLAS and CMS in Worldwide LHC Computing Grid  Reliant on production and advanced networking from ESNET, LHCNET and.
High Performance Computing Course Notes Grid Computing.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
EU-GRID Work Program Massimo Sgaravatto – INFN Padova Cristina Vistoli – INFN Cnaf as INFN members of the EU-GRID technical team.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
Robust Tools for Archiving and Preserving Digital Data Joseph JaJa, Mike Smorul, and Mike McGann Institute for Advanced Computer Studies Department of.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Workload Management Massimo Sgaravatto INFN Padova.
Magda – Manager for grid-based data Wensheng Deng Physics Applications Software group Brookhaven National Laboratory.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
XCAT Science Portal Status & Future Work July 15, 2002 Shava Smallen Extreme! Computing Laboratory Indiana University.
Experiment Requirements for Global Infostructure Irwin Gaines FNAL/DOE.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
DISTRIBUTED COMPUTING
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
Cluster Reliability Project ISIS Vanderbilt University.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
A Grid Computing Use case Datagrid Jean-Marc Pierson.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL U.S. ATLAS Physics and Computing Advisory Panel Review Argonne National Laboratory Oct 30, 2001.
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
GriPhyN EAC Meeting (Apr. 12, 2001)Paul Avery1 University of Florida Opening and Overview GriPhyN External.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
ATLAS Grid Data Processing: system evolution and scalability D Golubkov, B Kersevan, A Klimentov, A Minaenko, P Nevski, A Vaniachine and R Walker for the.
GridPP Presentation to AstroGrid 13 December 2001 Steve Lloyd Queen Mary University of London.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
Authors: Ronnie Julio Cole David
CLRC and the European DataGrid Middleware Information and Monitoring Services The current information service is built on the hierarchical database OpenLDAP.
…building the next IT revolution From Web to Grid…
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
WP3 Information and Monitoring Rob Byrom / WP3
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
U.S. ATLAS Computing Facilities Overview Bruce G. Gibbard Brookhaven National Laboratory U.S. LHC Software and Computing Review Brookhaven National Laboratory.
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
NORDUnet NORDUnet e-Infrastrucure: Grids and Hybrid Networks Lars Fischer CTO, NORDUnet Fall 2006 Internet2 Member Meeting, Chicago.
Grid Status - PPDG / Magda / pacman Torre Wenaus BNL DOE/NSF Review of US LHC Software and Computing Fermilab Nov 29, 2001.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
Experiment Support CERN IT Department CH-1211 Geneva 23 Switzerland t DBES The Common Solutions Strategy of the Experiment Support group.
CMS Experience with the Common Analysis Framework I. Fisk & M. Girone Experience in CMS with the Common Analysis Framework Ian Fisk & Maria Girone 1.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
10-Feb-00 CERN HepCCC Grid Initiative ATLAS meeting – 16 February 2000 Les Robertson CERN/IT.
Workload Management Workpackage
LCG middleware and LHC experiments ARDA project
Leigh Grundhoefer Indiana University
SDM workshop Strawman report History and Progress and Goal.
Wide Area Workload Management Work Package DATAGRID project
Gridifying the LHCb Monte Carlo production system
Status of Grids for HEP and HENP
gLite The EGEE Middleware Distribution
Presentation transcript:

1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory NOVEMBER 14-17, 2000

2 Motivation  The distributed LHC computing model developed by MONARC, and adopted by all 4 LHC experiments and the Hoffman review panel consists of distributed computational and data handling stores organized hierarchically (but flexibly) into “tiers” Tier 0 CERN Tier 1 A national facility Tier 2 A regional facility Tier 3 An individual institute Tier 4 An individual  The software which enables the experiments’ software frameworks to work effectively, and to provide efficient access to the and computation in a distributed environment is the called Grid Middleware  Middleware is broadly: Distributed Computing Management and Intersite Load Balancing Distributed Data Management Task and System Tracking, Workflow Management

3 Grid Middleware  Distributed computing Workflow scripts, schedulers: intersite load balancing Resource estimation, allocation Security, authentication, authorization, prioritization  Distributed data management replication transport mass store APIs for integration  Packages : CONDOR – distributed computing package Globus – wide security and authentication, file replication, information services  Projects: PPDG – adapt existing tools; first HEP Data Grid components GriPhyN – petascale virtual data; the Grid DataGrid – large scale EU initiative to develop and deploy the “Grid”

4 Grid Middleware functionality  Physicist submits a job  Middleware: Estimates resource requirements and performance Finds convenient places for it run Organizes efficient access to data  Caching, migration, replication Deals with authentication to the different sites Interface to local site resource allocation mechanisms, policies Run jobs Monitor progress Recover from problems Collects, manages output Archival bookkeeping (catalogs)

5 Software for Tier 2 Centers  Local site management Automated installation, configuration management, system maintenance Automated monitoring and error recovery Performance monitoring Expressing and managing local Grid resources  Mass storage management data storage uniform mass storage interface exchange of data and meta-data between mass storage systems

6 Frameworks, Interfaces and Adapters  Middleware is co-developed and written by “off project” computer scientists and some software engineers  Interfaces to Software Framework are specified by core software engineers and physicists in consultation with grid software integration engineers and physicists– all “on project”  Adapters, HEP or experiment specific, are written by grid software integration engineers and physicists  Requirement: Core software can function completely independent of middleware; likewise, an app can be built grid-enabled without without specialized knowledge.  WBS organization:  In CMS, grid software integration is initiated done by CAS engineers and by off project researchers (such as GriPhyN staff and postdocs), deployment and operation of final production codes in UF  In ATLAS, grid software integration engineers are accounted for under Facilities

7

8

9 EU DataGrid  Work areas: Workload management Grid data management Grid monitoring services Grid fabric management Grid mass storage management Grid integration testbed Two application areas (HEP & Bioinformatics)  Scale of effort: 3 year effort Many man-years  National initiatives to build Tier 1 & 2 facilities across the EU

10

11 GriPhyN and LHC Computing  Request planning and execution in a large scale production (ordered) and chaotic user analysis Large numbers of LHC physicists Wide area infrastructure  Execution management & fault tolerance  Performance analysis  Strategy: Physicists interact closely with CS and middleware developers Develop and deploy prototype  tier 2 centers and testbeds to provide platform for testing and performance assessment, comparison with MONARC simulations Integrate toolkits into Core software from the beginning  Grow the infrastructure adiabatically

12 GriPhyN Management External Advisory Panel VD Toolkit Development Requirements Definition & Scheduling Integration & Testing Documentation & Support Outreach/Education CS Research Execution Management Performance Analysis Request Planning & Scheduling Virtual Data Applications ATLAS CMS LSC/LIGO SDSS NSF Review Committee US LHC DCWG US LHC DCWG Industrial Programs Technical Coordination Committee Networks Databases Visualization Digital Libraries Grids Collaborative Systems Project Directors Internet 2DOE ScienceNSF PACIs Collaboration Board Other Grid Projects Project Coordination Group System Integration Project Coordinator

13 US LHC Distributed Computing Working Group US LHC DCWG DCWG GriPhyN PPDG EU DataGrid management lines worklines

14 Summary  The hierarchical LHC computing model is essential for physics and requires software which works in a distributed environment  Close interaction and collaboration are required Between physicists and computer scientists Between the two LHC experiments  Many common problems  GriPhyN collaboration is off to a good start and is an excellent opportunity for US ATLAS and US CMS to collaborate on common projects