DataGrid is a project funded by the European Commission under contract IST-2000-25182 F. Harris DataGrid EU Review 19 Feb 2004 WP8 HEP Applications Final.

Slides:



Advertisements
Similar presentations
An open source approach for grids Bob Jones CERN EU DataGrid Project Deputy Project Leader EU EGEE Designated Technical Director
Advertisements

1 User Analysis Workgroup Update  All four experiments gave input by mid December  ALICE by document and links  Very independent.
RLS Production Services Maria Girone PPARC-LCG, CERN LCG-POOL and IT-DB Physics Services 10 th GridPP Meeting, CERN, 3 rd June What is the RLS -
A conceptual model of grid resources and services Authors: Sergio Andreozzi Massimo Sgaravatto Cristina Vistoli Presenter: Sergio Andreozzi INFN-CNAF Bologna.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
LCG Milestones for Deployment, Fabric, & Grid Technology Ian Bird LCG Deployment Area Manager PEB 3-Dec-2002.
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
EU 2nd Year Review – Feb – Title – n° 1 WP8: Progress and testbed evaluation F Harris (Oxford/CERN) (WP8 coordinator )
UCL workshop – 4-5 March 2004 – HEP Assessment of EDG – n° 1 HEP Applications Evaluation of the EDG Testbed and Middleware Stephen Burke (EDG HEP Applications.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Andrew McNab - Manchester HEP - 5 July 2001 WP6/Testbed Status Status by partner –CNRS, Czech R., INFN, NIKHEF, NorduGrid, LIP, Russia, UK Security Integration.
DataGrid is a project funded by the European Commission under contract IST GridPP-2 Middleware 4 th -5 th Mar 2004 Information and Monitoring.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
DataGrid is a project funded by the European Commission under contract IST GridPP Edinburgh 4 Feb 2004 WP8 HEP Applications Final Project evaluation.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.
INFSO-RI Enabling Grids for E-sciencE SA1: Cookbook (DSA1.7) Ian Bird CERN 18 January 2006.
LCG and HEPiX Ian Bird LCG Project - CERN HEPiX - FNAL 25-Oct-2002.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Cosener’s House – 30 th Jan’031 LHCb Progress & Plans Nick Brook University of Bristol News & User Plans Technical Progress Review of deliverables.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
WP8 Status – Stephen Burke – 30th January 2003 WP8 Status Stephen Burke (RAL) (with thanks to Frank Harris)
DOSAR Workshop, Sao Paulo, Brazil, September 16-17, 2005 LCG Tier 2 and DOSAR Pat Skubic OU.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
5 Sep 2002F Harris Plenary Budapest1 WP8 Report F Harris (Oxford/CERN)
Status of the LHCb MC production system Andrei Tsaregorodtsev, CPPM, Marseille DataGRID France workshop, Marseille, 24 September 2002.
Responsibilities of ROC and CIC in EGEE infrastructure A.Kryukov, SINP MSU, CIC Manager Yu.Lazin, IHEP, ROC Manager
13 May 2004EB/TB Middleware meeting Use of R-GMA in BOSS for CMS Peter Hobson & Henry Nebrensky Brunel University, UK Some slides stolen from various talks.
LCG EGEE is a project funded by the European Union under contract IST LCG PEB, 7 th June 2004 Prototype Middleware Status Update Frédéric Hemmer.
INFSO-RI Enabling Grids for E-sciencE OSG-LCG Interoperability Activity Author: Laurence Field (CERN)
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
CERN-IT Oracle Database Physics Services Maria Girone, IT-DB 13 December 2004.
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
LCG ARDA status Massimo Lamanna 1 ARDA in a nutshell ARDA is an LCG project whose main activity is to enable LHC analysis on the grid ARDA is coherently.
INFSO-RI Enabling Grids for E-sciencE Experience of using gLite for analysis of ATLAS combined test beam data A. Zalite / PNPI.
Presenter Name Facility Name UK Testbed Status and EDG Testbed Two. Steve Traylen GridPP 7, Oxford.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Jens G Jensen RAL, EDG WP5 Storage Element Overview DataGrid Project Conference Heidelberg, 26 Sep-01 Oct 2003.
2-Sep-02Steve Traylen, RAL WP6 Test Bed Report1 RAL and UK WP6 Test Bed Report Steve Traylen, WP6
DataTAG is a project funded by the European Union DataTAG WP4 meeting, Bologna 29/07/2003 – n o 1 GLUE Schema - Status Report DataTAG WP4 meeting Bologna,
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Julia Andreeva on behalf of the MND section MND review.
2 Sep 2002F Harris EDG/WP6 meeeting at Budapest LHC experiments use of EDG Testbed F Harris (Oxford/CERN)
David Foster LCG Project 12-March-02 Fabric Automation The Challenge of LHC Scale Fabrics LHC Computing Grid Workshop David Foster 12 th March 2002.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
DIRAC Project A.Tsaregorodtsev (CPPM) on behalf of the LHCb DIRAC team A Community Grid Solution The DIRAC (Distributed Infrastructure with Remote Agent.
DataGrid is a project funded by the European Commission under contract IST rd EU Review – 19-20/02/2004 The EU DataGrid Project Three years.
EGEE is a project funded by the European Union under contract IST Information and Monitoring Services within a Grid R-GMA (Relational Grid.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Gennaro Tortone, Sergio Fantinel – Bologna, LCG-EDT Monitoring Service DataTAG WP4 Monitoring Group DataTAG WP4 meeting Bologna –
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
Towards deploying a production interoperable Grid Infrastructure in the U.S. Vicky White U.S. Representative to GDB.
CERN Certification & Testing LCG Certification & Testing Team (C&T Team) Marco Serra - CERN / INFN Zdenek Sekera - CERN.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE and gLite are registered trademarks EGEE Operations: Evolution of the Role of.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
SAM architecture EGEE 07 Service Availability Monitor for the LHC experiments Simone Campana, Alessandro Di Girolamo, Nicolò Magini, Patricia Mendez Lorenzo,
GDB Meeting CERN 09/11/05 EGEE is a project funded by the European Union under contract IST A new LCG VO for GEANT4 Patricia Méndez Lorenzo.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Bob Jones EGEE Technical Director
Regional Operations Centres Core infrastructure Centres
Eleonora Luppi INFN and University of Ferrara - Italy
EGEE Middleware Activities Overview
EO Applications Parallel Session
Bernd Panzer-Steindel CERN/IT
LHC Data Analysis using a worldwide computing grid
Report on GLUE activities 5th EU-DataGRID Conference
The LHCb Computing Data Challenge DC06
Presentation transcript:

DataGrid is a project funded by the European Commission under contract IST F. Harris DataGrid EU Review 19 Feb 2004 WP8 HEP Applications Final Project evaluation of EDG middleware, and summary of workpackage achievements F HARRIS (OXFORD/CERN)

F Harris DataGrid EU Review 19 Feb n° 2 Outline u Overview of objectives and achievements u Key points in the achievements of the 6 WP8 experiments u Lessons learned from the three years u Summary of the exploitation of WP8 work, and of future HEP applications activity in EGEE/LCG u Concluding comments u Questions and discussion

F Harris DataGrid EU Review 19 Feb n° 3 Overview of objectives and achievements OBJECTIVES ACHIEVEMENTS Continued work in Architectural Task Force (ATF) u Walkthroughs of HEP use cases helped to clarify interfacing problems. Reactivation of the Application Working Group (AWG) u Extension of HEPCAL use cases covering key areas in Biomedicine and Earth Sciences. u Basis of first proposal for common application work in EGEE Work with LCG/GAG (Grid Applications group) in further refinement of HEP requirements u HEPCAL-2 requirements document for the use of grid by thousands of individual users. u In addition further refined the original HEPCAL document Developments of tutorials and documentation for the user community u WP8 has played a substantial role in course design, implementation and delivery

F Harris DataGrid EU Review 19 Feb n° 4 OBJECTIVESACHIEVEMENTS Evaluate EDG Application Testbed, and integrate into experiment tests as appropriate. u Further successful evaluation of 1.4.n throughout the summer. u Evaluation of EDG 2.0 on the EDG Application Testbed since October, and of EDG 2.1 since December Liase with LCG regarding EDG/LCG integration and the development of the LCG service. u EIPs (Loose Cannons) helped testing of EDG components on the LCG Cert TB prior to LCG-1 start in September. u Performed stress tests on LCG-1. Continue work with experiments on data challenges throughout the year. u All 6 experiments have conducted data challenges of different scales throughout 2003 on EDG App TB or LCG/Grid.it.

F Harris DataGrid EU Review 19 Feb n° 5 Comments on experiment work Expt EDG LCG/Grid.it US Grid 3 NorduGrid Experiments are living in an international multi-grid world using other Grids DataTag project is very important for inter-operability (GLUE schema used for inter-operability with US grids) Have used EDG software in a number of grids EDG Application Testbed LCG Service (LCG-1 evolving to LCG-2) Italian Grid.it (identical with LCG-1 release) Having 2 running experiments (in addition to the 4 LHC experiments) involved in the evaluations has proved very useful BaBar and work on Grid.it D0 and work on EDG App TB

F Harris DataGrid EU Review 19 Feb n° 6 Evolution in the use of EDG App TB and the LCG service (and Grid.it) Feb-Sept 2003 EDG 1.4 evolved with production use by Atlas, CMS and LHCb with efficiencies ranging from 40 – 90% - (higher in self-managed configurations) LHCb 300K events Feb/Mar Atlas 500K events May CMS 2M events in LCG-0 configuration in summer Sep LCG-1 service open, (and Grid.it installed LCG-1 release) Used EDG 2.0 job and data management and partitioned MDS (+VDT) All experiments have installed software in LCG-1 and accomplished positive tests ATLAS,ALICE,BaBar,CMS performed ‘production’ runs on LCG-1/Grid.it Oct 20 – HEP formal evaluations of EDG 2.0 commenced R-GMA instead of ‘partitioned’ MDS D0 and CMS evaluated ‘monitoring’ features of RGMA D0 did ‘production’ running Regular generic tests Oct December 2003 Feb 2004 LCG-2 service release Move to new C++ compiler Enhanced data management Will include R-GMA for job and network monitoring All experiments will use it for data challenges Move to EDG 2.1 on EDG App TB Fixing known problems Move to new C++ compiler Enhance Data Management VOMS Secure services

F Harris DataGrid EU Review 19 Feb n° 7 The next slides show recent achievements of the 6 WP8 experiments

F Harris DataGrid EU Review 19 Feb n° 8 u evaluation on LCG-1 and Grid.it Sep-Nov 2003 n Significant improvement in terms of stability with respect to tests in Spring 2003 n Jobs were sensitive to space on worker nodes u Projected load on LCG2 during ALICE DC(start Feb 2004) n 10 5 events (1 event/job) n Generate ~30 TB output n Test LCG Mass Storage n Parallel data analysis (AliEN/PROOF) including LCG Efficiency was generally a step function for batches (close to 0 or close to 100). With long jobs and multi files very sensitive to long-term system stability

F Harris DataGrid EU Review 19 Feb n° 9 Main features of new DC2 system for multi-grid environment Common production database, supervisor and data management system for all of ATLAS Executors developed by middleware experts (LCG, NorduGrid, US). ATLAS n Use of EDG (mod for RH7.3) in May 2003 s Reconstructed 500 K events in 250 jobs with 85% 1 st pass efficiency s With privately managed configuration of 7 sites in Italy, Lyon and Cambridge n LCG-1(+ Grid.it) production in Jan-Feb 2004 s Have simulated events in hr. jobs of 200 events each with efficiency ~80% n LCG-2 plans Start around April

F Harris DataGrid EU Review 19 Feb n° 10 u LCG-0 (summer 2003) n Components from VDT and EDG n DataTAG (GLUE) n VOMS + RLS + R-GMA n 14 sites configured and managed by CMS n Substantial improvements in efficiency compared to first EDG stress test (~80%) n CPU hours on LCG-0 s 500K Pythia 2000 jobs 8h s 1.5M CMSIM 6000 jobs 10h u LCG-1 n Ran for 9 days on LCG-1 over Xmas n In total 600,000 events (30-40h jobs) were produced n Sites used mainly in Italy, Spain n Efficiency around 75% over XMAS n Used GENIUS portal u LCG-2 -data challenge Mar 1

F Harris DataGrid EU Review 19 Feb n° 11 u Tests on the EDG1.4 application testbed (Feb-Mar 2003): n Standard LHCb production tasks, 300K events produced; n ~35% success rate. (TB support running down) n Software installation by the running job; u EDG2.0 tests (November 2003): Submission of the jobs: n To EDG RB; n and experimented with directly to a CE with the CE status information obtained from the CE GRIS server: 90% efficiency u GETTING READY NOW FOR LCG- 2 and DC in April ( tests are positive) DIRAC Job Management DIRAC Job Management DIRAC CE Resource Broker Resource Broker CE 1 Agent CE 2 CE 3 Production manager Production manager GANGA UI User CLI LCG DIRAC Sites

F Harris DataGrid EU Review 19 Feb n° 12 Babar u Strategy for first integration n Created ‘simulation’ RPM to be installed at sites n Data output stored on closest SE n Data copied to Tier-1 or SLAC using edg-copy u Scheme first tested with EDG on 5 Italian sites u Operation on Grid.it with LCG-1 release n RB at CNAF - farms at 8 sites n 1 week test with ~ 500 jobs n 95% success at Ferrara(site with central DB) n 60% success elsewhere n 33% failures due to network saturation due to simultaneous requests to remote applications database n Positive experience with use of GENIUS portal s u Analysis applications also have been successfully tested on EDG App TB

F Harris DataGrid EU Review 19 Feb n° 13 u Interfaced EDG software and resources to D0 re-processing n Frequent software updates so don’t use RPMs s Registered compressed tar archives in RLS as grid files for installation by jobs n Use RGMA for monitoring s Allows users and programs to publish information for inspection by other users, and for archiving in production database u Found EDG s/w generally satisfactory for task (with caveats)  Used ‘Classic’ SE s/w while waiting for developments to interface to SARA mass-store  Very sensitive to RGMA stability. Since December good progress with RGMA, and can run at ~90% efficiency when RGMA is up

F Harris DataGrid EU Review 19 Feb n° 14 Summary of middleware evaluations u Workload management n Tests have shown that software is more robust and scalable s Stress tests were successful with up to 1600 jobs in multiple streams – efficiencies over 90% s Problems with new sites during tests– VOs not set up properly (though site accepted job) u Data Management n Has worked well with respect to functionality and scalability (have registered ~100K files in ongoing tests) s Tests so far with only 1 LRC per VO implemented n Performance needs enhancement s Registrations and simple queries can take up to 10 seconds n We have lost (with GDMP) bulk transfer functions n Some functions needed inbound IP connectivity (Globus). D0 had to program round this (problem since fixed)

F Harris DataGrid EU Review 19 Feb n° 15 Summary of middleware evaluations(2) u Information System n Partitioned MDS has worked well for LCG following on from work accomplished within EDG (BD II work), but limited to ~100 sites probably. n R-GMA work is very promising for ‘life after MDS’, but needs ‘hardening’. u Mass Storage support (mission critical for data challenges) n We await ‘accepted’ uniform interface to disk and tape systems s Solution coming with SRM/GFAL software s WP5 have made important contribution to the development of SRM interface n EDG 2.0 had mass storage access to CERN (Castor) and RAL(ADS) s The ‘Classic-SE’ has been a useful fallback (gridftp server) while waiting for commissioning of developments

F Harris DataGrid EU Review 19 Feb n° 16 Site Related Issues (major factors in overall efficiency) u Site Certification n Official, standard procedure as part of release n Consistency checks of published information u Site Configuration n Large parameter space with insufficient defaults so please can we have… s Automated configuration s Automated tests s Run-time checks of parameters u Space management and publishing n Running out of space on SEs and WNs is still a problem. Jobs need to check availability before running

F Harris DataGrid EU Review 19 Feb n° 17 The Deliverables + ‘extra’ outputs from WP8 u The formal EU deliverables n D8.1 The original HEP requirements document n D8.2 ‘Evaluation by experiments after 1 st year’ n D8.3 ‘Evaluation by experiments after 2 nd year’ n D8.4 ‘Evaluation after 3 rd year’ u Extra key documents (being used as input to EGEE) n HEPCAL Use cases May 2002 (revised Oct 2003) n AWG Recommendations for middleware (June 2003) n AWG Enhanced use cases (for Biomed,Earth Science) Sep 2003 n HEPCAL2 Use cases for analysis (several WP8 people) u Generic HEP test suite used by EDG/LCG u Ongoing consultancy from ‘loose cannons’ to all applications u Interfacing of 6 experiment systems to middleware

F Harris DataGrid EU Review 19 Feb n° 18 Main lessons learned Architecture & Software Life-cycle n Information system is nerve centre of grid. We look to R-GMA developments for long term solution to scaling problems n Globally HEP applications feel it would have been ‘better’ to start with simpler prototype, and to have more frequent incremental releases n Applications should have played larger role in architecture in defining interfaces (so we could all learn together!). Deployment & Operations of the Middleware n Formation of Task Forces (applications+middleware) was a very important step midway in project n Loose Cannons (team of 5) were crucial to all developments. Worked across experiments. This team comprised all the funded effort of WP8.

F Harris DataGrid EU Review 19 Feb n° 19 Main lessons learned (cont’d) Site Related Lessons n Site configuration must be automated. n Site certification needs to be improved. Incompliant sites make correct brokering impossible. n Space management on SEs and WNs is an outstanding problem n We look to SRM/GFAL as a solution to uniform mass storage interfacing n Must have flexible application s/w installation. Application needs and site policies vary.

F Harris DataGrid EU Review 19 Feb n° 20 Exploitation of the work of WP8, and future HEP applications work in LCG/EGEE u All experiments have exploited the EDG middleware using WP8 effort, and this exploitation is being carried into the data challenges in 2004 u The HEPCAL and AWG documents are essential inputs to the future EGEE/LCG work u Future developments will be in the context of EGEE/LCG infrastructure carrying over the important experience from WP8 u The NA4 activity in EGEE will include dedicated people for interfacing middleware to experiments software (8 people at CERN + others distributed in the community) u Within the EGEE project middleware will be ‘hardened’ (including EDG components) and evaluated by the HEP applications, in parallel with the use of current EDG software on LCG for the physics data challenges

F Harris DataGrid EU Review 19 Feb n° 21 Concluding comments u Over the past 3 years the HEP community has moved to the use of grid services in physics production systems using world- wide configurations u Experiments are using several managed grids (EGEE/LCG,US Grids, Nordugrid) so inter-operability is crucial u We have learned very important lessons in Datagrid which we carry forward into the EGEE project, and we will learn more lessons from the use of EDG/EGEE software in the forthcoming experiment data challenges in 2004