Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE.

Slides:



Advertisements
Similar presentations
EGEE-II INFSO-RI Enabling Grids for E-sciencE The gLite middleware distribution OSG Consortium Meeting Seattle,
Advertisements

Plateforme de Calcul pour les Sciences du Vivant SRB & gLite V. Breton.
Analysis demos from the experiments. Analysis demo session Introduction –General information and overview CMS demo (CRAB) –Georgia Karapostoli (Athens.
LHC Experiment Dashboard Main areas covered by the Experiment Dashboard: Data processing monitoring (job monitoring) Data transfer monitoring Site/service.
Stefano Belforte INFN Trieste 1 CMS SC4 etc. July 5, 2006 CMS Service Challenge 4 and beyond.
Ganga Developments Karl Harrison (University of Cambridge) 18th GridPP Meeting University of Glasgow, 20th-21st March 2007
GRACE Project IST EGAAP meeting – Den Haag, 25/11/2004 Giuseppe Sisto – Telecom Italia Lab.
US ATLAS Western Tier 2 Status and Plan Wei Yang ATLAS Physics Analysis Retreat SLAC March 5, 2007.
Computing Infrastructure Status. LHCb Computing Status LHCb LHCC mini-review, February The LHCb Computing Model: a reminder m Simulation is using.
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
SRM 2.2: status of the implementations and GSSD 6 th March 2007 Flavia Donno, Maarten Litmaath INFN and IT/GD, CERN.
CERN IT Department CH-1211 Genève 23 Switzerland t Internet Services Job Monitoring for the LHC experiments Irina Sidorova (CERN, JINR) on.
OSG Area Coordinator’s Report: Workload Management April 20 th, 2011 Maxim Potekhin BNL
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
Alex Read, Dept. of Physics Grid Activity in Oslo CERN-satsingen/miljøet møter MN-fakultetet Oslo, 8 juni 2009 Alex Read.
The ILC And the Grid Andreas Gellrich DESY LCWS2007 DESY, Hamburg, Germany
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
CCRC’08 Weekly Update Jamie Shiers ~~~ LCG MB, 1 st April 2008.
MW Readiness Verification Status Andrea Manzi IT/SDC 21/01/ /01/15 2.
LCG EGEE is a project funded by the European Union under contract IST LCG PEB, 7 th June 2004 Prototype Middleware Status Update Frédéric Hemmer.
Owen SyngeTitle of TalkSlide 1 Storage Management Owen Synge – Developer, Packager, and first line support to System Administrators. Talks Scope –GridPP.
Stefano Belforte INFN Trieste 1 Middleware February 14, 2007 Resource Broker, gLite etc. CMS vs. middleware.
1 LHCb on the Grid Raja Nandakumar (with contributions from Greig Cowan) ‏ GridPP21 3 rd September 2008.
1 User Analysis Workgroup Discussion  Understand and document analysis models  Best in a way that allows to compare them easily.
INFSO-RI Enabling Grids for E-sciencE Enabling Grids for E-sciencE Pre-GDB Storage Classes summary of discussions Flavia Donno Pre-GDB.
INFSO-RI Enabling Grids for E-sciencE Experience of using gLite for analysis of ATLAS combined test beam data A. Zalite / PNPI.
Karsten Köneke October 22 nd 2007 Ganga User Experience 1/9 Outline: Introduction What are we trying to do? Problems What are the problems? Conclusions.
ATLAS Distributed Analysis Dietrich Liko. Thanks to … pathena/PANDA: T. Maneo, T. Wenaus, K. De DQ2 end user tools: T. Maneo GANGA Core: U. Edege, J.
A PanDA Backend for the Ganga Analysis Interface J. Elmsheuser 1, D. Liko 2, T. Maeno 3, P. Nilsson 4, D.C. Vanderster 5, T. Wenaus 3, R. Walker 1 1: Ludwig-Maximilians-Universität.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
EGEE-III INFSO-RI Enabling Grids for E-sciencE Ricardo Rocha CERN (IT/GS) EGEE’08, September 2008, Istanbul, TURKEY Experiment.
INFSO-RI Enabling Grids for E-sciencE The gLite File Transfer Service: Middleware Lessons Learned form Service Challenges Paolo.
Alex Read, Dept. of Physics Grid Activities in Norway R-ECFA, Oslo, 15 May, 2009.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Plans for Service Challenge 3 Ian Bird LHCC Referees Meeting 27 th June 2005.
Doug Benjamin Duke University. 2 ESD/AOD, D 1 PD, D 2 PD - POOL based D 3 PD - flat ntuple Contents defined by physics group(s) - made in official production.
ANALYSIS TOOLS FOR THE LHC EXPERIMENTS Dietrich Liko / CERN IT.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI Monitoring of the LHC Computing Activities Key Results from the Services.
David Adams ATLAS ATLAS-ARDA strategy and priorities David Adams BNL October 21, 2004 ARDA Workshop.
MND review. Main directions of work  Development and support of the Experiment Dashboard Applications - Data management monitoring - Job processing monitoring.
LCG Service Challenges SC2 Goals Jamie Shiers, CERN-IT-GD 24 February 2005.
U.S. ATLAS Facility Planning U.S. ATLAS Tier-2 & Tier-3 Meeting at SLAC 30 November 2007.
The ATLAS Strategy for Distributed Analysis on several Grid Infrastructures D. Liko, IT/PSS for the ATLAS Distributed Analysis Community.
T3g software services Outline of the T3g Components R. Yoshida (ANL)
EGEE-II INFSO-RI Enabling Grids for E-sciencE Practical using WMProxy advanced job submission.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
ATLAS Distributed Analysis Dietrich Liko IT/GD. Overview  Some problems trying to analyze Rome data on the grid Basics Metadata Data  Activities AMI.
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
Distributed Analysis Tutorial Dietrich Liko. Overview  Three grid flavors in ATLAS EGEE OSG Nordugrid  Distributed Analysis Activities GANGA/LCG PANDA/OSG.
INFSO-RI Enabling Grids for E-sciencE gLite Test and Certification Effort Nick Thackray CERN.
Enabling Grids for E-sciencE Experience Supporting the Integration of LHC Experiments Computing Systems with the LCG Middleware Simone.
Grid Deployment Board 5 December 2007 GSSD Status Report Flavia Donno CERN/IT-GD.
The Grid Storage System Deployment Working Group 6 th February 2007 Flavia Donno IT/GD, CERN.
WLCG Status Report Ian Bird Austrian Tier 2 Workshop 22 nd June, 2010.
Status of gLite-3.0 deployment and uptake Ian Bird CERN IT LCG-LHCC Referees Meeting 29 th January 2007.
INFSO-RI Enabling Grids for E-sciencE File Transfer Software and Service SC3 Gavin McCance – JRA1 Data Management Cluster Service.
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
ATLAS Computing Model Ghita Rahal CC-IN2P3 Tutorial Atlas CC, Lyon
J. Shank DOSAR Workshop LSU 2 April 2009 DOSAR Workshop VII 2 April ATLAS Grid Activities Preparing for Data Analysis Jim Shank.
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
(Prague, March 2009) Andrey Y Shevel
Data Challenge with the Grid in ATLAS
New monitoring applications in the dashboard
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Readiness of ATLAS Computing - A personal view
LCG middleware and LHC experiments ARDA project
R. Graciani for LHCb Mumbay, Feb 2006
LHC Data Analysis using a worldwide computing grid
Presentation transcript:

Acronyms GAS - Grid Acronym Soup, LCG - LHC Computing Project EGEE - Enabling Grids for E-sciencE gLite - Lightweight Middleware for Grid Computing (EGEE) OSG - Open Science Grid PANDA - Production and Distributed Analysis (USATLAS & OSG) Nordugrid - Nordic Grid ARC - Advanced Resource Connector (Nordugrid) WMS - Workload Management System GANGA - Gaudi/Athena and Grid Alliance pathena – parallel Athena DDM - Distributed Data Management DPM – Disk Pool Manager dCache – disk pool management system CASTOR - CERN Advanced STORage manager SRM - Storage Resource Manager FTS - File Transfer System AOD - Analysis Object Data TAG - Event Selection Data

2007 is a busy year … Many Distributed Analysis Tutorials – Edinburgh Milano – Lyon – Munich – Toronto – Bergen – Valencia – CERN – Stanford - Sharply rising number of grid users Increasing demand for user support

Overview The Grid Infrastructure – EGEE/gLite, OSG/PANDA, Nordugrid/ARC – Data Distribution – Data Access The Distributed Analysis Activity – pathena/PANDA – GANGA/EGEE & Nordugrid – GANGA/PANDA – Dashboard Conclusions

Three Grids EGEE – gLite Middleware OSG in the US – PANDA Nordugrid in Scandinavia – ARC Middleware

Data Distribution More manpower for the developer team New Data Placement Coordinator New DDM version 0.3 Disk space has been reviewed Eagerly awaiting – SRM 2.2 – New version of FTS Aim: 100 % of data on all T1’s and then further copies on T2’s Today: On 5 T1’s we have 80 to 92 % Problems are related to disk space and to changes to new storage systems

Data Access We need direct access to the data – TAG analysis – Faster IO using Athena v13 (2 to 10 MB/sec) Together with the LCG Application Area – Solve issues with the access to various Storage Systems (DPM, dCache, Castor,…) We are also investigating xrootd technology

pathena/PANDA Lightweight submission client to PANDA Web pages for monitoring pathena AnalysisSkeleton_jobOptions.py --inDS trig1_misal1_mc T1_McAtNlo_Jimmy.recon.AOD.v outDS user.DietrichLiko.test.01

PANDA since Feb 2006 Close to 200 users since February 2006, about 38 users last month Large number of jobs BNL is main site

GANGA/EGEE + Nordugrid GANGA job submission tool – ATLAS and LHCb GUI ganga athena AnalysisSkeleton_jobOptions.py --inDS trig1_misal1_mc T1_McAtNlo_Jimmy.recon.AOD.v outDS user.DietrichLiko.test.01 --lcg –ce ce-fzk.gridka.de:2119/jobmanager-pbspro-atlasS

GANGA since Feb ATLAS Users: In total about 440, about 80 users per week

GANGA/PANDA GANGA plug-in has been developed to interface GANGA with PANDA pathena is an essential part of the plug-in The same job can be sent now to all ATLAS grid infrastructures – First prototype is available – Working on support for all options ….

ARDA Dashboard Provides an overview of all ATLAS activities – LCG System Monitoring WG – PANDA – Nordugrid

Conclusions Usage of the Grid for Distributed Analysis has been steadily rising in ATLAS – Many Distributed Analysis tutorials – Many new users – Increasing demand for user support All ATLAS grids are being used for analysis – OSG, EGEE, Nordugrid Good collaboration between developers of the analysis tools

gLite WMS A nearly final version is available at CERN – We expect it to be fully certified soon Bulksubmission – Without Condor DAG’s Improved sandbox handling Support for new CE etc. WMS site efficiency at IN2P3 for different experiments CMS/CrabLHCb/DIRAC Alice/Alien ATLAS/GANGA

PANDA on LCG PANDA is available as test installation on several LCG sites – Lyon + French cloud – Several other sites The issue is also related to ATLAS Production Ongoing discussion between the involved parties

User evaluations At March ATLAS software week At a recent PAT forum Notes are in preparations – CSC Note of SUSY team

17 Conclusions Both schemes are user-friendly. Differences between the GANGA and PanDA exist, but user can switch fairly easy from one to the other. Full user control over the job configuration Performance wise, analysis over large datasets can be performed in few hours. Many thanks to Caroline Collard, Stathes Paganis, Mark Hodgkinson, Dimitris Varouchas, Nurcan Ozturk and the H/A  μμ group for their feedback. Having the experience of running alone a small computer cluster and using the old-fashioned distributed analysis scheme, my conclusion is that using PanDA and GANGA is easier and less time consuming. Moreover, all feed-back received was very positive! (although there are still a few missing pieces) Concerns: 1) Data replication uniformity over the sites. 2) Integrity of data-files. Recovery of crashed jobs due to corrupted files not easy. (currently, user manually isolate the problematic files and re-submit the job in the old way) 3) Distinction between LCG and OSG not very productive. (situation is improving though, eg PanDA now runs on site ANALY_LYON) 4) Provide stable environment for users during development. 5) Maintain good performance with increasing number of users.