BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy.

Slides:



Advertisements
Similar presentations
Your university or experiment logo here BaBar Status Report Chris Brew GridPP16 QMUL 28/06/2006.
Advertisements

B A B AR and the GRID Roger Barlow for Fergus Wilson GridPP 13 5 th July 2005, Durham.
INFN - Ferrara BaBarGrid Meeting SPGrid Efforts in Italy BaBar Collaboration Meeting - SLAC December 11, 2002 Enrica Antonioli - Paolo Veronesi.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
DataGrid Kimmo Soikkeli Ilkka Sormunen. What is DataGrid? DataGrid is a project that aims to enable access to geographically distributed computing power.
1 Use of the European Data Grid software in the framework of the BaBar distributed computing model T. Adye (1), R. Barlow (2), B. Bense (3), D. Boutigny.
K.Harrison CERN, 23rd October 2002 HOW TO COMMISSION A NEW CENTRE FOR LHCb PRODUCTION - Overview of LHCb distributed production system - Configuration.
Workload Management Massimo Sgaravatto INFN Padova.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
ATLAS Off-Grid sites (Tier-3) monitoring A. Petrosyan on behalf of the ATLAS collaboration GRID’2012, , JINR, Dubna.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
5 November 2001F Harris GridPP Edinburgh 1 WP8 status for validating Testbed1 and middleware F Harris(LHCb/Oxford)
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Computing for ILC experiment Computing Research Center, KEK Hiroyuki Matsunaga.
Computing and LHCb Raja Nandakumar. The LHCb experiment  Universe is made of matter  Still not clear why  Andrei Sakharov’s theory of cp-violation.
3 Sept 2001F HARRIS CHEP, Beijing 1 Moving the LHCb Monte Carlo production system to the GRID D.Galli,U.Marconi,V.Vagnoni INFN Bologna N Brook Bristol.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
F.Fanzago – INFN Padova ; S.Lacaprara – LNL; D.Spiga – Universita’ Perugia M.Corvo - CERN; N.DeFilippis - Universita' Bari; A.Fanfani – Universita’ Bologna;
INFSO-RI Enabling Grids for E-sciencE Logging and Bookkeeping and Job Provenance Services Ludek Matyska (CESNET) on behalf of the.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Distribution After Release Tool Natalia Ratnikova.
Nick Brook Current status Future Collaboration Plans Future UK plans.
1 DIRAC – LHCb MC production system A.Tsaregorodtsev, CPPM, Marseille For the LHCb Data Management team CHEP, La Jolla 25 March 2003.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
11 December 2000 Paolo Capiluppi - DataGrid Testbed Workshop CMS Applications Requirements DataGrid Testbed Workshop Milano, 11 December 2000 Paolo Capiluppi,
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
Ashok Agarwal University of Victoria 1 GridX1 : A Canadian Particle Physics Grid A. Agarwal, M. Ahmed, B.L. Caron, A. Dimopoulos, L.S. Groer, R. Haria,
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
And Tier 3 monitoring Tier 3 Ivan Kadochnikov LIT JINR
OSG Tier 3 support Marco Mambelli - OSG Tier 3 Dan Fraser - OSG Tier 3 liaison Tanya Levshina - OSG.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
Evolution of a High Performance Computing and Monitoring system onto the GRID for High Energy Experiments T.L. Hsieh, S. Hou, P.K. Teng Academia Sinica,
US LHC OSG Technology Roadmap May 4-5th, 2005 Welcome. Thank you to Deirdre for the arrangements.
A GRID solution for Gravitational Waves Signal Analysis from Coalescing Binaries: preliminary algorithms and tests F. Acernese 1,2, F. Barone 2,3, R. De.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Andrew McNab - Manchester HEP - 17 September 2002 UK Testbed Deployment Aim of this talk is to the answer the questions: –“How much of the Testbed has.
INFSO-RI Enabling Grids for E-sciencE ARDA Experiment Dashboard Ricardo Rocha (ARDA – CERN) on behalf of the Dashboard Team.
Testing and integrating the WLCG/EGEE middleware in the LHC computing Simone Campana, Alessandro Di Girolamo, Elisa Lanciotti, Nicolò Magini, Patricia.
Tier3 monitoring. Initial issues. Danila Oleynik. Artem Petrosyan. JINR.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
INFSO-RI Enabling Grids for E-sciencE CRAB: a tool for CMS distributed analysis in grid environment Federica Fanzago INFN PADOVA.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
Enabling Grids for E-sciencE CMS/ARDA activity within the CMS distributed system Julia Andreeva, CERN On behalf of ARDA group CHEP06.
Gennaro Tortone, Sergio Fantinel – Bologna, LCG-EDT Monitoring Service DataTAG WP4 Monitoring Group DataTAG WP4 meeting Bologna –
D.Spiga, L.Servoli, L.Faina INFN & University of Perugia CRAB WorkFlow : CRAB: CMS Remote Analysis Builder A CMS specific tool written in python and developed.
LHCb 2009-Q4 report Q4 report LHCb 2009-Q4 report, PhC2 Activities in 2009-Q4 m Core Software o Stable versions of Gaudi and LCG-AA m Applications.
Geant4 GRID production Sangwan Kim, Vu Trong Hieu, AD At KISTI.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
Job Priorities and Resource sharing in CMS A. Sciabà ECGI meeting on job priorities 15 May 2006.
BaBar & Grid Eleonora Luppi for the BaBarGrid Group TB GRID Bologna 15 febbraio 2005.
Bob Jones EGEE Technical Director
Workload Management Workpackage
Regional Operations Centres Core infrastructure Centres
BaBar-Grid Status and Prospects
Eleonora Luppi INFN and University of Ferrara - Italy
EGEE Middleware Activities Overview
U.S. ATLAS Grid Production Experience
Moving the LHCb Monte Carlo production system to the GRID
INFN-GRID Workshop Bari, October, 26, 2004
Ruslan Fomkin and Tore Risch Uppsala DataBase Laboratory
Job workflow Pre production operations:
DØ MC and Data Processing on the Grid
Gridifying the LHCb Monte Carlo production system
MonteCarlo production for the BaBar experiment on the Italian grid
The LHCb Computing Data Challenge DC06
Presentation transcript:

BaBar Grid Computing Eleonora Luppi INFN and University of Ferrara - Italy

Eleonora Luppi - HEPDG May BaBar Grid Project Overview HEP is one of the more interesting and interested testbed for GRID fully distributed technology BaBar is a running experiments and needs a huge amount of computing resources BaBar Grid Project motivations : Simulation Production (SP) : –25 production sites running the same executable –Simulation Production (SP) in BaBar is a typical distributed effort –The Grid concept is very appealing for SP Analysis: BaBar uses several Tier-A in analysis context –A tool to automatically submit jobs where the data and CPU cycles are available could be very useful The GridKa Tier-A will only be accessible through Grid Data Distribution: –Can be seen as a Grid application –The extraction from a Tier-A and transfer to a Tier-C of a deep-copy collection could nicely make use of Grid tools

Eleonora Luppi - HEPDG May BaBar Grid Project Overview (cont.) The Babar Grid group has started to study the possibility to gridify Babar software At the beginning, no stable middleware was available Different approaches to start: Soft migration from “standard” BaBar software to Grid compatible software Full gridification of BaBar software Grid tools for data distribution Simulation production, simpler than analysis, was the starting point

Eleonora Luppi - HEPDG May Standard simulation production Collect and centrally manage requests for simulation –Many user requests, many production jobs (>500,000), many events –Web interface to relational DB Distribute requests to production sites –Separate request into runs, usually 2K events, each of them requires 5-10 hours of computing –Assign to production sites “allocations” whose size depends on the size of the production farm (typically, a few million events) Jobs require: –A specific release of the BaBar software installed –Configuration and Conditions information for the BaBar detector, stored in an Objectivity database at each site. –Detector noise information (Background Trigger), stored in Root files, accessed from an xrootd server at each site Manage jobs at each site (ProdTools) –Set of perl scripts, interacting with relational DB –Automatic procedure to produce, merge and import runs to SLAC

Eleonora Luppi - HEPDG May Full Grid approach In Italy Simulation Production was concentrated in one farm An independent group started to implement SP on the Grid Current Resources LCG middleware with SLC OS About 400 shared CPUs available for production AMS, for Objectivity db, and Xrootd servers installed in Ferrara, Naples and Padua ProdTools installed on User Interface in Ferrara BaBar Resource Broker in Ferrara 4 sites involved during production performance tests: Ferrara, Naples, Bari and Padua

Eleonora Luppi - HEPDG May Full Grid approach Italy case BaBar SP can use almost all Grid.it resources (not only dedicated ones)

Eleonora Luppi - HEPDG May Production scheme

Eleonora Luppi - HEPDG May Software Installation Simulation Software packaged and installed on involved sites Package distributed and tagged using LCG tools: –ManageSoftwareTool (submitted like a standard job) –Transfer packages from closest SE to WNs –Run scripts to install the simulation software –lcg-ManageVOTag Main script –Set all variables needed –Run Moose (SP executable) –Compress and transfer output to the SE ProdTools –spsub integrated with standard Grid submission –Output retrieved from SE to UI using a custom script –spmerge integrated in standard Grid submission Full Grid recipe

Eleonora Luppi - HEPDG May Test of true Production Results Produced some thousands of runs of real SP6 data Mean elapsed time for run: ~ 8 hours (average on different sites) A few jobs (<5%) automatically resubmitted after failure due to wrong sites configuration SP on Grid.it Future Plans Add new Grid.it sites to the BaBar VO Add specific BaBar resources (background trigger and Cdb servers) on more sites Gridify BaBar resources on Italian Tier 1 Start true production of new SP8 software release on the “Grid Farm” Coordination with other national Grid sites

Eleonora Luppi - HEPDG May Smooth migration approach BaBar in UK already had distributed infrastructure with 7 farms, each with the BaBar software installed and local Objy CondDB etc. Decision was to initially add “grid” to SP production on these farms rather than gridify the SP software Used Globus to turn the remote farms into a single pseudo batch system SP Jobs are built, merged and exported as at any other SP site Submission is to Globus rather than a local batch system The software is made up of three parts: –Submitter –Unpacker –Monitoring When a working prototype was ready, started moving that to LCG migration recipe

Eleonora Luppi - HEPDG May Completing migration… UK farms reinstalled with SL All BaBar UK Farm site are also GridPP Tier 2 sites LCG installed with minimal changes on UK farms Merging work done for UK-SPGrid and Italian-SPGrid Rewrite monitoring to monitor LCG Jobs Installation of BaBar SP8 software in LCG environment in progress

Eleonora Luppi - HEPDG May US and Canada Grid Activity On the west side of Atlantic Ocean… Slac is an Open Science Grid site BaBar should use OSG at Slac Sp Grid already tested in Europe, easier than analysis BaBar should adopt OSG for SPGrid In Canada BaBar Simulation Production is planned to use GridX1 resources A LCG interface is possible, like ATLAS

Eleonora Luppi - HEPDG May SPGrid Summary We started with unstable middleware and different approaches Now we can follow a common line We coordinate our work to obtain a common SPGrid structure running Monte Carlo production In Europe we can run SPgrid jobs Italy, UK and soon in Germany Canadian GridX1 infrastructure can be used for SP and interfaced to European one We want to maintain compatibility with US OSG path We will take into account the disappearing of Objectivity as Cdb SP8 can be done in SPGrid

Eleonora Luppi - HEPDG May “Analysis” Grid “Analysis” in the following means any kind of experimental data processing Problems to face: –Large amount of data to analyze –Remote access to input data –Different kinds and sizes of output depending on the type of analysis A step by step approach can be useful to reduce the complexity of the problem Skimming of data is a candidate starting point

Eleonora Luppi - HEPDG May Skim Production in the Grid - Outline Short definition : GRID = Submit a job locally and run it anywhere. Submit skim jobs from a central site (i.e.SLAC) Check job requirements and run the job at an appropriate Tier-site Move job output to the pre-defined output area (~ 100 MB/job) Import collection to be skimmed. Skim Operator Site Manager

Eleonora Luppi - HEPDG May Site Requirements Software release has to exist. Conditions database has to exists. Input collection has to be available. Local worker nodes need to have means to move data to remote locations using Grid Middleware. Similar to SP New

Eleonora Luppi - HEPDG May Data Distribution: Xrootd Analysis needs efficient distributed data access –Scalable –Fault tolerant –Secure Xrootd could be seen as an alternative to standard Grid (LCG...) solutions –Xrootd was proposed also as a solution for some LHC experiments –Xrootd is distributed as part of CERN’s root package

Eleonora Luppi - HEPDG May “Analysis” Summary Step by step approach for the analysis Skimming seems to be a good starting point We need to study data handling possibilities to obtain an efficient and simple way to handle and maintain data, also in non-BaBar sites Results of tests with grid-analysis special infrastructures (see Alien and children) tell us that they can be interesting but need more support and stability

Eleonora Luppi - HEPDG May Conclusions SPGrid is a reality for true SP production We are converging to a common solution With the Grid resources available in the different national grids, we could have a huge quantity of computing time for our simulations Our next goal is to have an Analysis Grid capable of running “standard” analysis programs (i.e. Skimming) We need to give special attention to data handling tools