ATLAS DC2 ISGC-2005 Taipei 27th April 2005

Slides:



Advertisements
Similar presentations
Resources for the ATLAS Offline Computing Basis for the Estimates ATLAS Distributed Computing Model Cost Estimates Present Status Sharing of Resources.
Advertisements

Enabling Grids for E-sciencE Analysis of the ATLAS Rome Production Experience on the LCG Computing Grid Simone Campana, CERN/INFN EGEE.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Santiago González de la Hoz on behalf of ATLAS DC2 Collaboration EGC 2005 Amsterdam, 14/02/2005.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Current Monte Carlo calculation activities in ATLAS (ATLAS Data Challenges) Oxana Smirnova LCG/ATLAS, Lund University SWEGRID Seminar (April 9, 2003, Uppsala)
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
CMS Report – GridPP Collaboration Meeting VI Peter Hobson, Brunel University30/1/2003 CMS Status and Plans Progress towards GridPP milestones Workload.
ATLAS Data Challenge Production and U.S. Participation Kaushik De University of Texas at Arlington BNL Physics & Computing Meeting August 29, 2003.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
The LHC Computing Grid – February 2008 The Worldwide LHC Computing Grid Dr Ian Bird LCG Project Leader 25 th April 2012.
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
ATLAS Data Challenge Production Experience Kaushik De University of Texas at Arlington Oklahoma D0 SARS Meeting September 26, 2003.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
Zprávy z ATLAS SW Week March 2004 Seminář ATLAS SW CZ Duben 2004 Jiří Chudoba FzÚ AV CR.
Ian Bird LHC Computing Grid Project Leader LHC Grid Fest 3 rd October 2008 A worldwide collaboration.
The LHC Computing Grid – February 2008 The Challenges of LHC Computing Dr Ian Bird LCG Project Leader 6 th October 2009 Telecom 2009 Youth Forum.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
Grid User Interface for ATLAS & LHCb A more recent UK mini production used input data stored on RAL’s tape server, the requirements in JDL and the IC Resource.
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
The CMS Simulation Software Julia Yarba, Fermilab on behalf of CMS Collaboration 22 m long, 15 m in diameter Over a million geometrical volumes Many complex.
Performance of The NorduGrid ARC And The Dulcinea Executor in ATLAS Data Challenge 2 Oxana Smirnova (Lund University/CERN) for the NorduGrid collaboration.
ATLAS Grid Computing Rob Gardner University of Chicago ICFA Workshop on HEP Networking, Grid, and Digital Divide Issues for Global e-Science THE CENTER.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
ATLAS-specific functionality in Ganga - Requirements for distributed analysis - ATLAS considerations - DIAL submission from Ganga - Graphical interfaces.
INFSO-RI Enabling Grids for E-sciencE The EGEE Project Owen Appleton EGEE Dissemination Officer CERN, Switzerland Danish Grid Forum.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Overview of ATLAS Data Challenge Oxana Smirnova LCG/ATLAS, Lund University GAG monthly, February 28, 2003, CERN Strongly based on slides of Gilbert Poulard.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
ALICE Physics Data Challenge ’05 and LCG Service Challenge 3 Latchezar Betev / ALICE Geneva, 6 April 2005 LCG Storage Management Workshop.
ATLAS Experience on Large Scale Productions on the Grid CHEP-2006 Mumbai 13th February 2006 Gilbert Poulard (CERN PH-ATC) on behalf of ATLAS Data Challenges;
EXPERIENCE WITH ATLAS DISTRIBUTED ANALYSIS TOOLS S. González de la Hoz L. March IFIC, Instituto.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
LHC collisions rate: Hz New PHYSICS rate: Hz Event selection: 1 in 10,000,000,000,000 Signal/Noise: Raw Data volumes produced.
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
Bob Jones EGEE Technical Director
WLCG Tier-2 Asia Workshop TIFR, Mumbai 1-3 December 2006
(Prague, March 2009) Andrey Y Shevel
The EDG Testbed Deployment Details
Oxana Smirnova, Jakob Nielsen (Lund University/CERN)
U.S. ATLAS Grid Production Experience
The LHC Computing Grid Visit of Mtro. Enrique Agüera Ibañez
Data Challenge with the Grid in ATLAS
INFN-GRID Workshop Bari, October, 26, 2004
Grid related projects CERN openlab LCG EDG F.Fluckiger
Long-term Grid Sustainability
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Readiness of ATLAS Computing - A personal view
The LHC Computing Grid Visit of Her Royal Highness
Computing Report ATLAS Bern
Dagmar Adamova (NPI AS CR Prague/Rez) and Maarten Litmaath (CERN)
US CMS Testbed.
US ATLAS Physics & Computing
Grid Canada Testbed using HEP applications
CERN, the LHC and the Grid
LHC Data Analysis using a worldwide computing grid
LHC Computing Grid Project
ATLAS DC2 & Continuous production
Computing activities at Victoria
The LHC Computing Grid Visit of Professor Andreas Demetriou
LHC Computing Grid Project
The LHCb Computing Data Challenge DC06
Presentation transcript:

ATLAS DC2 ISGC-2005 Taipei 27th April 2005 Gilbert Poulard (CERN PH-ATC) on behalf of ATLAS Data Challenges; Grid and Operations teams

Overview Introduction ATLAS production system Data Challenge 2 ATLAS experiment ATLAS Data Challenges program ATLAS production system Data Challenge 2 The 3 Grid flavors (LCG; Grid3 and NorduGrid) ATLAS DC2 production Conclusions ISGC-2005 G. Poulard - CERN PH

Introduction: LHC/CERN LHC (CERN) Mont Blanc, 4810 m Geneva ISGC-2005 G. Poulard - CERN PH

The challenge of the LHC computing Storage – Raw recording rate 0.1 – 1 GBytes/sec Accumulating at 5-8 PetaBytes/year 10 PetaBytes of disk Processing – 200,000 of today’s fastest PCs ISGC-2005 G. Poulard - CERN PH

Introduction: ATLAS Detector for the study of high-energy proton-proton collisions. The offline computing will have to deal with an output event rate of 200 Hz. i.e 2x109 events per year with an average event size of 1.6 Mbyte. Researchers are spread all over the world. ISGC-2005 G. Poulard - CERN PH

Introduction: ATLAS experiment ~ 2000 Collaborators ~150 Institutes 34 Countries Diameter 25 m Barrel toroid length 26 m Endcap end-wall chamber span 46 m Overall weight 7000 Tons ISGC-2005 G. Poulard - CERN PH

Introduction: Data Challenges Scope and Goals of Data Challenges (DCs): Validate Computing Model Software Data Model DC1 (2002-2003) Put in place the full software chain Simulation of the data; digitization; pile-up Reconstruction Production system Tools (bookkeeping; monitoring; …) Intensive use of Grid “Build” the ATLAS DC community DC2 (2004) Similar exercise as DC1 BUT Use of the Grid middleware developed in several projects LHC Computing Grid project (LCG) to which CERN is committed Grid3 on US NorduGrid in Scandinavian countries ISGC-2005 G. Poulard - CERN PH

ATLAS Production System The production database, which contains abstract job definitions The Windmill supervisor that reads the production database for job definitions and present them to the different Grid executors in an easy-to-parse XML format The Executors, one for each Grid flavor, that receives the job-definitions in XML format and converts them to the job description language of that particular Grid Don Quijote, the ATLAS Data Management System, moves files from their temporary output locations to their final destination on some Storage Elements and registers the files in the Replica Location Service of that Grid In order to handle the task of ATLAS DC2 an automated Production system was developed. It consists of 4 components: ISGC-2005 G. Poulard - CERN PH

DC2 production phases Task Flow for DC2 data Pythia Persistency: Bytestream Raw Digits Task Flow for DC2 data ESD Digits (RDO) MCTruth Bytestream Raw Digits Mixing Reconstruction Events HepMC Hits MCTruth Geant4 Digitization Bytestream Raw Digits ESD Digits (RDO) MCTruth Events HepMC Hits MCTruth Pythia Geant4 Digitization Reconstruction Digits (RDO) MCTruth Events HepMC Hits MCTruth Geant4 Pile-up Bytestream Raw Digits ESD Bytestream Raw Digits Mixing Reconstruction Events HepMC Hits MCTruth Digits (RDO) MCTruth Geant4 Pile-up Bytestream Raw Digits 20 TB 5 TB 20 TB 30 TB ~5 TB Event Mixing Detector Simulation Digitization (Pile-up) Reconstruction Event generation Byte stream Persistency: Athena-POOL TB Physics events Min. bias Events Piled-up events Mixed events Mixed events With Pile-up ISGC-2005 G. Poulard - CERN PH Volume of data for 107 events

DC2 production phases Process No. of events ~ Event size ~ CPU time per event ~ Volume of data MB kSI2k-s TB Event generation 5 x 107 0.06 3 Simulation 107 2. 520 30 Pile-up Digitization 3 x 106 3. 150 15 6 20 Event mixing & Byte-stream 5.4 ATLAS DC2 started in July 2004 The simulation part was finished by the end of September and the pile-up and digitization parts by the end of November 10 million events were generated, fully simulated and digitized and ~2 Million events were “piled-up” Event mixing and reconstruction was done for 2.4 Million events in December. The Grid technology as provided the tools to perform this “massive” worldwide production ISGC-2005 G. Poulard - CERN PH

The 3 Grid flavors LCG (http://lcg.web.cern.ch/LCG/) The job of LHC Computing Grid Project - LCG - is to prepare the computing infrastructure for the simulation, processing and analysis of LHC data for all four of the LHC collaborations. This includes both the common infrastructure of libraries, tools and frameworks required to support the physics application software, and the development and deployment of the computing services needed to store and process the data, providing batch and interactive facilities for the worldwide community of physicists involved in LHC. Grid3 (http://www.ivdgl.org/grid2003/) The Grid3 collaboration has deployed and international Data Grid with dozens of sites and thousand of processors. The facility jointly by the US Grid project iVDGL, GriPhyN and PPDG and the US participants in the LHC experiments ATLAS and CMS. NorduGrid (http://www.nordugrid.org/) The aim of the NorduGrid collaboration is to deliver a robust, scalable, portable and fully featured solution for a global computational and data Grid system. NorduGrid develops and deploys a set of tools and services - the so-called ARC middleware, which is a free software. Both Grid3 and NorduGrid have similar approaches using the same foundations (GLOBUS) as LCG with slightly different middleware ISGC-2005 G. Poulard - CERN PH

The 3 Grid flavors: LCG-2 Number of sites; resources are evolving quickly ISGC-2005 G. Poulard - CERN PH

The 3 Grid flavors: NorduGrid NorduGrid is a research collaboration established mainly across Nordic Countries but includes sites from other countries. They contributed to a significant part of the DC1 (using the Grid in 2002). It supports production on several operating systems (non-RedHat 7.3 platforms). > 10 countries, 40+ sites, ~4000 CPUs, ~30 TB storage ISGC-2005 G. Poulard - CERN PH

The 3 Grid flavors: Grid3 Sep 04 30 sites, multi-VO shared resources ~3000 CPUs (shared) The deployed infrastructure has been in operation since November 2003 At this moment running 3 HEP and 2 Biological applications Over 100 users authorized to run in GRID3 ISGC-2005 G. Poulard - CERN PH

ATLAS DC2: countries (sites) Australia (1) Austria (1) Canada (4) CERN (1) Czech Republic (2) Denmark (4) France (1) Germany (1+2) Italy (7) Japan (1) Netherlands (1) Norway (3) Poland (1) Slovenia (1) Spain (3) Sweden (7) Switzerland (1) Taiwan (1) UK (7) USA (19) 20 countries 69 sites 13 countries; 31 sites 7 countries; 19 sites ISGC-2005 G. Poulard - CERN PH

ATLAS DC2 production Total ISGC-2005 G. Poulard - CERN PH

ATLAS DC2 production ISGC-2005 G. Poulard - CERN PH

ATLAS Production (July 2004 - April 2005) Rome (mix jobs) DC2 (short jobs period) DC2 (long jobs period) ISGC-2005 G. Poulard - CERN PH

Jobs Total As of 30 November 2004 20 countries 69 sites ~ 260000 Jobs ~ 2 MSi2k.months ISGC-2005 G. Poulard - CERN PH

Lessons learned from DC2 Main problems The production system was in development during DC2 The beta status of the services of the Grid caused troubles while the system was in operation For example the Globus RLS, the Resource Broker and the information system were unstable at the initial phase Specially on LCG, lack of uniform monitoring system The mis-configuration of sites and site stability related problems But also Human errors (for example “expired proxy”; bad registration of files) Network problems (connection lost between two processes) Data Management System problems (eg. connection with mass storage system) ISGC-2005 G. Poulard - CERN PH

Lessons learned from DC2 Main achievements To have run a large scale production on Grid ONLY, using 3 Grid flavors To have an automatic production system making use of Grid infrastructure Few “10 TB” of data have been moved among the different Grid flavors using DonQuijote (ATLAS Data Management) servers ~260000 jobs were submitted by the production system ~260000 logical files were produced and ~2500 jobs were run per day ISGC-2005 G. Poulard - CERN PH

Conclusions The generation, simulation and digitization of events for ATLAS DC2 have been completed using 3 flavors of Grid Technology (LCG; Grid3; NorduGrid) They have been proven to be usable in a coherent way for a real production and this is a major achievement This exercise has taught us that all the involved elements (Grid middleware, production system, deployment and monitoring tools, …) need improvements From July to end November 2004, the automatic production system has submitted ~260000 jobs, they consumed ~2000 kSI2k months of CPU and produced more than 60 TB of data If one includes on-going production one reaches ~700000 jobs, more than 100 TB and ~500 kSI2k.years ISGC-2005 G. Poulard - CERN PH