ATLAS DC2 & Continuous production

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

1 Databases in ALICE L.Betev LCG Database Deployment and Persistency Workshop Geneva, October 17, 2005.
ATLAS Analysis Model. Introduction On Feb 11, 2008 the Analysis Model Forum published a report (D. Costanzo, I. Hinchliffe, S. Menke, ATL- GEN-INT )
Current Monte Carlo calculation activities in ATLAS (ATLAS Data Challenges) Oxana Smirnova LCG/ATLAS, Lund University SWEGRID Seminar (April 9, 2003, Uppsala)
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
ATLAS-Specific Activity in GridPP EDG Integration LCG Integration Metadata.
December 17th 2008RAL PPD Computing Christmas Lectures 11 ATLAS Distributed Computing Stephen Burke RAL.
ATLAS Data Challenge Production and U.S. Participation Kaushik De University of Texas at Arlington BNL Physics & Computing Meeting August 29, 2003.
Alexandre A. P. Suaide VI DOSAR workshop, São Paulo, 2005 STAR grid activities and São Paulo experience.
Data Import Data Export Mass Storage & Disk Servers Database Servers Tapes Network from CERN Network from Tier 2 and simulation centers Physics Software.
LHC: ATLAS Experiment meeting “Conditions” data challenge Elizabeth Gallas - Oxford - August 29, 2009 XLDB3.
The first year of LHC physics analysis using the GRID: Prospects from ATLAS Davide Costanzo University of Sheffield
ATLAS and GridPP GridPP Collaboration Meeting, Edinburgh, 5 th November 2001 RWL Jones, Lancaster University.
F. Fassi, S. Cabrera, R. Vives, S. González de la Hoz, Á. Fernández, J. Sánchez, L. March, J. Salt, A. Lamas IFIC-CSIC-UV, Valencia, Spain Third EELA conference,
ATLAS Data Challenge Production Experience Kaushik De University of Texas at Arlington Oklahoma D0 SARS Meeting September 26, 2003.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
David Adams ATLAS ADA, ARDA and PPDG David Adams BNL June 28, 2004 PPDG Collaboration Meeting Williams Bay, Wisconsin.
ATLAS Data Challenges US ATLAS Physics & Computing ANL October 30th 2001 Gilbert Poulard CERN EP-ATC.
Zprávy z ATLAS SW Week March 2004 Seminář ATLAS SW CZ Duben 2004 Jiří Chudoba FzÚ AV CR.
GDB Meeting - 10 June 2003 ATLAS Offline Software David R. Quarrie Lawrence Berkeley National Laboratory
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
Post-DC2/Rome Production Kaushik De, Mark Sosebee University of Texas at Arlington U.S. Grid Phone Meeting July 13, 2005.
SC4 Planning Planning for the Initial LCG Service September 2005.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
ATLAS Grid Computing Rob Gardner University of Chicago ICFA Workshop on HEP Networking, Grid, and Digital Divide Issues for Global e-Science THE CENTER.
The ATLAS TAGs Database - Experiences and further developments Elisabeth Vinek, CERN & University of Vienna on behalf of the TAGs developers group.
Computing for Alice at GSI (Proposal) (Marian Ivanov)
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
David Stickland CMS Core Software and Computing
Overview of ATLAS Data Challenge Oxana Smirnova LCG/ATLAS, Lund University GAG monthly, February 28, 2003, CERN Strongly based on slides of Gilbert Poulard.
Victoria, Sept WLCG Collaboration Workshop1 ATLAS Dress Rehersals Kors Bos NIKHEF, Amsterdam.
1 A Scalable Distributed Data Management System for ATLAS David Cameron CERN CHEP 2006 Mumbai, India.
ATLAS Distributed Analysis Dietrich Liko IT/GD. Overview  Some problems trying to analyze Rome data on the grid Basics Metadata Data  Activities AMI.
David Adams ATLAS ATLAS Distributed Analysis (ADA) David Adams BNL December 5, 2003 ATLAS software workshop CERN.
Finding Data in ATLAS. May 22, 2009Jack Cranshaw (ANL)2 Starting Point Questions What is the latest reprocessing of cosmics? Are there are any AOD produced.
L. Perini DATAGRID WP8 Use-cases 19 Dec ATLAS short term grid use-cases The “production” activities foreseen till mid-2001 and the tools to be used.
ATLAS Distributed Analysis DISTRIBUTED ANALYSIS JOBS WITH THE ATLAS PRODUCTION SYSTEM S. González D. Liko
Dario Barberis: Conclusions ATLAS Software Week - 10 December Conclusions Dario Barberis CERN & Genoa University.
ATLAS Experience on Large Scale Productions on the Grid CHEP-2006 Mumbai 13th February 2006 Gilbert Poulard (CERN PH-ATC) on behalf of ATLAS Data Challenges;
Joe Foster 1 Two questions about datasets: –How do you find datasets with the processes, cuts, conditions you need for your analysis? –How do.
CERN IT Department CH-1211 Genève 23 Switzerland t EGEE09 Barcelona ATLAS Distributed Data Management Fernando H. Barreiro Megino on behalf.
Monthly video-conference, 18/12/2003 P.Hristov1 Preparation for physics data challenge'04 P.Hristov Alice monthly off-line video-conference December 18,
LHCb Computing 2015 Q3 Report Stefan Roiser LHCC Referees Meeting 1 December 2015.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
LHCb computing model and the planned exploitation of the GRID Eric van Herwijnen, Frank Harris Monday, 17 July 2000.
ATLAS Distributed Analysis S. González de la Hoz 1, D. Liko 2, L. March 1 1 IFIC – Valencia 2 CERN.
Wouter Verkerke, NIKHEF Status of DC3 (CSC) production Wouter Verkerke (NIKHEF)
Database Replication and Monitoring
U.S. ATLAS Grid Production Experience
Data Challenge with the Grid in ATLAS
ALICE analysis preservation
INFN-GRID Workshop Bari, October, 26, 2004
European Organization for Nuclear Research
Bernd Panzer-Steindel, CERN/IT
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
Philippe Charpentier CERN – LHCb On behalf of the LHCb Computing Group
Readiness of ATLAS Computing - A personal view
Savannah to Jira Migration
ATLAS DC2 ISGC-2005 Taipei 27th April 2005
Computing Report ATLAS Bern
Zhongliang Ren 12 June 2006 WLCG Tier2 Workshop at CERN
Simulation use cases for T2 in ALICE
US ATLAS Physics & Computing
R. Graciani for LHCb Mumbay, Feb 2006
Grid Computing in CMS: Remote Analysis & MC Production
The ATLAS Computing Model
HEC Beam Test Software schematic view T D S MC events ASCII-TDS
Short to middle term GRID deployment plan for LHCb
Presentation transcript:

ATLAS DC2 & Continuous production ATLAS Physics Plenary 26 February 2004 Gilbert Poulard CERN PH-ATC

G. Poulard - ATLAS Physics plenary Outline Goals Operation; scenario; time scale Resources Production; Grid and Tools Role of Tiers Comments on schedule Analysis will be covered in tomorrow’s talk and in the next software week. February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary DC2: goals At this stage the goal includes: Full use of Geant4; POOL; LCG applications Pile-up and digitization in Athena Deployment of the complete Event Data Model and the Detector Description Simulation of full ATLAS and 2004 combined Testbeam Test the calibration and alignment procedures Use widely the GRID middleware and tools Large scale physics analysis Computing model studies (document end 2004) Run as much as possible of the production on LCG-2 February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary DC2 operation Consider DC2 as a three-part operation: part I: production of simulated data (May-June 2004) needs Geant4, digitization and pile-up in Athena, POOL persistency “minimal” reconstruction just to validate simulation suite will run on any computing facilities we can get access to around the world part II: test of Tier-0 operation (July 2004) needs full reconstruction software following RTF report design, definition of AODs and TAGs (calibration/alignment and) reconstruction will run on Tier-0 prototype as if data were coming from the online system (at 10% of the rate) output (ESD+AOD) will be distributed to Tier-1s in real time for analysis part III: test of distributed analysis on the Grid (August-Oct. 2004) access to event and non-event data from anywhere in the world both in organized and chaotic ways in parallel: run distributed reconstruction on simulated data February 26, 2004 G. Poulard - ATLAS Physics plenary

DC2: Scenario & Time scale September 03: Release7 March 17th: Release 8 (production) May 3rd 04: July 1st 04: “DC2” August 1st: Put in place, understand & validate: Geant4; POOL; LCG applications Event Data Model Digitization; pile-up; byte-stream Conversion of DC1 data to POOL; large scale persistency tests and reconstruction Testing and validation Run test-production Start final validation Start simulation; Pile-up & digitization Event mixing Transfer data to CERN Intensive Reconstruction on “Tier0” Distribution of ESD & AOD Calibration; alignment Start Physics analysis Reprocessing February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary Task Flow for DC2 data Bytestream Raw Digits ESD Digits (RDO) MCTruth Bytestream Raw Digits Mixing Reconstruction Events HepMC Hits MCTruth Geant4 Digitization Bytestream Raw Digits ESD Digits (RDO) MCTruth Events HepMC Hits MCTruth Pythia Geant4 Digitization Reconstruction Digits (RDO) MCTruth Events HepMC Hits MCTruth Geant4 Pile-up Bytestream Raw Digits ESD Bytestream Raw Digits Mixing Reconstruction Events HepMC Hits MCTruth Digits (RDO) MCTruth Geant4 Pile-up Bytestream Raw Digits 18 TB 5 TB 24 TB 75 TB ~2 TB Event Mixing Detector Simulation Digitization (Pile-up) Reconstruction Event generation Byte stream Persistency: Athena-POOL TB Physics events Min. bias Events Piled-up events Mixed events Mixed events With Pile-up February 26, 2004 G. Poulard - ATLAS Physics plenary Volume of data for 107 events

G. Poulard - ATLAS Physics plenary Scenario for DC2 For Tier-0 exercise Event generation Detector simulation Pile-up Detector response RDO Event mixing Byte-stream No MCTruth Data transfer Reconstruction ESD production AOD production Streaming Distribution of data For Physics & Detectors studies Event generation Detector simulation Pile-up Detector response RDO Event mixing Byte-stream Data transfer Reconstruction ESD production AOD production Q1: Trigger & Filtering in the chain? Q2: ATLFAST? Both are needed for Computing model studies February 26, 2004 G. Poulard - ATLAS Physics plenary

Still working on scenario Production scenario Input Output Comments Event generation none Generated events < 2 GB files G4 simulation Generated Events “part of” Hits + MCTruth ~ 400-500 jobs/day ~ 500 GB/day ~ 6 MB/s Detector response (Generated events) 1 file Digits +MCTruth RDO (or BS) No MCTruth if BS Pile-up Hits “signal” Hits “min.b” Several 10 files ~ 1000 jobs/day ~ 10 GB/job ~ 10 TB/day ~ 125 MB/s Byte-stream “pile-up” data RDO 1 (or few) files BS Still working on scenario Events mixing RDO or BS Several files “ Reconstruction ESD AOD production AOD Streaming? February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary DC2 resources Process No. of events Time duration CPU power Volume of data At CERN Off site months kSI2k TB Simulation 107 2 600 25 5 20 Phase I (May-June) Pile-up (*) Digitization 400 75 15 60 Byte-stream (small) 16 Total Phase I 1000 120 40 96 Reconstruction Tier-0 0.5 10 Phase II (July) Tier-1 Total 130 45 111 February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary DC2 resources (based on Geant3 numbers) Process Time duration Input Output Event size I/O rate months MB MB/s GB/day Simulation 2 2.4 5.8 500 Pile-up Digitization () no “0” suppression 120 11000 1.8 (7.5) 4.3 (18.2) 365 (1600) Reconstruction 0.5 17.4 1500 4.8 420 Process Time duration Volume of data (Bytestream) Bandwith Weeks TB MB/s Data Transfer 6 12 3.3 February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary DC2; Grid & Production tools We foresee to use: 3 Grid flavors (LCG; Grid3; Nordu-Grid) “batch” systems (LSF; …) Automated production system New production DB (Oracle) Supervisor-executer component model Windmill supervisor project Executers for each Grid and LSF Data management system Don Quijote DMS project Successor of Magda … but uses native catalogs AMI for bookkeeping Going to web services Integrated to POOL February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary Atlas Production System schema Task = [job]* Dataset = [partition]* AMI JOB DESCRIPTION Location Hint (Task) Task (Dataset) Task Transf. Definition Data Management System + physics signature Human intervention Job Run Info Location Hint (Job) Job (Partition) Partition Transf. Definition Executable name Release version signature Supervisor 1 Supervisor 2 Supervisor 3 Supervisor 4 US Grid Executer LCG Executer NG Executer LSF Executer Chimera RB RB US Grid LCG NG Local Batch February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary Tiers in DC2 Tier-0 20% of simulation will be done at CERN All data in ByteStream format (~16 TB) will be copied to CERN Reconstruction will be done at CERN (in ~10 days). Reconstruction output (ESD) will be exported in 2 copies from Tier-0 ( 2 X ~5 TB). Primary AOD can be produced at CERN Still under discusssion February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary Tiers in DC2 Tier-1s will have to Host simulated data produced by them or coming from Tier-2; plus ESD (& AOD) coming from Tier-0 Run reconstruction in parallel to Tier-0 exercise (~2 months) This will include links to MCTruth Produce and host ESD and AOD Provide access to the ATLAS V.O. members Tier-2s Run simulation (and other components if they wish to) Copy (replicate) their data to Tier-1 It would be easier if both Tier-1s and Tier-2s are on Grid Strongly recommended ATLAS is committed to LCG All information should be entered into the relevant database and catalog February 26, 2004 G. Poulard - ATLAS Physics plenary

G. Poulard - ATLAS Physics plenary Comments on schedule The change of the schedule has been “driven” by ATLAS side: the readiness of the software Combined test beam has a highest priority The availability of the production tools The integration with grid is not always easy Grid side The readiness of LCG We would prefer run Grid only! Priorities are not defined by ATLAS only For the Tier-0 exercise It will be difficult to define the starting date before we have a better idea how work the “pile-up” and the “event-mixing” processes. February 26, 2004 G. Poulard - ATLAS Physics plenary