Monte Carlo Production and Reprocessing at DZero

Slides:



Advertisements
Similar presentations
GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Advertisements

CMS Applications Towards Requirements for Data Processing and Analysis on the Open Science Grid Greg Graham FNAL CD/CMS for OSG Deployment 16-Dec-2004.
McFarm: first attempt to build a practical, large scale distributed HEP computing cluster using Globus technology Anand Balasubramanian Karthik Gopalratnam.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Rod Walker IC 13th March 2002 SAM-Grid Middleware  SAM.  JIM.  RunJob.  Conclusions. - Rod Walker,ICL.
6/2/2015 Michael Diesburg HCP Distributed Computing at the Tevatron D0 Computing and Event Model Michael Diesburg, Fermilab For the D0 Collaboration.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Grid Job and Information Management (JIM) for D0 and CDF Gabriele Garzoglio for the JIM Team.
TechFair ‘05 University of Arlington November 16, 2005.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Deploying and Operating the SAM-Grid: lesson learned Gabriele Garzoglio for the SAM-Grid Team Sep 28, 2004.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Installation Sites. Active SAM-Grid Sites  The following sites are actively used for different types of jobs Monte Carlo – production of simulated events.
SamGrid– A Reality of “Grid” Computing –SamGrid– Adam Lyon (Fermilab Computing Division and DØ Experiment) GridKa School’04 September, 2004 Outline Introduction.
1 st December 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
DOSAR Workshop at Sao Paulo Dick Greenwood What’s Next for DOSAR? Dick Greenwood Louisiana Tech University 1 st DOSAR Workshop at the Sao Paulo, Brazil.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
DØ Computing Model & Monte Carlo & Data Reprocessing Gavin Davies Imperial College London DOSAR Workshop, Sao Paulo, September 2005.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
Status of UTA IAC + RAC Jae Yu 3 rd DØSAR Workshop Apr. 7 – 9, 2004 Louisiana Tech. University.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
The SAM-Grid and the use of Condor-G as a grid job management middleware Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
22 nd September 2003 JIM for CDF 1 JIM and SAMGrid for CDF Mòrag Burgon-Lyon University of Glasgow.
26SEP03 2 nd SAR Workshop Oklahoma University Dick Greenwood Louisiana Tech University LaTech IAC Site Report.
São Paulo Regional Analysis Center SPRACE Status Report 22/Aug/2006 SPRACE Status Report 22/Aug/2006.
Dzero MC production on LCG How to live in two worlds (SAM and LCG)
The SAM-Grid / LCG Interoperability Test Bed Gabriele Garzoglio ( ) Speaker: Pierre Girard (
DØSAR a Regional Grid within DØ Jae Yu Univ. of Texas, Arlington THEGrid Workshop July 8 – 9, 2004 Univ. of Texas at Arlington.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
Amber Boehnlein, FNAL Accelerator Based Physics: ATLAS CDF CMS DO STAR Amber Boehnlein OSG Consortium Meeting January 24, 2006.
High Energy Physics & Computing Grids TechFair Univ. of Arlington November 10, 2004.
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
The CMS Simulation Software Julia Yarba, Fermilab on behalf of CMS Collaboration 22 m long, 15 m in diameter Over a million geometrical volumes Many complex.
Jan 10, 2007, Clermont-Ferrand Tibor Kurca, Tutorial Grille1 DØ Computing Introduction - Fermilab & Tevatron & DØ Experiment DØ Computing Model 1. data.
High Energy FermiLab Two physics detectors (5 stories tall each) to understand smallest scale of matter Each experiment has ~500 people doing.
GRID activities in Wuppertal D0RACE Workshop Fermilab 02/14/2002 Christian Schmitt Wuppertal University Taking advantage of GRID software now.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
Worldwide Data Processing with SAMGrid. As experiments refine their understanding of raw data, a point is reached where it becomes desirable to reanalyze.
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
Analysis Tools at D0 PPDG Analysis Grid Computing Project, CS 11 Caltech Meeting Lee Lueking Femilab Computing Division December 19, 2002.
December 07, 2006Parag Mhashilkar, Fermilab1 Samgrid – OSG Interoperability Parag Mhashilkar, Fermi National Accelerator Laboratory.
MC Production in Canada Pierre Savard University of Toronto and TRIUMF IFC Meeting October 2003.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
July 26, 2007Parag Mhashilkar, Fermilab1 DZero On OSG: Site And Application Validation Parag Mhashilkar, Fermi National Accelerator Laboratory.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
DØ Computing Model and Operational Status Gavin Davies Imperial College London Run II Computing Review, September 2005.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
5/12/06T.Kurca - D0 Meeting FNAL1 p20 Reprocessing Introduction Computing Resources Architecture Operational Model Technical Issues Operational Issues.
PROTO-GRID Status of Grid-enabled UTA McFarm software Tomasz Wlodek University of the Great State of Texas At Arlington.
Production Resources & Issues p20.09 MC-data Regeneration
ALICE analysis preservation
OUHEP STATUS Hardware OUHEP0, 2x Athlon 1GHz, 2 GB, 800GB RAID
R. Graciani for LHCb Mumbay, Feb 2006
Tibor Kurca, Tutorial CPPM
 YongPyong-High Jan We appreciate that you give an opportunity to have this talk. Our Belle II computing group would like to report on.
DØ MC and Data Processing on the Grid
Lee Lueking D0RACE January 17, 2002
HEP Data Grid for DØ and Its Regional Grid, DØ SAR
The LHCb Computing Data Challenge DC06
Presentation transcript:

Monte Carlo Production and Reprocessing at DZero SAMGrid Monte Carlo Production and Reprocessing at DZero Eduardo Gregores VII DOSAR Workshop

Overview DZero Production Software SAM and SAMGrid, the Fermilab Grid Monte Carlo Production with SAMGrid Reprocessing with SamGrid SAMGrid on OSG and LHCG Eduardo Gregores VII DOSAR Workshop

DZero Software Analysis Software Production Software All analyze packages (topanalyze, bana, cafe, ...) Takes data from any stage of processing (raw, tmb or root). Run on different resources from laptops to dedicated farms. Customized by the end user. Production Software Common use by the collaboration. Heavy work; must run on computer farms or grids. Real Data processing. Simulated Data processing. Eduardo Gregores VII DOSAR Workshop

DZero Production Software Real Data: Online Data Aquisition System Select and store the events of interest. Triggers L1, L2, L3 and filters. Builds the Raw data – collection of hits and ADC counts. Software independent. Data Reconstruction. Identify physical objetcs and its kinematical properties. Transforms the Raw data into Thumbnail format. Software dependent. Data Skimming: Select the events based on its physical characteristics. Transform the raw data into Root format for the analysis. Simulated Data: Event generation: Generates physical events of interest. Pythia, Alpgen, Herwig, Isajet, Comphep, etc. External software. Experiment Simulation: Full simulation of DZero RunII detector D0GSTAR -- DØ GEANT Simulation of the Total Apparatus Response How much energy is deposited in the active areas of the detector by the particle produced by the generator. Raw data simulation: D0SIM – Electronics Simulation. Adds noise and inefficiencies to the events from D0GSTAR. Merges hard scatter and minimum bias events. Adds calorimeter pileup from previous crossings. Same functionality as for real data. Eduardo Gregores VII DOSAR Workshop

DZero Production Activities Production of simulated data: Production chain managed by MC_Runjob. Data catalog, storage and retrieval managed by SAM. Produced mostly by offsite farms on SAMGrid, the Fermilab Grid. More then 12 sites in 8 countries LU, LTU, OU, UTA, MSU, UW, Sprace, WestGrid, Nikhef, ccin2p3, Gridka, FZU and Tata to list some. Submission and monitoring by SAMGrid, the DZero production Grid. Based on JIM (Job Information Management) suite. Use of LCG and OSG as SAMGrid resources has just begun. Reprocessing of real data. Reconstruct the raw data according to latest software. All raw data collect to date must be reprocessed. First reprocessing (p14) made at Fermilab facility. Second reprocessing (p17) made mostly off site using SAMGrid. Refixing of produced thumbnails. Both simulated and real data. Last refixing used LCG and OSG as SAMGrid resources. Data Skimming. Uses Fermilab computing resources. Real Data Reconstruction: Increasing luminosity increases the load for online reconstruction. On the move into SAMGrid. First time online data is shipped off site processing. Eduardo Gregores VII DOSAR Workshop

SAM: A Grid-like Data Handling System Organized system of distributed servers. Store and retrieve files and associated metadata. Track locations and metadata for each file in the system. Interface to a mass storage system on a robotic tape store. Cache files on local disk and make then avaialable for other stations. GridFTP enabled for file transfers using globus (prefered method). Provide methods of job submission for local or grid-aware systems. Adaptable to grid submission systems using Condor and Globus grid tools. Used by DZero and CDF Experiments More then 50 SAM stations worldwide Eduardo Gregores VII DOSAR Workshop

SAM in Numbers: The Petabyte Era 18 Million Files Delivered by the Stations 9 PetaBytes of Data Delivered by the Stations 8 Million Files Stored on Tape 1.8 PetaBytes of Data Stored on Tape Eduardo Gregores VII DOSAR Workshop

SAMGrid Eduardo Gregores VII DOSAR Workshop

SAMGrid Architecture Eduardo Gregores VII DOSAR Workshop

SAMGrid Job Flow Eduardo Gregores VII DOSAR Workshop

SAMGrid Execution Sites  Station Name   Type   Site   Universe   Experiment     alicenext   SAMGRID     Wuppertal   prd   d0     ccin2p3-grid1     Wuppertal2LCG     Wuppertal2LCG_Devel   test_prd     cms-grid     CMS-FNAL-WC1     d0_fzu_prague     FZU_GRID     d0karlsruhe     GridKa     d0ppdg-wisconsin-2     Wisconsin     fnal-farm     FNAL     imperial-prd     Imperial College London   station_prd     ltu-cct     LTU-CCT     luhep     LUHEP     mcgill     Zephir-McGill     msu2     MSU     oscer     OSCER     osg-ouhep     Samgrid-OSG     ouhep     OUHEP     sprace     SPRACE     tata-d0-mcfarm     Tata     westgrid-ubc     WestGrid  Eduardo Gregores VII DOSAR Workshop

SAMGrid Submission Sites   Scheduler Name   Machine Name   Platform   Users  Max Jobs     d0farm.hecr.tifr.res.in   d0farm.hecr.tifr.res.in  I386-LINUX-RH72    1   200     hep.westgrid.ca   hep.westgrid.ca   I386-LINUX-RH9    4     luhep03.lunet.edu   luhep03.lunet.edu   14     sam.farm.particle.cz   sam.farm.particle.cz     samgrid.fnal.gov   samgrid.fnal.gov   50    Eduardo Gregores VII DOSAR Workshop

MC Production on SAMGrid DZero Software Latest Cut SAMGrid Software Latest Cuts Package Version Tarball in SAM D0 Release p17.09.01 d0_MC_p17.09.01_v9.tar.gz MC_Runjob v06-05-19 mc_runjob_v06-05-19_v2.tar.gz Magnetic Field v00-01-00 MagField_v00-01-00.tar.gz Product 2 Sep 2005 22 Nov 2005 jim_advertise v1_4_7 jim_client v2_1_27 v2_1_37 jim_config v1_2_16 v1_2_17 jim_job_managers v2_2_70 v2_2_79 jim_merge v1_18 jim_sandbox v2_4_10 samgrid_util v2_0_12 v2_0_13 sam_batch_adapter v1_0_1_18 v1_0_1_21 sam_client v1_0_60 v1_0_66 sam_fcp v1_0_19 v1_0_20 sam_gsi_config v2_2_19 -q vdt v2_2_22 -q vdt tomcat v4_3_3_9 vdt v1_1_14_13 xmldb_client v2_0_11 xmldb_server v1_0_3 Job description file for Monte Carlo production +notify_user = gregores@fnal.gov check_consistency = true d0_release_version = p17.09.01 group = mcc99 instances = 1 job_type = dzero_monte_carlo jobfiles_dataset = sprace_mcp17_v00-09-20_e minbias_dataset = Zerobias_p17MC_avelumi_1_rename2 runjob_numevts = 200000 runjob_requestid = 31118 sam_experiment = d0 sam_universe = prd station_name = sprace Job Submission > Samg submit file.jdf Eduardo Gregores VII DOSAR Workshop

MC Production on SAMGrid Eduardo Gregores VII DOSAR Workshop

Reprocessing and Refixing on SAMGrid Raw Data Reprocessing About 1 Billion events. 500 fb-1 of RunII data. 250 TB of raw data. All data taken up to 2005. Reprocessed on 12 locations distributed over 6 countries on 3 continents during 6 months. Gathered 3500 GHz equivalent of computer power. Completelly run on SAMGrid infrastructure. First use of Grid architecture to process real data on large scale. Data Refixing All data up to 09/2005. 1.5 Billion events refixed. Use of OSG and LHCG infrastucture by SAMGrid. First real world use case of large scale interoperability. Eduardo Gregores VII DOSAR Workshop

SAMGrid – LCG – OSG C S SAM-Grid Network Boundaries Forwarding Node FW SAM-Grid C S Expanding SAMGrid . Increase SAMGrid resources using LCG and OSG Keep SAMGrid data handling. Use of Forwarding Nodes The FW node is a usual gateway for SamGrid. Grid-to-Grid as evolution of Grid-to-Fabric interface. Job Management is implemented via an Grid specific “scheduler” handler. Uses some regional SAM station for data retrieval and storage. LCG uses station at CCIN2P3. US OSG sites uses OSG_OUHEP station. Status: DZero data has been refixed on LCG and OSG resources. MC is being produced on LCG. Joel may have some good news for MC production on OSG. Still lots of open questions concerning operation on external resources. Network Boundaries Forwarding Node LCG Cluster VO-Service (SAM) Job Flow Offers Service FW C S Eduardo Gregores VII DOSAR Workshop