Data Management with SAM at DØ The 2 nd International Workshop on HEP Data Grid Kyunpook National University Daegu, Korea August 22-23, 2003 Lee Lueking.

Slides:



Advertisements
Similar presentations
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
Advertisements

6/2/2015 Michael Diesburg HCP Distributed Computing at the Tevatron D0 Computing and Event Model Michael Diesburg, Fermilab For the D0 Collaboration.
Meta-Computing at DØ Igor Terekhov, for the DØ Experiment Fermilab, Computing Division, PPDG ACAT 2002 Moscow, Russia June 28, 2002.
F Fermilab Database Experience in Run II Fermilab Run II Database Requirements Online databases are maintained at each experiment and are critical for.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
Zhiling Chen (IPP-ETHZ) Doktorandenseminar June, 4 th, 2009.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
HEP Experiment Integration within GriPhyN/PPDG/iVDGL Rick Cavanaugh University of Florida DataTAG/WP4 Meeting 23 May, 2002.
CHEP 03 UCSDd0db.fnal.gov/sam1 DØ Data Handling Operational Experience CHEP03 UCSD March 24-28, 2003 Lee Lueking Fermilab Computing Division DØ overview.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
An Overview of PHENIX Computing Ju Hwan Kang (Yonsei Univ.) and Jysoo Lee (KISTI) International HEP DataGrid Workshop November 8 ~ 9, 2002 Kyungpook National.
28 April 2003Lee Lueking, PPDG Review1 BaBar and DØ Experiment Reports DOE Review of PPDG January 28-29, 2003 Lee Lueking Fermilab Computing Division D0.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
May 12-15, 2003Lee Lueking, EDG Int. Proj. Conf.1 DØ Computing Experience and Plans for SAM-Grid EU DataGrid Internal Project Conference May 12-15, 2003.
DØ Data Handling Operational Experience GridPP8 Sep 22-23, 2003 Rod Walker Imperial College London Computing Architecture Operational Statistics Challenges.
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
SAMGrid as a Stakeholder of FermiGrid Valeria Bartsch Computing Division Fermilab.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ScotGRID:The Scottish LHC Computing Centre Summary of the ScotGRID Project Summary of the ScotGRID Project Phase2 of the ScotGRID Project Phase2 of the.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
GridPP18 Glasgow Mar 07 DØ – SAMGrid Where’ve we come from, and where are we going? Evolution of a ‘long’ established plan Gavin Davies Imperial College.
DØ Computing Model & Monte Carlo & Data Reprocessing Gavin Davies Imperial College London DOSAR Workshop, Sao Paulo, September 2005.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
D C a c h e Michael Ernst Patrick Fuhrmann Tigran Mkrtchyan d C a c h e M. Ernst, P. Fuhrmann, T. Mkrtchyan Chep 2003 Chep2003 UCSD, California.
Manchester HEP Desktop/ Laptop 30 Desktop running RH Laptop Windows XP & RH OS X Home server AFS using openafs 3 DB servers Kerberos 4 we will move.
Workshop on Computing for Neutrino Experiments - Summary April 24, 2009 Lee Lueking, Heidi Schellman NOvA Collaboration Meeting.
16 September GridPP 5 th Collaboration Meeting D0&CDF SAM and The Grid Act I: Grid, Sam and Run II Rick St. Denis – Glasgow University Act II: Sam4CDF.
1 DØ Grid PP Plans – SAM, Grid, Ceiling Wax and Things Iain Bertram Lancaster University Monday 5 November 2001.
DØSAR a Regional Grid within DØ Jae Yu Univ. of Texas, Arlington THEGrid Workshop July 8 – 9, 2004 Univ. of Texas at Arlington.
What is SAM-Grid? Job Handling Data Handling Monitoring and Information.
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
SAM - Sequential Data Access via Metadata Schema Metadata Functionality Workshop Glasgow University April 26-28,2004.
Computing plans from UKDØ. Iain Bertram 8 November 2000.
Lee Lueking 1 The Sequential Access Model for Run II Data Management and Delivery Lee Lueking, Frank Nagy, Heidi Schellman, Igor Terekhov, Julie Trumbo,
GridPP11 Liverpool Sept04 SAMGrid GridPP11 Liverpool Sept 2004 Gavin Davies Imperial College London.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
EGEE-II INFSO-RI Enabling Grids for E-sciencE EGEE Site Architecture Resource Center Deployment Considerations MIMOS EGEE Tutorial.
DCAF (DeCentralized Analysis Farm) Korea CHEP Fermilab (CDF) KorCAF (DCAF in Korea) Kihyeon Cho (CHEP, KNU) (On the behalf of HEP Data Grid Working Group)
DCAF(DeCentralized Analysis Farm) for CDF experiments HAN DaeHee*, KWON Kihwan, OH Youngdo, CHO Kihyeon, KONG Dae Jung, KIM Minsuk, KIM Jieun, MIAN shabeer,
International Workshop on HEP Data Grid Aug 23, 2003, KNU Status of Data Storage, Network, Clustering in SKKU CDF group Intae Yu*, Joong Seok Chae Department.
Frank Wuerthwein, UCSD Update on D0 and CDF computing models and experience Frank Wuerthwein UCSD For CDF and DO collaborations October 2 nd, 2003 Many.
Analysis Tools at D0 PPDG Analysis Grid Computing Project, CS 11 Caltech Meeting Lee Lueking Femilab Computing Division December 19, 2002.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Adapting SAM for CDF Gabriele Garzoglio Fermilab/CD/CCF/MAP CHEP 2003.
April 25, 2006Parag Mhashilkar, Fermilab1 Resource Selection in OSG & SAM-On-The-Fly Parag Mhashilkar Fermi National Accelerator Laboratory Condor Week.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
CDF SAM Deployment Status Doug Benjamin Duke University (for the CDF Data Handling Group)
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
DØ Computing Model and Operational Status Gavin Davies Imperial College London Run II Computing Review, September 2005.
SAM: Past, Present, and Future Lee Lueking All Dzero Meeting November 2, 2001.
DØ Grid Computing Gavin Davies, Frédéric Villeneuve-Séguier Imperial College London On behalf of the DØ Collaboration and the SAMGrid team The 2007 Europhysics.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
SAM at CCIN2P3 configuration issues
DØ Computing & Analysis Model
DØ MC and Data Processing on the Grid
Lee Lueking D0RACE January 17, 2002
Presentation transcript:

Data Management with SAM at DØ The 2 nd International Workshop on HEP Data Grid Kyunpook National University Daegu, Korea August 22-23, 2003 Lee Lueking Fermilab, USA Computing Division DØ in brief SAM feature Overview SAM operation at DØ Summary Roadmap of Talk

22 Aug. 2003Lee Lueking, HEP Data Grid2 The DØ Experiment D0 Collaboration –18 Countries; 80 institutions –>600 Physicists Detector Data (Run 2a end mid ‘04) –1,000,000 Channels –Event size 250KB –Event rate 25 Hz avg. –Est. 2 year data totals (incl. Processing and analysis): 1 x 10 9 events, ~1.2 PB Monte Carlo Data (Run 2a) –6 remote processing centers –Estimate ~0.3 PB. Run 2b, starting 2005: >1PB/year Tevatron Chicago  pp p CDF DØ

SAM Features

22 Aug. 2003Lee Lueking, HEP Data Grid4 Managing Resources in SAM Data Resources (Storage + Network) Compute Resources (CPU + Memory) Local Batch SAM Station Servers Data and Compute Co-allocation SAM Global Optimizer SAM metadata Fair-share Resource allocation User groups Consumer(s) Project= DS on Station Dataset Definitions Datasets (DS) Batch scheduler SAM Meta-dataSAM servers Batch + SAM Replica Catalog Replica Management Service

22 Aug. 2003Lee Lueking, HEP Data Grid5 Simplified SAM Database Schema (SAM Metadata) Files ID Name Format Size # Events Files ID Name Format Size # Events Events ID Event Number Trigger L1 Trigger L2 Trigger L3 Off-line Filter Thumbnail Events ID Event Number Trigger L1 Trigger L2 Trigger L3 Off-line Filter Thumbnail Volume Project Data Tier Physical Data Stream Physical Data Stream Trigger Configuration Trigger Configuration Creation & Processing Info Creation & Processing Info Run Event-File Catalog Event-File Catalog Run Conditions Luminosity Calibration Trigger DB Alignment Run Conditions Luminosity Calibration Trigger DB Alignment Group and User information Group and User information Station Config. & Cache info Station Config. & Cache info File Storage Locations File Storage Locations MC Request & Info MC Request & Info SAM schema has over 100 tables There are several other related table spaces also available

22 Aug. 2003Lee Lueking, HEP Data Grid6 Monte Carlo Request System User defines required data in terms of a set of metadata keyword/values which define the physics details of the requested MC sample. This is then stored in SAM and when the request is processed, this physics data is extracted, and augmented with further 'processing mechanics' information and converted into executable jobs which are tailored to the resource they are executed on. The resulting data is stored in SAM with the physics metadata augmented by the details of the workflow and data provenance. Essentially it provides a metadata materialization service (a.k.a. virtual data system).

22 Aug. 2003Lee Lueking, HEP Data Grid7 SAM File Forwarding and Routing MSS SAM Station 1 SAM Station 2 SAM Station 3 SAM Station 4 Remote SAM Station Station Responsibilities Pre-stage files for consumers. Manage local cache Store files for producers File Forwarding File stores can be forwarded through other stations File Routing Routes for file transfers are configurable Extra-domain transfers use bbftp or GridFTP (parallel transfer protocols) Remote SAM Station Remote SAM Station

SAM at DØ d0db.fnal.gov/sam

22 Aug. 2003Lee Lueking, HEP Data Grid9 Overview of DØ Data Handling Registered Users600 Number of SAM Stations56 Registered Nodes900 Total Disk Cache40 TB Number Files - physical1.5M Number Files - virtual0.7M Robotic Tape Storage400 TB Regional Center Analysis site Summary of DØ Data HandlingIntegrated Files Consumed vs Month (DØ) Integrated GB Consumed vs Month (DØ) 4.0 M Files Consumed 1.2 PB Consumed Mar2002 Mar2003

22 Aug. 2003Lee Lueking, HEP Data Grid10 Great Britain 200 All Monte Carlo Production Netherlands 50 France 100 Texas 64Czech R. 32 fnal.gov DØ Data Flows UNIX hosts ENSTORE movers LINUX farm 300+ dual PIII/IV nodes Startap Chicago switch a: production c: development ADIC AML/2 STK 9310 powderhorn ClueDØ Linux desktop user cluster 227 nodes Fiber to experiment switch DEC4000 d0ola,b,c L3 nodes RIP data logger collector/router a b c SUN 4500 Linux quad d0ora1 d0lxac1 Linux d0dbsrv1 switch SGI Origin R12000 processors 27 TB fiber channel disks Central Analysis Backend (CAB) 160 dual 2GHz Linux nodes 35 GB cache ea. Experimental Hall/office complex CISCO Datalogger Worldwide Analysis

22 Aug. 2003Lee Lueking, HEP Data Grid11 DØ SAM Station Summary NameLocationNodes/cpuCacheUse/comments Central- analysis FNAL128 SMP*, SGI Origin TBAnalysis & D0 code development CAB (CA Backend) FNAL16 dual 1 GHz dual 1.8 GHz 6.2 TBAnalysis and general purpose FNAL-FarmFNAL100 dual GHz +240 dual 1.8 GHz 3.2 TBReconstruction CLueD0FNAL50 mixed PIII, AMD. (may grow >200) 2 TBUser desktop, General analysis D0karlsruhe (GridKa) Karlsruhe, Germany 1 dual 1.3 GHz gateway, >160 dual PIII & Xeon 3 TB NFS shared General/Workers on PN. Shared facility D0umich (NPACI) U Mich. Ann Arbor 1 dual 1.8 GHz gateway, 100 x dual AMD XP TB NFS shared Re-reconstruction. workers on PN. Shared facility Many Others > 4 dozen WorldwideMostly dual PIII, Xeon, and AMD XP MC production, gen. analysis, testing *IRIX, all others are Linux

22 Aug. 2003Lee Lueking, HEP Data Grid12 Station Stats: GB Consumed (by jobs) Daily Feb 14 – Mar 15 Central-Analysis FNAL-farm ClueD0 CAB 2.5 TB Feb GB Feb TB Mar 6 >1.6 TB Feb 28

22 Aug. 2003Lee Lueking, HEP Data Grid13 Station Stats: MB Delivered/Sent Daily Feb 14 – March 15 Central-Analysis FNAL-farm ClueD0 CAB Delivered to Sent from 1 TB Feb GB Feb TB Mar GB Feb TB Feb 22 Consumed 270 GB Feb 17 Consumed 1.1 TB Mar 6 Consumed 1.6 TB Feb 28 Consumed

22 Aug. 2003Lee Lueking, HEP Data Grid14 Challenges (1) Getting SAM to meet the needs of DØ in the many configurations is and has been an enormous challenge. –Automating Monte Carlo Production and Cataloging with MC request system in conjunction with MC RunJob meta system. –File corruption issues. Solved with CRC. –Preemptive distributed caching is prone to race conditions and log jams. These have been solved. –Private networks sometimes require “border” naming services. This is understood. –NFS shared cache configuration provides additional simplicity and generality, at the price of scalability (star configuration). This works. –Global routing completed.

22 Aug. 2003Lee Lueking, HEP Data Grid15 Challenges (2) –Convenient interface for users to build their own applications. SAM user api is provided for python. –Installation procedures for the station servers have been quite complex. They are improving and we plan to soon have “push button” and even “opportunistic deployment” installs. –Lots of details with opening ports on firewalls, OS configurations, registration of new hardware, and so on. –Username clashing issues. Moving to GSI and Grid Certificates. –Interoperability with many MSS. –Network attached files. Consumer is given file URL and data is delivered to consumer over the network via RFIO, dCap, etc.

22 Aug. 2003Lee Lueking, HEP Data Grid16 Summary SAM is a well-hardened, multi-featured, distributed Data Management and Delivery system. The DØ Experiment has many challenging data management needs which are being met by SAM on a worldwide scale. (CDF is also using SAM). Many complex issues have been solved to provide the needed level of service to the experiment. Now, on to SAMGrid…

Thank You