Fermilab June 29, 2001Data collection and handling for HEP1 Matthias Kasemann Fermilab Overview of Data collection and handling for High Energy Physics.

Slides:



Advertisements
Similar presentations
1 AMY Detector (eighties) A rather compact detector.
Advertisements

First results from the ATLAS experiment at the LHC
The Biggest Experiment in History. Well, a tiny piece of it at least… And a glimpse 12bn years back in time To the edge of the observable universe So.
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
HLT - data compression vs event rejection. Assumptions Need for an online rudimentary event reconstruction for monitoring Detector readout rate (i.e.
Randall Sobie The ATLAS Experiment Randall Sobie Institute for Particle Physics University of Victoria Large Hadron Collider (LHC) at CERN Laboratory ATLAS.
Grids: Why and How (you might use them) J. Templon, NIKHEF VLV T Workshop NIKHEF 06 October 2003.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
1 Andrew Hanushevsky - HEPiX, October 6-8, 1999 Mass Storage For BaBar at SLAC Andrew Hanushevsky Stanford.
DATA PRESERVATION IN ALICE FEDERICO CARMINATI. MOTIVATION ALICE is a 150 M CHF investment by a large scientific community The ALICE data is unique and.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
LHC’s Second Run Hyunseok Lee 1. 2 ■ Discovery of the Higgs particle.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
25 February 2000Tim Adye1 Using an Object Oriented Database to Store BaBar's Terabytes Tim Adye Particle Physics Department Rutherford Appleton Laboratory.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
Instrumentation of the SAM-Grid Gabriele Garzoglio CSC 426 Research Proposal.
Fermilab Root2001 June 15, 2001Root2001 Workshop, FNAL1 An incomplete view of the future of HEP Computing Matthias Kasemann Fermilab A view of the future.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
P5 and the HEP Program A. Seiden Fermilab June 2, 2003.
European Organization for Nuclear Research Organisation Européenne pour la Recherche Nucléaire High-Energy Physics Data Delivering Data in Science ICSTI.
Fermi National Accelerator Laboratory SC2006 Fermilab Data Movement & Storage Multi-Petabyte tertiary automated tape store for world- wide HEP and other.
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Collider Detector at Fermilab Sung-hyun chang High Energy Physics lab. KNU.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
The LHCb CERN R. Graciani (U. de Barcelona, Spain) for the LHCb Collaboration International ICFA Workshop on Digital Divide Mexico City, October.
Les Les Robertson LCG Project Leader High Energy Physics using a worldwide computing grid Torino December 2005.
DESY Photon Science XFEL official start of project: 5 June 2007 FLASH upgrade to 1 GeV done, cool down started PETRA III construction started 2 July 2007.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
FSU Experimental HEP Faculty Todd Adams Susan Blessing Harvey Goldman S Sharon Hagopian Vasken Hagopian Kurtis Johnson Harrison Prosper Horst Wahl.
F Detector and Computing Operations Hugh Montgomery DOE Tevatron Operations Review March 27, 2007.
Searching for New Matter with the D0 Experiment Todd Adams Department of Physics Florida State University September 19, 2004.
Computing for LHC Physics 7th March 2014 International Women's Day - CERN- GOOGLE Networking Event Maria Alandes Pradillo CERN IT Department.
LHC Computing, CERN, & Federated Identities
Computing Division FY03 Budget and budget outlook for FY04 + CDF International Finance Committee April 4, 2003 Vicky White Head, Computing Division.
John Womersley 1/13 Fermilab’s Future John Womersley Fermilab May 2004.
Physics Results from CDF and Prospects for a FY 2011 Run Kevin Pitts / University of Illinois DOE S&T Review of Scientific User Facilities June 30 – July.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
LHCbComputing Computing for the LHCb Upgrade. 2 LHCb Upgrade: goal and timescale m LHCb upgrade will be operational after LS2 (~2020) m Increase significantly.
1 Experimental Particle Physics PHYS6011 Fergus Wilson, RAL 1.Introduction & Accelerators 2.Particle Interactions and Detectors (2) 3.Collider Experiments.
Hans Wenzel CDF CAF meeting October 18 th -19 th CMS Computing at FNAL Hans Wenzel Fermilab  Introduction  CMS: What's on the floor, How we got.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
Nigel Lockyer Fermilab Operations Review 16 th -18 th May 2016 Fermilab in the Context of the DOE Mission.
1 Particle Physics Data Grid (PPDG) project Les Cottrell – SLAC Presented at the NGI workshop, Berkeley, 7/21/99.
Grid technologies for large-scale projects N. S. Astakhov, A. S. Baginyan, S. D. Belov, A. G. Dolbilov, A. O. Golunov, I. N. Gorbunov, N. I. Gromova, I.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
CrossGrid Workshop, Kraków, 5 – 6 Nov-2001 Distributed Data Analysis in HEP Piotr MALECKI Institute of Nuclear Physics Kawiory 26A, Kraków, Poland.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
Hall D Computing Facilities Ian Bird 16 March 2001.
Particle Physics Sector Young-Kee Kim / Greg Bock Leadership Team Strategic Planning Winter Workshop January 29, 2013.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
evoluzione modello per Run3 LHC
ALICE Computing Upgrade Predrag Buncic
ILD Ichinoseki Meeting
High Energy Physics at UTA
Nuclear Physics Data Management Needs Bruce G. Gibbard
High Energy Physics at UTA
Particle Physics Theory
Using an Object Oriented Database to Store BaBar's Terabytes
Experimental Particle Physics PHYS6011 Putting it all together Lecture 4 28th April 2008 Fergus Wilson. RAL.
Development of LHCb Computing Model F Harris
Presentation transcript:

Fermilab June 29, 2001Data collection and handling for HEP1 Matthias Kasemann Fermilab Overview of Data collection and handling for High Energy Physics Experiments

Fermilab June 29, 2001Data collection and handling for HEP2 Fermilab Mission Statement (see Web) Fermi National Accelerator Laboratory advances the understanding of the fundamental nature of matter and energy by providing leadership and resources for qualified researchers to conduct basic research at the frontiers of high energy physics and related disciplines. Fermilab operates the world's highest-energy particle accelerator, the Tev-atron. More than 2,200 scientists from 36 states and 20 countries use Fermi-lab's facilities to carry out research at the frontiers of particle physics.

Fermilab June 29, 2001Data collection and handling for HEP3 Fermilab Community: Collaborations u Fermilab is an open site (no fences) and acts as a host to the many universities and institutions pursuing research here. u Given that, the culture of the lab is very university-like, which is one of its big strengths for scientific research. u Collaborations: u 2,716 Physicists work at Fermilab u 224 institutions from: u 38 states (1,703 physicists) u 23 foreign countries (1,014 physicists) u 555 graduate students u (probably a similar number of postdocs) u It is interesting to note that only 10% of CDF and D0 physicists work for Fermilab

Fermilab June 29, 2001Data collection and handling for HEP4 Collaborations u Detectors are designed and built by large collaborations of scientists and technicians u Many tens of institutions (mainly universities) u Many hundreds of people u Many countries u Important Run2 Milestone: u CDF and D0 (  =800+ scientists) started data taking 03/01/2001 after 4 years of preparation u Unique scientific opportunity to make major HEP discovery – don’t miss it!!

Fermilab June 29, 2001Data collection and handling for HEP5 D0 Experiment CDF Experiment Computing Center

Fermilab June 29, 2001Data collection and handling for HEP6 From Physics to Raw Data: what happens in a detector Fragmentation, Decay Interaction with detector material Multiple scattering, interactions Raw data (Bytes) Read-out addresses, ADC, TDC values, Bit patterns e+e+ e-e- f f Z0Z0 _ Detector response Noise, pile-up, cross-talk, inefficiency, ambiguity, resolution, response function, alignment, temperature Theoretical Model of Particle interaction Particle production and decays observed in detectors are Quantum Mechanical processes. Hundreds or thousands of different production- and decay-channels possible, all with different probabilities. In the end all we measure are probabilities!! 250Kb – 1 Mb

Fermilab June 29, 2001Data collection and handling for HEP7 Data reduction and recording: here CMS in 2006 On-line System Multi-level trigger Filter out background Reduce data volume 24 x 7 operation Level 1 - Special Hardware Level 2 - Embedded Processors 40 MHz (1000 TB/sec) Level 3 – Farm of commodity CPUs 75 KHz (75 GB/sec) 5 KHz (5 GB/sec) 100 Hz (100 MB/sec) Data Recording & Offline Analysis protons anti- protons

Fermilab June 29, 2001Data collection and handling for HEP8 From Raw Data to Physics: what happens during analysis Fragmentation, Decay Physics analysis Interaction with detector material Pattern, recognition, Particle identification Detector response apply calibration, alignment, Raw data Convert to physics quantities Reconstruction Simulation (Monte-Carlo) Analysis _ e+e+ e-e- f f Z0Z0 Basic physics Results 250Kb – 1 Mb100 Kb25 Kb5 Kb500 b

Fermilab June 29, 2001Data collection and handling for HEP9 Data flow from detector to analysis MBps 20 MBps 100 Mbps 400 MBps Experiment “reconstruction” “analysis CPU” “analysis disks” Permanent storage

Fermilab June 29, 2001Data collection and handling for HEP10 u Mass storage: robotics, tape drives + interface computing. u Production farms u Analysis computers: support for many users for high statistics analysis (single system image, multi-CPU). u Disk storage: permanent storage for frequently accessed data, staging pool for data stored on tape. u Miscellaneous: networking, infrastructure,... (Total for both CDF & D0 experiments) Run IIa Equipment Spending Profile

Fermilab June 29, 2001Data collection and handling for HEP11 RUN IIa Equipment u Analysis servers u Disk storage u Robots with tape drives

Fermilab June 29, 2001Data collection and handling for HEP12 Fermilab HEP Program Year: Neutrinos: MiniBooNE NuMI/Minos Run IIa Collider: Run IIb BTeV MI Fixed Target: Testbeam KaMI/CKM? Astrophysics: Sloan CDMS Auger LHC physics

Fermilab June 29, 2001Data collection and handling for HEP13 Run 2 Data Volumes u First Run 2b costs estimates based on scaling arguments u Use predicted luminosity profile u Assume technology advance (Moore’s law) u CPU and data storage requirements both scale with data volume stored u Data volume depends on physics selection in trigger u Can vary between 1 – 8 PB (Run 2a: 1 PB) per experiment u Have to start preparation by 2002/2003

Fermilab June 29, 2001Data collection and handling for HEP14 D0: Data Volume collected...in the past 12 Months Events collected [M] Data Volume [GB]...in the past 12 Months Fermilab Stations uCentral Analysis uOnline uFarm uLinux analysis stations (3) Remote Stations uLyon (France) uAmsterdam (Netherlands) uLancaster (UK) uPrague (Czech R.) uMichigan State uU. T. Arlington

Fermilab June 29, 2001Data collection and handling for HEP15 Data Volume per experiment per year (in units of Gbytes) Data Volume doubles every 2.4 years

Fermilab June 29, 2001Data collection and handling for HEP16 Improving our knowledge: better experiments in HEP Desired Improvement Higher energy More collisions Better detectors More events Better analysis Simulation Theory Computing Technique Accelerator Design/simulation Acc. Design and controls Triggers (networks, CPU), simulation Disk, tape, CPU, networks Disk, tape, CPU, networks, algorithms CPU, algorithms, OO

Fermilab June 29, 2001Data collection and handling for HEP17 How long are data scientifically interesting? “Lifetime of data” u 1. Month after recording: u Verification of data integrity u Verification of detector performance and integrity u months after recording: u Collect more data u Process and reconstruct “interpret the bits” u Perform data analysis u Compare to simulated data u Publish!! u >2 years after recording: u Data often superseded by more precise experiments u Combine results for high statistics measurements and publish!! u Archive for comparison and possible re-analysis u >5 years after recording: u Decide on long-term storage for re-analysis

Fermilab June 29, 2001Data collection and handling for HEP18 Tape storage history: the last 10 years u 1990: “New tape retention policy” u Maximize accessibility of tapes actively used u Provide off-site storage for data which have finite probability of being needed in the future u Default retention period in FCC set to 5 years, extended on justified request u Disposition of redundant and obsolete tapes decided together with experiment spokespeople (based on scientific value) u 1992: “Too many tapes, need room to store new data” u Tapes not accessed within years moved to off-site storage u Tapes retrievable within few working days u 1998: New “Fermilab Tapes Purchasing Policy” u Tapes intended for “long term” storage are purchased and owned by FNAL u Tapes cannot be removed from FCC/storage by experimenters + users u 1999 – 2000: remove 9-track round tapes from FCC archive u Data needed in the future moved to off-site storage u Disposition of redundant and obsolete tapes decided together with experiment spokespeople (based on scientific value)

Fermilab June 29, 2001Data collection and handling for HEP19 Relevant questions for tape storage: u All questions wrt. Storage or Disposal are discussed and procedures established in concurrence with DOE records manager in FAO/CH u In order to satisfy DOE requirements for disposal we request written approval from spokesperson: u What is stored on the individual tape? (raw data, summary data, etc) u When were the tapes written? u Do subsequent summary tapes exist (are tapes redundant)? u Do you foresee future research needs for these tapes? u Are software and computers still available for reanalysis or is it feasible to write new software to do so? u Do you have any reason to retain the tapes? Explain…