Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Amber Boehnlein, FNAL D0 Computing Model and Plans Amber Boehnlein D0 Financial Committee November 18, 2002.
6/2/2015 Michael Diesburg HCP Distributed Computing at the Tevatron D0 Computing and Event Model Michael Diesburg, Fermilab For the D0 Collaboration.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
L3 Filtering: status and plans D  Computing Review Meeting: 9 th May 2002 Terry Wyatt, on behalf of the L3 Algorithms group. For more details of current.
The D0 Monte Carlo Challenge Gregory E. Graham University of Maryland (for the D0 Collaboration) February 8, 2000 CHEP 2000.
High Energy Physics At OSCER A User Perspective OU Supercomputing Symposium 2003 Joel Snow, Langston U.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
The SAMGrid Data Handling System Outline:  What Is SAMGrid?  Use Cases for SAMGrid in Run II Experiments  Current Operational Load  Stress Testing.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
November 7, 2001Dutch Datagrid SARA 1 DØ Monte Carlo Challenge A HEP Application.
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
D0 SAM – status and needs Plagarized from: D0 Experiment SAM Project Fermilab Computing Division.
3rd June 2004 CDF Grid SAM:Metadata and Middleware Components Mòrag Burgon-Lyon University of Glasgow.
Snapshot of the D0 Computing and Operations Planning Process Amber Boehnlein For the D0 Computing Planning Board.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
A Design for KCAF for CDF Experiment Kihyeon Cho (CHEP, Kyungpook National University) and Jysoo Lee (KISTI, Supercomputing Center) The International Workshop.
Jan. 17, 2002DØRAM Proposal DØRACE Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Remote Analysis Station ArchitectureRemote.
DØ RAC Working Group Report Progress Definition of an RAC Services provided by an RAC Requirements of RAC Pilot RAC program Open Issues DØRACE Meeting.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
ALICE Upgrade for Run3: Computing HL-LHC Trigger, Online and Offline Computing Working Group Topical Workshop Sep 5 th 2014.
DØ Computing Model & Monte Carlo & Data Reprocessing Gavin Davies Imperial College London DOSAR Workshop, Sao Paulo, September 2005.
Stephen Wolbers CHEP2000 February 7-11, 2000 Stephen Wolbers CHEP2000 February 7-11, 2000 CDF Farms Group: Jaroslav Antos, Antonio Chan, Paoti Chang, Yen-Chu.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
International Workshop on HEP Data Grid Nov 9, 2002, KNU Data Storage, Network, Handling, and Clustering in CDF Korea group Intae Yu*, Junghyun Kim, Ilsung.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Status of UTA IAC + RAC Jae Yu 3 rd DØSAR Workshop Apr. 7 – 9, 2004 Louisiana Tech. University.
9 February 2000CHEP2000 Paper 3681 CDF Data Handling: Resource Management and Tests E.Buckley-Geer, S.Lammel, F.Ratnikov, T.Watts Hardware and Resources.
KLOE Computing Update Paolo Santangelo INFN LNF KLOE General Meeting University of Rome 2, Tor Vergata 2002, December
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Data reprocessing for DZero on the SAM-Grid Gabriele Garzoglio for the SAM-Grid Team Fermilab, Computing Division.
DØ Data Handling & Access The DØ Meta-Data Browser Pushpa Bhat Fermilab June 4, 2001.
Computing plans from UKDØ. Iain Bertram 8 November 2000.
Lee Lueking 1 The Sequential Access Model for Run II Data Management and Delivery Lee Lueking, Frank Nagy, Heidi Schellman, Igor Terekhov, Julie Trumbo,
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Integration of the ATLAS Tag Database with Data Management and Analysis Components Caitriana Nicholson University of Glasgow 3 rd September 2007 CHEP,
1 EIR Nov 4-8, 2002 DAQ and Online WBS 1.3 S. Fuess, Fermilab P. Slattery, U. of Rochester.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
DØ Online Workshop3-June-1999S. Fuess Online Computing Overview DØ Online Workshop 3-June-1999 Stu Fuess.
Feb. 14, 2002DØRAM Proposal DØ IB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) Introduction Partial Workshop Results DØRAM Architecture.
Status report of the KLOE offline G. Venanzoni – LNF LNF Scientific Committee Frascati, 9 November 2004.
Collaboration meeting, October 11, 2002Jianming Qian, University of Michigan Computing & Software DØ main executables Data and MC production Data-handling.
Frank Wuerthwein, UCSD Update on D0 and CDF computing models and experience Frank Wuerthwein UCSD For CDF and DO collaborations October 2 nd, 2003 Many.
Remote Institute Tasks Frank Filthaut 11 February 2002  Monte Carlo production  Algorithm development  Alignment, calibration  Data analysis  Data.
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
Data Management with SAM at DØ The 2 nd International Workshop on HEP Data Grid Kyunpook National University Daegu, Korea August 22-23, 2003 Lee Lueking.
Feb. 13, 2002DØRAM Proposal DØCPB Meeting, Jae Yu 1 Proposal for a DØ Remote Analysis Model (DØRAM) IntroductionIntroduction Partial Workshop ResultsPartial.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
Status Report on Data Reconstruction May 2002 C.Bloise Results of the study of the reconstructed events in year 2001 Data Reprocessing in Y2002 DST production.
DØ Computing Model and Operational Status Gavin Davies Imperial College London Run II Computing Review, September 2005.
1 P. Murat, Mini-review of the CDF Computing Plan 2006, 2005/10/18 An Update to the CDF Offline Plan and FY2006 Budget ● Outline: – CDF computing model.
ATLAS – statements of interest (1) A degree of hierarchy between the different computing facilities, with distinct roles at each level –Event filter Online.
DØ Computing & Analysis Model
ALICE Computing Upgrade Predrag Buncic
Lee Lueking D0RACE January 17, 2002
DØ RAC Working Group Report
Proposal for a DØ Remote Analysis Model (DØRAM)
Development of LHCb Computing Model F Harris
Presentation transcript:

Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002

Jianming Qian, UM/DØ Organization

Jianming Qian, UM/DØ Road to Glories Reconstruction (RECO) detector algorithms particle identifications production farm data format … pedestals gains, linearity … Calibration detector trigger system data acquisition … Raw Data:Monte Carlo event generation detector simulation … Physics Analysis event selections efficiencies & backgrounds … Databases Network Releases Operation Data handling & access Trigger simulations passes through software and computing …

Jianming Qian, UM/DØ Where We Are Now All building blocks are in place … PC/Linux based Level-3 is taking a 500 Hz input and the DAQ is is capable of writing at ~50 Hz. The offline reconstruction program (DØRECO) has all the basic reconstruction functionalities, is deployed at all farms, and writes DØRECO output in a temporary format. Data processing is done at Fermilab. The central farm is presently capable of processing event at ~12 Hz DC. Monte Carlo needs of physics analyses are met by remote farms (Czech, France, Netherlands, UK, USA, UK) Basic infrastructures and functionalities of data handling and access are in place. Data are stored in the Enstore system via SAM. Analysis CPUs are provided by an SGI O2000 system (dømino) and Linux desktops (ClueDØ, …). … we are producing first physics results!

Jianming Qian, UM/DØ Physics Distributions - 1

Jianming Qian, UM/DØ Physics Distributions - 2

Jianming Qian, UM/DØ Level-3 and Reconstruction Level-3 Status: Basic functionalities: event building, filtering through partial reconstruction, streaming, monitoring in a time budget of 0.1 sec. Current Farm: 48 1 GHz P-III processors and capable of providing a factor of ??? rejection Implemented filters: Calorimeter, muon and ??? Based filters Output rate: ?? Hz peak, ?? Hz average DØRECO Status: Basic functionalities: reconstructing all physics object candidates, produce DST and Thumbnails (TMB). The current DØRECO has all reconstruction functionalities, but does not yet produce DST or TMB. Current version takes 20 CPU seconds per event on a 500 MHz benchmark processor. New version takes longer. Algorithm speed is a concern. Sec/event luminosity On 500 MHz processor 396 ns 132 ns

Jianming Qian, UM/DØ Global Production System Fermilab Farm: Data Processing An SGI O2000 as the I/O server, used to buffer all output back to central storage facility Two Gb and one 100 Mb Ethernet interfaces providing inside and outside connections 122 dual processor Intel systems as worker nodes with a total of THz CPU power 128 dual processor 2.0 GHz nodes are on order, this will bring the CPU capacity to 0.35 THz Remote Farms: Monte Carlo Production Remote Monte Carlo farms have been meeting our physics needs. Current sites: Boston, CCIN2P3, Lancaster, NIKHEF, Prague, UTA. All sites are running full GEANT simulation Potential future sites: Univ. of Michigan, Manchester, Oklahoma, UCD Ireland, Karlsruhe

Jianming Qian, UM/DØ Data Handling & Databases SAM (Sequential Access to data via Meta-data) is used to catalog data files produced by the experiment and provides distributed data storage and access to production and analysis centers. Data storage and access are provided through SAM stations deployed at production and analysis centers. The central database contains all metadata for each file registered with the system, as well as station configuration and operational information. The raw data from DAQ, data RECO outputs from the central production farm and MC RECO outputs from remote production farms are stored at the Fermilab Enstore facility via SAM. Both ADIC robot with LTO drives and STK robot with 9940 drives are in use. The system has handled a peak volume of > 2 TB a day so far. An 8 processor SUN 4500 provides access to Oracle databases for both onsite and offsite needs. The implementation of DB servers bypasses many offsite license issues. The current system works well and needs to be scaled, evolved and improved to meet Run IIa and IIb needs

Jianming Qian, UM/DØ Computing System Now Central Data Storage SGI 2000 … Fermilab Farm Remote Farm DØDØ Linux Desktops (ClueDØ) dØmino

Jianming Qian, UM/DØ Where We Want to Go Level-3 capable of taking 1 kHz input and DAQ capable of writing out ~50 Hz for Run IIa and ~100 Hz for Run IIb, resulting a DC rate of 25 Hz for Run IIa and 50 Hz for Run IIb. A DØRECO running at a reasonable speed (50 sec/event on a 500 MHz processor) writing out reconstructed information in DST (~150 kB/event) and TMB (~10 kB/event) format. A central Fermilab farm capable of keeping up with the Online and providing some reprocessing power. A global Monte Carlo production systems capable of running at ½ of the data rate and possibly providing secondary data processing. An analysis model of central and regional centers with sufficient CPU and disk resources to meet analysis needs. A reliable and fast data handling and access system capable of providing full TMB and partial DST datasets to all analysis centers. An automatic and tiered software release system serving all production and analysis centers as well as individual institutions. …

Jianming Qian, UM/DØ Computing Architecture Central Data Storage dØmino … …… Central Analysis System Central Farm Remote Farm Remote Analysis Linux Desktops (ClueDØ) ClueDØ Back End (CluB) DØ

Jianming Qian, UM/DØ Data Tier DST (Data Summary Tape): all high level physics objects, some detector-level information to allow calibration and limited re-reconstruction. 100% on tape, partial set on disk. TMB (Thumbnail): all high level physics objects, good for most physics analyses, 100% on tape, 100% on disk at central and regional centers. Derived Datasets: Physics/ID groups or their subgroups may create their derived datasets from either DST or TMB in their chosen format and are responsible for maintaining these datasets. DØRECO DST TMB Raw Data 300 kb/event 100 kb/evt 10 kb/evt … Derived Datasets

Jianming Qian, UM/DØ Primary Data Path Data Handling System Remote CentersCentral SystemClueD0/CluBRegional FarmCentral Farm Robotic Storage Raw Data RECO Data RECO MC

Jianming Qian, UM/DØ Analysis Model DST stripping: Resource intensive, limited to physics, ID, and detector groups, done at CAS TMB based: Expect to be done mostly by subgroups at RACs Derived datasets: Individuals at their desktops dØmino … CAS Analysis CPUs to be provided by: Central Analysis System (CAS) at FCC: A PC/Linux dØmino back-end supplied and administrated by the computing division ClueDØ/CluB at DØ: Linux PCs contributed mostly by institutions for interactive and batch jobs Regional Analysis Centers (RAC): Institutions with CPU, disk and personnel resources to serve collaborators … … ClueDØ CluB (RAC) … RAC (Regional Analysis Center) Layered Analysis Approach:

Jianming Qian, UM/DØ Cost Assumptions Event Rates: Run IIa (2003 – 2004): 25 Hz DC Run IIb (2005 – 2008): 50 Hz DC Monte Carlo: ½ of data rate ¼ Geant and ¾ fast simulations Event Size: Raw Data: 300 kB DST Data: 150 kB TMB Data: 10 kB TMB MC: 20 kB Farm Processing: Data Run Iia:40 a 500 MHz PC Data Run Iib: 50 sec/event Monte Carlo:85 sec/event (average over Geant and fast simulations) Farm efficiency: 70% These estimates are our ‘first’ attempt, Will improve as we learn more…

Jianming Qian, UM/DØ Summary All building blocks are in place. Have met most analysis needs so far. A road map to high luminosity and large dataset. Many systems scale. Many issues remain: Full Level-3 functionality Online and offline streaming DØRECO output DST and TMB content and format DØRECO quality and efficiency DØRECO speed at high luminosity Expand Monte Carlo production capacity Central and regional analysis centers, analysis CPU needs Data handling and access for large dataset A reasonable event display graphics … We are on our way…