ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.

Slides:



Advertisements
Similar presentations
Beowulf Supercomputer System Lee, Jung won CS843.
Advertisements

CHEP 2012 – New York City 1.  LHC Delivers bunch crossing at 40MHz  LHCb reduces the rate with a two level trigger system: ◦ First Level (L0) – Hardware.
23/04/2008VLVnT08, Toulon, FR, April 2008, M. Stavrianakou, NESTOR-NOA 1 First thoughts for KM3Net on-shore data storage and distribution Facilities VLV.
Title US-CMS User Facilities Vivian O’Dell US CMS Physics Meeting May 18, 2001.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
Hall D Online Data Acquisition CEBAF provides us with a tremendous scientific opportunity for understanding one of the fundamental forces of nature. 75.
CERN - European Laboratory for Particle Physics HEP Computer Farms Frédéric Hemmer CERN Information Technology Division Physics Data processing Group.
Shuei MEG review meeting, 2 July MEG Software Status MEG Software Group Framework Large Prototype software updates Database ROME Monte Carlo.
Test Of Distributed Data Quality Monitoring Of CMS Tracker Dataset H->ZZ->2e2mu with PileUp - 10,000 events ( ~ 50,000 hits for events) The monitoring.
GLAST LAT ProjectDOE/NASA Baseline-Preliminary Design Review, January 8, 2002 K.Young 1 LAT Data Processing Facility Automatically process Level 0 data.
A TCP/IP transport layer for the DAQ of the CMS Experiment Miklos Kozlovszky for the CMS TriDAS collaboration CERN European Organization for Nuclear Research.
Remote Production and Regional Analysis Centers Iain Bertram 24 May 2002 Draft 1 Lancaster University.
Boosting Event Building Performance Using Infiniband FDR for CMS Upgrade Andrew Forrest – CERN (PH/CMD) Technology and Instrumentation in Particle Physics.
CDF data production models 1 Data production models for the CDF experiment S. Hou for the CDF data production team.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina, L.Lueking,
Farm Management D. Andreotti 1), A. Crescente 2), A. Dorigo 2), F. Galeazzi 2), M. Marzolla 3), M. Morandin 2), F.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Copyright © 2000 OPNET Technologies, Inc. Title – 1 Distributed Trigger System for the LHC experiments Krzysztof Korcyl ATLAS experiment laboratory H.
MiniBooNE Computing Description: Support MiniBooNE online and offline computing by coordinating the use of, and occasionally managing, CD resources. Participants:
ILC Trigger & DAQ Issues - 1 ILC DAQ issues ILC DAQ issues By P. Le Dû
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
The Computing System for the Belle Experiment Ichiro Adachi KEK representing the Belle DST/MC production group CHEP03, La Jolla, California, USA March.
Stephen Wolbers CHEP2000 February 7-11, 2000 Stephen Wolbers CHEP2000 February 7-11, 2000 CDF Farms Group: Jaroslav Antos, Antonio Chan, Paoti Chang, Yen-Chu.
CDF Offline Production Farms Stephen Wolbers for the CDF Production Farms Group May 30, 2001.
21 st October 2002BaBar Computing – Stephen J. Gowdy 1 Of 25 BaBar Computing Stephen J. Gowdy BaBar Computing Coordinator SLAC 21 st October 2002 Second.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
Tier-2  Data Analysis  MC simulation  Import data from Tier-1 and export MC data CMS GRID COMPUTING AT THE SPANISH TIER-1 AND TIER-2 SITES P. Garcia-Abia.
October 2002 INFN Catania 1 The (LHCC) Grid Project Initiative in Prague Dagmar Adamova INP Rez near Prague.
Virtualization for the LHCb Online system CHEP Taipei Dedicato a Zio Renato Enrico Bonaccorsi, (CERN)
EGEE is a project funded by the European Union under contract IST HEP Use Cases for Grid Computing J. A. Templon Undecided (NIKHEF) Grid Tutorial,
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
Clara Gaspar, March 2005 LHCb Online & the Conditions DB.
3rd April 2001A.Polini and C.Youngman1 GTT status Items reviewed: –Results of GTT tests with 3 MVD-ADC crates. Aims Hardware and software setup used Credit.
The KLOE computing environment Nuclear Science Symposium Portland, Oregon, USA 20 October 2003 M. Moulson – INFN/Frascati for the KLOE Collaboration.
LHCb DAQ system LHCb SFC review Nov. 26 th 2004 Niko Neufeld, CERN.
PC clusters in KEK A.Manabe KEK(Japan). 22 May '01LSCC WS '012 PC clusters in KEK s Belle (in KEKB) PC clusters s Neutron Shielding Simulation cluster.
IPHC - DRS Gilles CLAUS 04/04/20061/20 EUDET JRA1 Meeting, April 2006 MAPS Test & DAQ Strasbourg OUTLINE Summary of MimoStar 2 Workshop CCMOS DAQ Status.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
June 17th, 2002Gustaaf Brooijmans - All Experimenter's Meeting 1 DØ DAQ Status June 17th, 2002 S. Snyder (BNL), D. Chapin, M. Clements, D. Cutts, S. Mattingly.
Online Monitoring for the CDF Run II Experiment T.Arisawa, D.Hirschbuehl, K.Ikado, K.Maeshima, H.Stadie, G.Veramendi, W.Wagner, H.Wenzel, M.Worcester MAR.
UTA MC Production Farm & Grid Computing Activities Jae Yu UT Arlington DØRACE Workshop Feb. 12, 2002 UTA DØMC Farm MCFARM Job control and packaging software.
DØ Online Workshop3-June-1999S. Fuess Online Computing Overview DØ Online Workshop 3-June-1999 Stu Fuess.
Predrag Buncic Future IT challenges for ALICE Technical Workshop November 6, 2015.
COMP381 by M. Hamdi 1 Clusters: Networks of WS/PC.
The 2001 Tier-1 prototype for LHCb-Italy Vincenzo Vagnoni Genève, November 2000.
Week1: Introduction to Computer Networks. Copyright © 2012 Cengage Learning. All rights reserved.2 Objectives 2 Describe basic computer components and.
1 Electronics Status Trigger and DAQ run successfully in RUN2006 for the first time Trigger communication to DRS boards via trigger bus Trigger firmware.
Large scale data flow in local and GRID environment Viktor Kolosov (ITEP Moscow) Ivan Korolko (ITEP Moscow)
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
LHCb datasets and processing stages. 200 kB100 kB 70 kB 0.1 kB 10kB 150 kB 0.1 kB 200 Hz LHCb datasets and processing stages.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
CWG13: Ideas and discussion about the online part of the prototype P. Hristov, 11/04/2014.
D0 Farms 1 D0 Run II Farms M. Diesburg, B.Alcorn, J.Bakken, R. Brock,T.Dawson, D.Fagan, J.Fromm, K.Genser, L.Giacchetti, D.Holmgren, T.Jones, T.Levshina,
Cluster Computers. Introduction Cluster computing –Standard PCs or workstations connected by a fast network –Good price/performance ratio –Exploit existing.
Markus Frank (CERN) & Albert Puig (UB).  An opportunity (Motivation)  Adopted approach  Implementation specifics  Status  Conclusions 2.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
10/18/01Linux Reconstruction Farms at Fermilab 1 Steven C. Timm--Fermilab.
Jianming Qian, UM/DØ Software & Computing Where we are now Where we want to go Overview Director’s Review, June 5, 2002.
HTCC coffee march /03/2017 Sébastien VALAT – CERN.
CMS High Level Trigger Configuration Management
Enrico Gamberini, Giovanna Lehmann Miotto, Roland Sipos
Controlling a large CPU farm using industrial tools
Enrico Bonaccorsi, (CERN) Loic Brarda, (CERN) Gary Moine, (CERN)
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
Toward a costing model What next? Technology decision n Schedule
The Performance and Scalability of the back-end DAQ sub-system
Cluster Computers.
Presentation transcript:

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen2 Motivation  Commodity hardware (PC farms, Ethernet networks) and Linux OS, used in online environment, allow to blur the sharp border online/offline  HERA-B uses successfully Linux PC Farms in the Trigger and Data Acquisition systems  Traditionally in HEP experiments, online and offline computing and software are sharply separated  Different environment and requirements  Dedicated hardware and software in DAQ and trigger  The reconstruction, typically offline task, is done online at HERA-B  HERA-B uses online computing and software resources to perform offline data reprocessing and MC production in the online PC Farms

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen3 HERA-B DAQ DSP switch High Bandwidth (10 Gbps) Low Latency (<10  s) Online PC Farms

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen4 Online PC Farms  L2/L3 trigger step  240 nodes  Intel Celeron 1.3 GHz  256 MB RAM  Fast Ethernet NIC  Linux OS  No real time extensions  CAN card for slow control  Temperature, Power up/down  Online reconstruction & L4 Trigger  100 dual-CPU nodes  Intel PIII 550 MHz  256 MB RAM  Fast Ethernet NIC  Linux OS  No real time extensions  CAN card for slow control  Temperature, Power up/down L2/L3 FarmL4 Farm  Diskless PCs  PROM in NIC loads Linux  Extremely ease maintenance  DSP-to-PCI interface  data link to DSP switch (40 MB/s, 1  s driver latency)

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen5 L4 FARM tasks  Full online event reconstruction  Allow immediate physics analysis  Avoid relatively slow access to tape (20 TB/year)  Full online reconstruction allows online Data Quality Monitoring and Online Calibration and Alignment  Online Event Classification and Selection  Mark events in physics categories (event directories)  L4 trigger step  Data logging  Add reconstruction info to event and send to logger

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen6 L4 FARM Software  Linux environment  Process server  Frame Program ARTE  Reconstruction, analysis and MC  Same code online and offline  Data I/O shm memory based (online) and file based (offline)  Event reconstruction time ~ 4 sec  50 Hz output rate Process Server

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen7 Online DQM and CnA  Online CnA to keep trigger performance & online reconstruction  DQM from reconstructed data  Gathering system to increase statistics  CnA version tag in event data  CnA constants multicasted to L2 nodes by DAQ  CnA constants retrieved from DB by L4 nodes when new CnA tag in events

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen8 Booting and State Machine  Each run has ~2000 process. (~ 400 are under State Machine).  The run is booted in 3 minutes (~10 process/s ).  Different machine types: Linux, Lynx and DSP. & the same protocol.  The State Machine maps different transition different levels in the State Machine tree.  All procresses are booted remotely in different machines using the messaging system.

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen9 Offline  Online  Idea: Use online idle time to perform offline mass data processing using the online computing resources  Shutdown periods, time between spills, accelerator down time  Use vast online computing resources  440 CPUs, high network bandwidth  Use not only online hardware but also online processes and protocols:  Use online boot and control systems  Use online data transmission protocols  Perform “online” Data Quality Monitoring  Run “quasi-online” Data re-processing and Monte Carlo production

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen10 Data Taking L2 Buffers DSP switch L2/L3 Farm Ethernet switch L4 Farm EVC L4C Archiver TAPE

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen11 Data Re-processing L2/L3 Farm Ethernet switch L4 Farm Archiver TAPE Provider

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen12 Monte Carlo Production L2/L3 Farm Ethernet switch L4 Farm Archiver TAPE  Full Monte Carlo Production:  Generation, Detector Simulation, Digitization, Trigger Simulation and Full Reconstruction  GHz node. 300 KB/evt  1 Million evts/day, 300 Gbytes/day

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen13 Quasi-Online Processing  System fully integrated in the Run Control System  Shift crew can use efficiently the online idle time  Same online processes and protocols used for booting, control, monitoring, data reconstruction, data quality, logging and archiving Data Reprocessing = Online Reconstruction

ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen14 Summary  Efficient use of online computing resources at HERA-B to perform mass offline data processing  Not only the online hardware is used but also the online boot, control, monitoring and data transmission processes and protocols LHC experiments might consider to include the online computing power as GRID resources in order to use the online idle time for offline mass data processing