SuperBelle Event Building System Takeo Higuchi (IPNS, KEK) S.Y.Suzuki (CC, KEK) 2008/12/12 1st Open Meeting of the SuperKEKB Collaboration.

Slides:



Advertisements
Similar presentations
Network II.5 simulator ..
Advertisements

TELE202 Lecture 8 Congestion control 1 Lecturer Dr Z. Huang Overview ¥Last Lecture »X.25 »Source: chapter 10 ¥This Lecture »Congestion control »Source:
MIGSOCK Migratable TCP Socket in Linux Demonstration of Functionality Karthik Rajan Bryan Kuntz.
A CHAT CLIENT-SERVER MODULE IN JAVA BY MAHTAB M HUSSAIN MAYANK MOHAN ISE 582 FALL 2003 PROJECT.
Page: 1 Director 1.0 TECHNION Department of Computer Science The Computer Communication Lab (236340) Summer 2002 Submitted by: David Schwartz Idan Zak.
ACAT 2002, Moscow June 24-28thJ. Hernández. DESY-Zeuthen1 Offline Mass Data Processing using Online Computing Resources at HERA-B José Hernández DESY-Zeuthen.
CHEP04 - Interlaken - Sep. 27th - Oct. 1st 2004T. M. Steinbeck for the Alice Collaboration1/27 A Control Software for the ALICE High Level Trigger Timm.
Takeo Higuchi Institute of Particle and Nuclear Studies, KEK on behalf of the COPPER working group Apr.22,2005Super B Factory Workshop Detector (DAQ/Computing)
I/O Systems ◦ Operating Systems ◦ CS550. Note:  Based on Operating Systems Concepts by Silberschatz, Galvin, and Gagne  Strongly recommended to read.
Server Design Discuss Design issues for Servers Review Server Creation in Linux.
Cli/Serv.: JXTA/151 Client/Server Distributed Systems v Objective –explain JXTA, a support environment for P2P services and applications ,
Chapter 17 Networking Dave Bremer Otago Polytechnic, N.Z. ©2008, Prentice Hall Operating Systems: Internals and Design Principles, 6/E William Stallings.
FTP (File Transfer Protocol) & Telnet
MSS, ALICE week, 21/9/041 A part of ALICE-DAQ for the Forward Detectors University of Athens Physics Department Annie BELOGIANNI, Paraskevi GANOTI, Filimon.
Designing For Testability. Incorporate design features that facilitate testing Include features to: –Support test automation at all levels (unit, integration,
SVT workshop October 27, 1998 XTF HB AM Stefano Belforte - INFN Pisa1 COMMON RULES ON OPERATION MODES RUN MODE: the board does what is needed to make SVT.
+ discussion in Software WG: Monte Carlo production on the Grid + discussion in TDAQ WG: Dedicated server for online services + experts meeting (Thusday.
Operating Systems Recitation 9, May 19-20, Iterative server Handle one connection request at a time. Connection requests stored in queue associated.
C.Combaret, L.Mirabito Lab & beamtest DAQ with XDAQ tools.
Guide to Linux Installation and Administration, 2e1 Chapter 10 Managing System Resources.
Lecture 3 Process Concepts. What is a Process? A process is the dynamic execution context of an executing program. Several processes may run concurrently,
Data Acquisition System of SVD2.0 This series of slides explains how we take normal, calibration and system test data of the SVD 2.0 and monitor the environment.
ALICE, ATLAS, CMS & LHCb joint workshop on
Network Architecture for the LHCb DAQ Upgrade Guoming Liu CERN, Switzerland Upgrade DAQ Miniworkshop May 27, 2013.
1 Network Administration Module 3 ARP/RARP. 2 Address Resolution The problem Physical networks use physical addresses, not IP addresses Need the physical.
Agilent Technologies Copyright 1999 H7211A+221 v Capture Filters, Logging, and Subnets: Module Objectives Create capture filters that control whether.
Management of the LHCb DAQ Network Guoming Liu * †, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
CE Operating Systems Lecture 13 Linux/Unix interprocess communication.
Overview of DAQ at CERN experiments E.Radicioni, INFN MICE Daq and Controls Workshop.
1 © 2003, Cisco Systems, Inc. All rights reserved. CCNA 3 v3.0 Module 8 Virtual LANs Cisco Networking Academy.
STORE AND FORWARD & CUT THROUGH FORWARD Switches can use different forwarding techniques— two of these are store-and-forward switching and cut-through.
Image Builder Design Tao Qian UIUC April 10, 2007.
Silberschatz, Galvin and Gagne  Operating System Concepts UNIT II Operating System Services.
Sep. 17, 2002BESIII Review Meeting BESIII DAQ System BESIII Review Meeting IHEP · Beijing · China Sep , 2002.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Online Monitoring for the CDF Run II Experiment T.Arisawa, D.Hirschbuehl, K.Ikado, K.Maeshima, H.Stadie, G.Veramendi, W.Wagner, H.Wenzel, M.Worcester MAR.
CSI 3125, Preliminaries, page 1 SERVLET. CSI 3125, Preliminaries, page 2 SERVLET A servlet is a server-side software program, written in Java code, that.
Source Controller software Ianos Schmidt The University of Iowa.
Protocol Layering Chapter 11.
Status & development of the software for CALICE-DAQ Tao Wu On behalf of UK Collaboration.
TELL1 command line tools Guido Haefeli EPFL, Lausanne Tutorial for TELL1 users : 25.February
October Test Beam DAQ. Framework sketch Only DAQs subprograms works during spills Each subprogram produces an output each spill Each dependant subprogram.
1 Farm Issues L1&HLT Implementation Review Niko Neufeld, CERN-EP Tuesday, April 29 th.
Management of the LHCb DAQ Network Guoming Liu *†, Niko Neufeld * * CERN, Switzerland † University of Ferrara, Italy.
January 9, 2001 Router Plugins (Crossbow) 1 Washington WASHINGTON UNIVERSITY IN ST LOUIS Exercises.
Event Management. EMU Graham Heyes April Overview Background Requirements Solution Status.
R.Divià, CERN/ALICE 1 ALICE off-line week, CERN, 9 September 2002 DAQ-HLT software interface.
1 DAQ.IHEP Beijing, CAS.CHINA mail to: The Readout In BESIII DAQ Framework The BESIII DAQ system consists of the readout subsystem, the.
1 15-Mar-16 VLAN Trunking protocol CCNA Exploration Semester 3 Chapter 4.
DAQ & ConfDB Configuration DB workshop CERN September 21 st, 2005 Artur Barczyk & Niko Neufeld.
PXD DAQ News S. Lange (Univ. Gießen) Belle II Trigger/DAQ Meeting (Jan 16-18, 2012, Hawaii, USA) Today: only topics important for CDAQ - GbE Connection.
Demo system of Belle2link Sun Dehui, Zhao Jingzhou,Liu zhen’an Trigger Lab IHEP.
Segments Introduction: slides 2–6, 8 10 minutes
Module 12: I/O Systems I/O hardware Application I/O Interface
Controlling a large CPU farm using industrial tools
RT2003, Montreal Niko Neufeld, CERN-EP & Univ. de Lausanne
Multi-PCIe socket network device
Ch > 28.4.
IS3440 Linux Security Unit 9 Linux System Logging and Monitoring
I/O Systems I/O Hardware Application I/O Interface
Operating System Concepts
CS703 - Advanced Operating Systems
CS580 Special Project: IOS Firewall Setup using CISCO 1600 router
An Introduction to Software Architecture
The LHCb High Level Trigger Software Framework
Network Processors for a 1 MHz Trigger-DAQ System
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Chapter 13: I/O Systems I/O Hardware Application I/O Interface
Module 12: I/O Systems I/O hardwared Application I/O Interface
Message Passing Systems
Presentation transcript:

SuperBelle Event Building System Takeo Higuchi (IPNS, KEK) S.Y.Suzuki (CC, KEK) 2008/12/12 1st Open Meeting of the SuperKEKB Collaboration

Belle DAQ DiagramBelle Event Builder COPPER/FASTBUS readout Construction of event record Data transmission to RFARM Construction of event record Data transmission to RFARM

Configuration of SB Event Builder CPR crateROPC #01 RFARM #01 RFARM #02 RFARM #03 EFARM #01 SW EFARM #02 SW EFARM #03 SW EFARM #n CPR crateROPC #02 CPR crateROPC #03 CPR crateROPC #04 CPR crateROPC #m … …… SW RFARM #n The m and n depend on the detector design. From our experience, the performance limit here is at 1-2Gbps

Regulation on FINESSE by EB No event# skip / No event# shuffle – Any kind of data record must be sent out from the FINESSE upon the L1 trigger. At least a capsule with event# tag (and empty body) must be generated even for a totally zero-suppressed event or an L2-aborted event. – Event # by the FINESSE must be increased monotonically by … … … OK NG  FATAL

Inside EFARM #xx Multi-layer event building – Similar to the present EFARM configuration. EFARM #xx 1st layer2nd layeri-th layerExit layer … … PC

Performance Study Inside EB (1) Master thesis of K.Ito (2005) P2P connectionConnection over SW 100BaseT 1000BaseT 100BaseT GbE SW P2P connection Connection over SW SW works fine N=8 1000BaseT 100BaseT GbE SW Scalability up to N~10 …

Performance Study Inside EB (2) Master thesis of K.Ito (2005) 1000BaseT 100BaseT GbE SWROPC 1st Layer Pseudo 2nd Layer Event rateTotal data transfer rate N= N < 5 ~

Performance Study Inside EB (3) Master thesis of K.Ito (2005) N=5 N=1..10 ROPC 1st Layer 2nd Layer 800 bytes/ev1600 bytes/ev3200 bytes/ev6400 bytes/ev vs. # of 1st later PCs Event rateTotal data transfer rate

Concepts of Implementation Switch from house-made technologies to generally existing technologies – E.g. (1): event buffering Now:House made (shared memory + semaphore). SB:Socket buffer within Linux kernel. – E.g.(2): state machine Now:All EB nodes are managed by NSM. SB: rwho or finger +.plan Only one NSM node to respond to master. This collect all EB nodes’ status via those commands. Less effort for documentation

Concepts of Implementation Quit from complexity – NSM:Only one NSM node to respond to master. – D2:Own error reporting / logging  syslog. – No (or minimum) configuration file: the present EB requires large effort to add new ROPCs for its multiple and complicated configuration files... – Common software for all EB PCs in each layer. Less effort for trouble shooting We have just started the conceptual design. Following slides includes many unimplemented ideas.

Network Link Propagation … … Network connections propagate at RUN START from DOWNstream to UPstream.  # or ports awaiting for connection can be unity. COPPERs E1 01 E2 01 E2 01 RO 01 RO 0n E1 02 E2 02 E2 02 En 01 En 01 RF

Determination of PCs to Connect Idea #1: Determine EB PCs to connect from its own hostname + rule. –EB PC connects to all available (= accepting connection) PCs at run start, which match to the rule. – Two new schemes must be created: Detection of PCs which are not booted but included to CDAQ. Detection of PCs which are booted but excluded from CDAQ. E02_01 E02_02 E03_01 E01_11 E01_12 E01_19 E01_13 E.g.: … E01_21 E01_22 E01_29 E01_23 … E02_11 E02_12 E02_19 E02_13 … Recall “no (or minimum) configuration file” Recall “no (or minimum) configuration file”

Determination of PCs to Connect Idea #2:“EB_CONFIG server”. – EB_CONFIG server knows which EB PCs are booted up and which are not. – EB_CONFIG server knows which subsystems are included and which are excluded (told from MASTER). – EB_CONFIG server knows which EB PCs belong to which subsystem (from a configuration file). But this is what we like to avoid.

EB PC Inside EB PC (Process Diagram) downstream  upstream port #nnn … connect to upstream connect to upstream connect to upstream output to downstream event builder xinetd eb_process fork(2) An EB PC in the exit layer may have to output the built-up record to one of multiple destinations (RFARM PCs), which consequently means the EB PC accepts multiple connections from each destination PC. We have found a way to pass socket descriptor to another process. This enables us to perform event building, destination selection, and data transmission in a single process. ¶

Input Buffer Use Linux kernel buffer instead of house-made buffering scheme – Much easier implementation/maintenance. – Unlike when the present EB developed, very large buffer (~80MB) can be allocated in the kernel now. Inside kernel Data shmop recv Shared memory Build Inside kernel Data recv Build NowSuperBelle

Output Buffer Recent Linux provides information about how much unsent data remain in the kernel output buffer. – Established by SK DAQ team (Hayato, Yamada).  Enables dynamic traffic balancing. ioctl( sd, SIOCOUTQ, &val); PC #00 in RFARM #01 EB PC in exit layer PC #01 in RFARM #01 PC #02 in RFARM #01 idle too busy to recv data unsent data PC #00 in RFARM #01 EB PC in exit layer PC #01 in RFARM #01 PC #02 in RFARM #01 Select data destination to balance the traffic load.

PC: EBNSM State Machine / Status Collection Possible statuses of EB PC Collection of EB PC statuses and EB Configuration Accepting side Not ready Ready Partially connected Fully connected From downstream to the EB PC From the EB PC to upstream Connecting side Not ready Ready Partially connected Fully connected Ready to run EB_CONF server Collection of EB PC statuses Query: which EB PC is ready? to connect? included? NSM command: EB_CONFIG … NSM reply: READY, NOTREADY … nsmd CDAQ side EFARM side rwho is used.

Error Reporting and Logging INFO WARN ERROR FATAL EB Process syslog(3) /var/log/messages tail –f grep Watch with Throw error messages to the NSM master.

BASF Substitution Separation of BASF from EFARM – Present EFARM strongly depends on BASF; BASF modules build up event records.  May cause BELLE_LEVEL problem. – SuperBelle EFARM is not to use BASF to build up event. Substitution of BASF – BASF-like substitution for online data quality check and/or L3 trigger use will be provided. const int (*userfunc[])(unsigned int *data, size_t size);

Summary We have started the design of event building farm for SuperBelle. Regulation on the FINESSE is given from the EFARM. Several performance studies about network data link are already made to input to the new EFARM design. The new EFARM fully utilizes pre-implemented Linux functions instead our own invention. The new EFARM tries to be self-configured to be free from configuration-file nightmare.