Presentation is loading. Please wait.

Presentation is loading. Please wait.

DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March 2014 1.

Similar presentations

Presentation on theme: "DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March 2014 1."— Presentation transcript:

1 DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March 2014 1

2 DAQ and Trigger TDR TDR planned for June 2014 Sub-systems involved: – Long distance optical fibres – Readout boards (PCIe40) and Event Builder Farm – The Event Filter Farm for the HLT – Firmware, LLT, TFC 2

3 Readout System Review Held on Feb 25 th, 9:00 – 16:00 4 reviewers – Christoph Schwick (PH/CMD – CMS DAQ) – Stefan Haas (PH/ESE) – Guido Haefeli (EPFL) – Jan Troska (PH/ESE) All documents can be found in EDMS 3

4 Trigger LHCb TB Meeting 4 LHCb Technical Board Meeting 18/03/2014 Full software trigger, software LLT R. Legac, Trigger Coordinator

5 Long distance fibres 5 The distance to cover with 850 nm OM multimode optical cables, from underground to the surface, is 300 m. Locate the Event Builder Farm and the Event Filter Farm for the HLT outside the cavern. Minimum required bandwidth: 32 Tbit/s – # of 100 Gigabit/s links > 320, – # of 40 Gigabit/s links > 800, – # of 10 Gigabit/s links > 3200 – # of 5Gbit GBT links (4.5 Gbit effective): > 10000 – Current estimate O(15000), DAQ, Controls and Timing system. Spares included

6 Long distance fibres (II) 144 fibres per cable. A total of 120 such cables. 3 patch panels (breakpoints) foreseen: expected attenuation ~ 3dB 4.8 Gbit/s signal produced on the detector by Versatile Link transmitters. VTTx to MiniPod for data acquisition. MiniPod to VTRx for control, configuration. 6 MiniPod

7 Optical fibres studies 7 Measurement of BER vs. receive OMA(*) on different OM3 and OM4 fibres. (*) optical modulation amplitude Transmission test at 4.8 Gb/s The target value of 3 dB on OM3 using the Versatile Link and MP is reachable.

8 Optical links 8 OM3 vs OM4 A shortest path seems possible

9 LHCb DAQ today 2 x F10 E1200i 56 sub-farms Readout boards: 313 x TELL1 9 Push-protocol with centralized flow-control # links (UTP Cat 6)~ 3000 Event-size (total – zero- suppressed) 65 kB Read-out rate1 MHz # read-out boards313 output bw / read-out boardup to 4 Gbit/s (4 Ethernet links) # farm-nodes1500 (up to 2000) max. input bw / farm-node1 Gbit/s # core-routers2 # edge routers56

10 DAQ Event builder High speed network HLT Event filter 10 PCIe40PCIe40 PCIe40PCIe40 PCIe40PCIe40 PCIe40PCIe40 PCIe40PCIe40 PCIe40PCIe40 PCIe40PCIe40 PCIe40PCIe40 16-lane PCIe-3 EBFEBF EFFEFF

11 Readout board review Basic question to reviewers: “We propose a change of baseline from ATCA/AMC to PCIe. Do the reviewers support the choice of baseline (PCIe) and back-up (ATCA)?” Answer: “Given the listed advantages the review committee endorses the choice of the PCIe based design as a baseline.” 11

12 PCIe Gen3 based readout  A main FPGA manages the input streams and transmits data to the event-builder server using PCIe Gen3.  PCIe Gen3 throughput: 16-lane × 8 Gb/s/lane = 128 Gb/s  The readout version of the board uses two de-serializers. 12 16-lane PCIe-3 edge connector to the motherboard of the host PC 24 optical links from the FEE DMA over 8-lane PCIe-3 hard IP blocks U. Marconi et al. The PCIe Gen3 readout for the LHCb upgrade, CHEP2013

13 PCIe-3 test board ALTERA development board, equipped with a Stratix V GX FPGA, model 5SGXEA7K2F40C2N 13 “The Stratix V GX FPGA development board is designed to fit entirely into a PC motherboard with a ×8 PCI Express slot that can accommodate a full height short form factor add-in card. This interface uses the Stratix V GX FPGA device's PCI Express hard IP block, saving logic resources for the user logic application”. 8-lane edge-connector

14 14 The PCIe-3 DMA test setup GPU used to test 16-lane PCIe-3 data transfer between the device and the host memory The ALTERA development board 8-lane PCIe-3

15 15 The DMA PCIe-3 effective bandwidth DMA over 8-lane PCIe-3 hard IP blocks ALTERA Stratix V

16 Test of PLX bridge Long-term test using GTX690 card / PLX 8747 bridge. Zero impact of using a bridge and two independent PCIe targets pushing data into a PC. Consistently around 110 Gb/s over long- term. No load balancing issues between the two competing links observed. Details on: /wiki/index.php/I/O_perfor mance_of_PC_servers#Upgr ade_to_GTX690 /wiki/index.php/I/O_perfor mance_of_PC_servers#Upgr ade_to_GTX690 10 h 110 Gb/s PLX GPU 1 GPU 2 PCIe 16 PLX 8-lane 16-lane PCIe switch 16

17 Event builder node 17 PCIe40 Event Building Network Interface Event Building Network Interface data from the detector ~ 100 Gb/s to the event builder Dual-port IB FDR – 110 Gb/s from the event builder events that are being built on this machine opportunity for doing pre-processing of full event here to the HLT empty DDR3 40-50 GB/s Half duplex 2x50 GB/s Full duplex DDR3 40-50 GB/s Half duplex

18 EVB performance 18 Event Builder execution requires around 6 logical cores 18 instances of the HLT software CPU consumption Memory I/O bandwidth Event Builder performs stably at 400 Gb/s PC sustains 100 Gb/s Event Builder today Ivy Bridge Intel dual CPU 12 cores We currently see 50% free resources for opportunistic triggering on EB nodes

19 Fast Control 19

20 Costing (preliminary) Long-distance fibres: ~ 1.6 MCHF 15000 fibres OM3, 300 m. – Excluding patch-cords to detector and TELL40, cable-ducts, but including patch-panels, installation and testing – No contingency, but several quotes. PCIe40: DAQ version: 5.8 kCHF ECS/TRG version: 7.9 kCHF – includes 15% contingency Readout network, Event-building (PCIe40) ~ 3.6 MCHF – Including event-builder PCs and event-filter network, excluding farm-servers – Model based on InfiniBand FDR (2013 quotes) Event-building (AMC40) ~ 9 MCHF – including event-filter network – Model based on 40G and 10G Ethernet (on 2013 quotes) 20

21 Running conditions Visible rate and mean visible interactions per crossing At 2.×10 33, 27 MHz, 5.2 interaction per crossing 21 Mean visible interactions per crossing

22 LHCb upgrade: HLT farm Trigger-less system at 40 MHz: A selective, efficient and adaptable software trigger. Average event size: 100 kB Expected data flux: 4 TB/s Total HLT trigger process latency: 14 ms – Tracking time budget (VELO + Tracking + PV searches): 50% – Tracking finds 99% of offline tracks with p T >500 MeV/c Number of running trigger process required: 4.×10 5 Number of core/CPU available in 2018: ~ 200 – Intel tick-tock plan: 7nm technology available by 2018-19, the n. of core accordingly scales as 12. × (32 nm/ 7 nm) 2 = 250 Number of computing nodes required: ~ 1000 22

23 LLT 23 Readout reviewers’ opinion … only the software option should be pursued if any.

24 LLT performance 24

25 Person-power Fibres: (CERN) – Tests and preparation of installation 0.5 py (person-year) – Installation & testing by external company  0.5 py for supervision and follow- up PCIe40: (Marseille, Bologna) – design, production, commissioning ~ 18 py LLT: (Marseille, Annecy, Clermont Fd., LAL) ~ 17 py Firmware: Global framework, TFC, DAQ (Marseille, Bologna, CERN, Annecy) ~ 18 py (estimated) DAQ: (CERN, Bologna) – ~ 15 py (excluding ECS and high-level trigger infrastructure) Overall person-power in involved institutes is sufficient, but does not allow for any “luxuries”, as many people are also involved in the operation of LHCb during Run 2 25 N. Neufeld

26 Schedule TELL40 26 Schedule readout board PCIe40 board plan: as INFN we should develop the PCB and prototypes, to qualify our production: be ready to produce the boards. Sblocco 25 kE s.j.

27 Schedule DAQ Assume start of data-taking in 2020 – System for SD-tests ready whenever needed  minimal investment 2013 – 16: technology following (PC and network) 2015 – 16: Large scale network IB and Ethernet tests 2017: tender preparations 2018: Acquisition of minimal system to be able to read out every GBT – Acquisition of modular data-center 2019: Acquisition and Commissioning of full system – starting with network – farm as needed 27 N. Neufeld

28 Schedule: firmware, LLT, TFC All these are FPGA firmware projects First versions of global firmware and TFC ready now (for MiniDAQ test-systems) – then ongoing development LLT – Software 2014 -2015 – Hardware 2016 – 2017 (?) 28

29 Schedule long-distance fibres Test installation 2014 – validate installation procedure and pre-select provider Long-term fibre test with AMC40/PCIe40 on long- distance 2014/2015 Full installation in LS1.5 or during winter-shutdowns / to be finished before LS2 Assumptions: – Installation can be done without termination in UX (cables terminated on at least one end), if blown, fibres can be blown from bottom to top 29

Download ppt "DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March 2014 1."

Similar presentations

Ads by Google