290 likes | 454 Vues
DAQ and Trigger upgrade. U. Marconi, INFN Bologna Firenze, March 2014. DAQ and Trigger TDR. TDR planned for June 2014 Sub-systems involved: Long distance optical fibres Readout boards (PCIe40) and Event Builder Farm The Event Filter Farm for the HLT Firmware, LLT, TFC.
E N D
DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March 2014
DAQ and Trigger TDR • TDR planned for June 2014 • Sub-systems involved: • Long distance optical fibres • Readout boards (PCIe40) and Event Builder Farm • The Event Filter Farm for the HLT • Firmware, LLT, TFC
Readout System Review • Held on Feb 25th, 9:00 – 16:00 • 4 reviewers • ChristophSchwick (PH/CMD – CMS DAQ) • Stefan Haas (PH/ESE) • Guido Haefeli (EPFL) • Jan Troska (PH/ESE) • https://indico.cern.ch/event/297003/ • All documents can be found in EDMShttps://edms.cern.ch/document/1357418/1
Trigger LHCb TB Meeting LHCb Technical Board Meeting 18/03/2014 R. Legac, Trigger Coordinator Full software trigger, software LLT
Long distance fibres Locate the Event Builder Farm and the Event Filter Farm for the HLT outside the cavern. • Minimum required bandwidth: 32 Tbit/s • # of 100 Gigabit/s links > 320, • # of 40 Gigabit/s links > 800, • # of 10 Gigabit/s links > 3200 • # of 5Gbit GBT links (4.5 Gbit effective): > 10000 • Current estimate O(15000), DAQ, Controls and Timing system. Spares included The distance to cover with 850 nm OM multimode optical cables, from underground to the surface, is 300 m.
Long distance fibres (II) • 4.8 Gbit/s signal produced on the detector by Versatile Link transmitters. • VTTx to MiniPod for data acquisition. • MiniPod to VTRx for control, configuration. • 144 fibres per cable. A total of 120 such cables. • 3 patch panels (breakpoints) foreseen:expected attenuation ~ 3dB MiniPod
Optical fibres studies Transmission test at 4.8 Gb/s Measurement of BER vs. receive OMA(*) on different OM3 and OM4 fibres. (*) optical modulation amplitude The target value of 3 dB on OM3 using the Versatile Link and MP is reachable.
Optical links A shortest path seems possible OM3 vs OM4
LHCb DAQ today Push-protocol with centralized flow-control Readout boards: 313 x TELL1 2 x F10 E1200i 56 sub-farms
DAQ Event builder High speed network E B F P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 16-lane PCIe-3 16-lane PCIe-3 HLT Event filter E F F
Readout board review • Basic question to reviewers: “We propose a change of baseline from ATCA/AMC to PCIe. Do the reviewers support the choice of baseline (PCIe) and back-up (ATCA)?” • Answer:“Given the listed advantages the review committee endorses the choice of the PCIe based design as a baseline.”
PCIe Gen3 based readout • A main FPGA manages the input streams and transmits data to the event-builder server using PCIe Gen3. • PCIe Gen3 throughput: 16-lane × 8 Gb/s/lane = 128 Gb/s • The readout version of the board uses two de-serializers. 24 optical links from the FEE DMA over 8-lane PCIe-3 hard IP blocks U. Marconi et al. The PCIe Gen3 readout for the LHCb upgrade, CHEP2013 16-lane PCIe-3 edge connector to the motherboard of the host PC
PCIe-3 test board • ALTERA development board, equipped with a Stratix V GX FPGA, model 5SGXEA7K2F40C2N 8-lane edge-connector http://www.altera.com/literature/manual/rm_svgx_fpga_dev_board.pdf “The Stratix V GX FPGA development board is designed to fit entirely into a PC motherboard with a ×8 PCI Express slot that can accommodate a full height short form factor add-in card. This interface uses the Stratix V GX FPGA device's PCI Express hard IP block, saving logic resources for the user logic application”.
The PCIe-3 DMA test setup GPU used to test 16-lane PCIe-3 data transfer between the device and the host memory The ALTERA development board 8-lane PCIe-3
The DMA PCIe-3 effective bandwidth DMA over 8-lane PCIe-3 hard IP blocks ALTERA Stratix V
Test of PLX bridge 8-lane 110 Gb/s PLX PCIe switch • Long-term test using GTX690 card / PLX 8747 bridge. • Zero impact of using a bridge and two independent PCIe targets pushing data into a PC. Consistently around 110 Gb/s over long-term. • No load balancing issues between the two competing links observed. • Details on: https://lbonupgrade.cern.ch/wiki/index.php/I/O_performance_of_PC_servers#Upgrade_to_GTX690 10 h 16-lane GPU 2 GPU 1 PLX PCIe 16
Event builder node DDR3 40-50 GB/s Half duplex DDR3 40-50 GB/s Half duplex events that are being built on this machine opportunity for doing pre-processing of full event here 2x50 GB/s Full duplex empty Event Building Network Interface PCIe40 data from the detector ~ 100 Gb/s to the event builder Dual-port IB FDR – 110 Gb/s to the HLT from the event builder
EVB performance CPU consumption Memory I/O bandwidth • Event Builder execution requires around 6 logical cores • 18 instances of the HLT software • Event Builder performs stably at 400 Gb/s PC sustains 100 Gb/s Event Builder today Ivy Bridge Intel dual CPU 12 cores We currently see 50% free resources for opportunistic triggering on EB nodes
Costing (preliminary) • Long-distance fibres:~ 1.6 MCHF 15000 fibresOM3, 300 m. • Excluding patch-cords to detector and TELL40, cable-ducts, but including patch-panels, installation and testing • No contingency, but several quotes. • PCIe40:DAQ version: 5.8 kCHFECS/TRG version: 7.9 kCHF • includes 15% contingency • Readout network,Event-building (PCIe40) ~ 3.6 MCHF • Including event-builder PCs and event-filter network, excluding farm-servers • Model based on InfiniBand FDR (2013 quotes) • Event-building (AMC40)~ 9 MCHF • including event-filter network • Model based on 40G and 10G Ethernet (on 2013 quotes)
Running conditions Mean visible interactions per crossing • Visible rate and mean visible interactions per crossing • At 2.×1033 , 27 MHz, 5.2 interaction per crossing
LHCb upgrade: HLT farm • Trigger-less system at 40 MHz:A selective, efficient and adaptable software trigger. • Average event size: 100 kB • Expected data flux: 4 TB/s • Total HLT trigger process latency: 14 ms • Tracking time budget (VELO + Tracking + PV searches): 50% • Tracking finds 99% of offline tracks with pT>500 MeV/c • Number of running trigger process required: 4.×105 • Number of core/CPU available in 2018: ~ 200 • Intel tick-tock plan: 7nm technology available by 2018-19, the n. of core accordingly scales as 12. × (32 nm/ 7 nm)2 = 250 • Number of computing nodes required: ~ 1000
LLT Readout reviewers’ opinion … only the software option should be pursued if any.
Person-power • Fibres: (CERN) • Tests and preparation of installation 0.5 py (person-year) • Installation & testing by external company 0.5 py for supervision and follow-up • PCIe40: (Marseille, Bologna) • design, production, commissioning ~ 18 py • LLT: (Marseille,Annecy, Clermont Fd., LAL) ~ 17 py • Firmware: Global framework, TFC, DAQ (Marseille, Bologna, CERN, Annecy) ~ 18 py (estimated) • DAQ: (CERN, Bologna) • ~ 15 py (excluding ECS and high-level trigger infrastructure) • Overall person-power in involved institutes is sufficient, but does not allow for any “luxuries”, as many people are also involved in the operation of LHCb during Run 2 N. Neufeld
Schedulereadout board Schedule TELL40 PCIe40 board plan: as INFN we should develop the PCB and prototypes, to qualify our production: be ready to produce the boards. Sblocco 25 kEs.j.
Schedule DAQ N. Neufeld • Assume start of data-taking in 2020 • System for SD-tests ready whenever needed minimal investment • 2013 – 16: technology following (PC and network) • 2015 – 16: Large scale network IB and Ethernet tests • 2017: tender preparations • 2018: Acquisition of minimal system to be able to read out every GBT • Acquisition of modular data-center • 2019: Acquisition and Commissioning of full system • starting with network • farm as needed
Schedule: firmware, LLT, TFC • All these are FPGA firmware projects • First versions of global firmware and TFC ready now (for MiniDAQ test-systems) • then ongoing development • LLT • Software 2014 - 2015 • Hardware 2016 – 2017 (?)
Schedule long-distance fibres • Test installation 2014 • validate installation procedure and pre-select provider • Long-term fibre test with AMC40/PCIe40 on long-distance 2014/2015 • Full installation in LS1.5 or during winter-shutdowns / to be finished before LS2 • Assumptions: • Installation can be done without termination in UX (cables terminated on at least one end), if blown, fibres can be blown from bottom to top