1 / 29

DAQ and Trigger upgrade

This document discusses the DAQ and Trigger upgrade for LHCb, focusing on the use of long-distance optical fibres and readout boards. It also includes information on the event builder and filter farms, firmware, LLT, and TFC readout system review.

llaw
Download Presentation

DAQ and Trigger upgrade

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. DAQ and Trigger upgrade U. Marconi, INFN Bologna Firenze, March 2014

  2. DAQ and Trigger TDR • TDR planned for June 2014 • Sub-systems involved: • Long distance optical fibres • Readout boards (PCIe40) and Event Builder Farm • The Event Filter Farm for the HLT • Firmware, LLT, TFC

  3. Readout System Review • Held on Feb 25th, 9:00 – 16:00 • 4 reviewers • ChristophSchwick (PH/CMD – CMS DAQ) • Stefan Haas (PH/ESE) • Guido Haefeli (EPFL) • Jan Troska (PH/ESE) • https://indico.cern.ch/event/297003/ • All documents can be found in EDMShttps://edms.cern.ch/document/1357418/1

  4. Trigger LHCb TB Meeting LHCb Technical Board Meeting 18/03/2014 R. Legac, Trigger Coordinator Full software trigger, software LLT

  5. Long distance fibres Locate the Event Builder Farm and the Event Filter Farm for the HLT outside the cavern. • Minimum required bandwidth: 32 Tbit/s • # of 100 Gigabit/s links > 320, • # of 40 Gigabit/s links > 800, • # of 10 Gigabit/s links > 3200 • # of 5Gbit GBT links (4.5 Gbit effective): > 10000 • Current estimate O(15000), DAQ, Controls and Timing system. Spares included The distance to cover with 850 nm OM multimode optical cables, from underground to the surface, is 300 m.

  6. Long distance fibres (II) • 4.8 Gbit/s signal produced on the detector by Versatile Link transmitters. • VTTx to MiniPod for data acquisition. • MiniPod to VTRx for control, configuration. • 144 fibres per cable. A total of 120 such cables. • 3 patch panels (breakpoints) foreseen:expected attenuation ~ 3dB MiniPod

  7. Optical fibres studies Transmission test at 4.8 Gb/s Measurement of BER vs. receive OMA(*) on different OM3 and OM4 fibres. (*) optical modulation amplitude The target value of 3 dB on OM3 using the Versatile Link and MP is reachable.

  8. Optical links A shortest path seems possible OM3 vs OM4

  9. LHCb DAQ today Push-protocol with centralized flow-control Readout boards: 313 x TELL1 2 x F10 E1200i 56 sub-farms

  10. DAQ Event builder High speed network E B F P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 P C I e 4 0 16-lane PCIe-3 16-lane PCIe-3 HLT Event filter E F F

  11. Readout board review • Basic question to reviewers: “We propose a change of baseline from ATCA/AMC to PCIe. Do the reviewers support the choice of baseline (PCIe) and back-up (ATCA)?” • Answer:“Given the listed advantages the review committee endorses the choice of the PCIe based design as a baseline.”

  12. PCIe Gen3 based readout • A main FPGA manages the input streams and transmits data to the event-builder server using PCIe Gen3. • PCIe Gen3 throughput: 16-lane × 8 Gb/s/lane = 128 Gb/s • The readout version of the board uses two de-serializers. 24 optical links from the FEE DMA over 8-lane PCIe-3 hard IP blocks U. Marconi et al. The PCIe Gen3 readout for the LHCb upgrade, CHEP2013 16-lane PCIe-3 edge connector to the motherboard of the host PC

  13. PCIe-3 test board • ALTERA development board, equipped with a Stratix V GX FPGA, model 5SGXEA7K2F40C2N 8-lane edge-connector http://www.altera.com/literature/manual/rm_svgx_fpga_dev_board.pdf “The Stratix V GX FPGA development board is designed to fit entirely into a PC motherboard with a ×8 PCI Express slot that can accommodate a full height short form factor add-in card. This interface uses the Stratix V GX FPGA device's PCI Express hard IP block, saving logic resources for the user logic application”.

  14. The PCIe-3 DMA test setup GPU used to test 16-lane PCIe-3 data transfer between the device and the host memory The ALTERA development board 8-lane PCIe-3

  15. The DMA PCIe-3 effective bandwidth DMA over 8-lane PCIe-3 hard IP blocks ALTERA Stratix V

  16. Test of PLX bridge 8-lane 110 Gb/s PLX PCIe switch • Long-term test using GTX690 card / PLX 8747 bridge. • Zero impact of using a bridge and two independent PCIe targets pushing data into a PC. Consistently around 110 Gb/s over long-term. • No load balancing issues between the two competing links observed. • Details on: https://lbonupgrade.cern.ch/wiki/index.php/I/O_performance_of_PC_servers#Upgrade_to_GTX690 10 h 16-lane GPU 2 GPU 1 PLX PCIe 16

  17. Event builder node DDR3 40-50 GB/s Half duplex DDR3 40-50 GB/s Half duplex events that are being built on this machine opportunity for doing pre-processing of full event here 2x50 GB/s Full duplex empty Event Building Network Interface PCIe40 data from the detector ~ 100 Gb/s to the event builder Dual-port IB FDR – 110 Gb/s to the HLT from the event builder

  18. EVB performance CPU consumption Memory I/O bandwidth • Event Builder execution requires around 6 logical cores • 18 instances of the HLT software • Event Builder performs stably at 400 Gb/s PC sustains 100 Gb/s Event Builder today Ivy Bridge Intel dual CPU 12 cores We currently see 50% free resources for opportunistic triggering on EB nodes

  19. Fast Control

  20. Costing (preliminary) • Long-distance fibres:~ 1.6 MCHF 15000 fibresOM3, 300 m. • Excluding patch-cords to detector and TELL40, cable-ducts, but including patch-panels, installation and testing • No contingency, but several quotes. • PCIe40:DAQ version: 5.8 kCHFECS/TRG version: 7.9 kCHF • includes 15% contingency • Readout network,Event-building (PCIe40) ~ 3.6 MCHF • Including event-builder PCs and event-filter network, excluding farm-servers • Model based on InfiniBand FDR (2013 quotes) • Event-building (AMC40)~ 9 MCHF • including event-filter network • Model based on 40G and 10G Ethernet (on 2013 quotes)

  21. Running conditions Mean visible interactions per crossing • Visible rate and mean visible interactions per crossing • At 2.×1033 , 27 MHz, 5.2 interaction per crossing

  22. LHCb upgrade: HLT farm • Trigger-less system at 40 MHz:A selective, efficient and adaptable software trigger. • Average event size: 100 kB • Expected data flux: 4 TB/s • Total HLT trigger process latency: 14 ms • Tracking time budget (VELO + Tracking + PV searches): 50% • Tracking finds 99% of offline tracks with pT>500 MeV/c • Number of running trigger process required: 4.×105 • Number of core/CPU available in 2018: ~ 200 • Intel tick-tock plan: 7nm technology available by 2018-19, the n. of core accordingly scales as 12. × (32 nm/ 7 nm)2 = 250 • Number of computing nodes required: ~ 1000

  23. LLT Readout reviewers’ opinion … only the software option should be pursued if any.

  24. LLT performance

  25. Person-power • Fibres: (CERN) • Tests and preparation of installation 0.5 py (person-year) • Installation & testing by external company  0.5 py for supervision and follow-up • PCIe40: (Marseille, Bologna) • design, production, commissioning ~ 18 py • LLT: (Marseille,Annecy, Clermont Fd., LAL) ~ 17 py • Firmware: Global framework, TFC, DAQ (Marseille, Bologna, CERN, Annecy) ~ 18 py (estimated) • DAQ: (CERN, Bologna) • ~ 15 py (excluding ECS and high-level trigger infrastructure) • Overall person-power in involved institutes is sufficient, but does not allow for any “luxuries”, as many people are also involved in the operation of LHCb during Run 2 N. Neufeld

  26. Schedulereadout board Schedule TELL40 PCIe40 board plan: as INFN we should develop the PCB and prototypes, to qualify our production: be ready to produce the boards. Sblocco 25 kEs.j.

  27. Schedule DAQ N. Neufeld • Assume start of data-taking in 2020 • System for SD-tests ready whenever needed  minimal investment • 2013 – 16: technology following (PC and network) • 2015 – 16: Large scale network IB and Ethernet tests • 2017: tender preparations • 2018: Acquisition of minimal system to be able to read out every GBT • Acquisition of modular data-center • 2019: Acquisition and Commissioning of full system • starting with network • farm as needed

  28. Schedule: firmware, LLT, TFC • All these are FPGA firmware projects • First versions of global firmware and TFC ready now (for MiniDAQ test-systems) • then ongoing development • LLT • Software 2014 - 2015 • Hardware 2016 – 2017 (?)

  29. Schedule long-distance fibres • Test installation 2014 • validate installation procedure and pre-select provider • Long-term fibre test with AMC40/PCIe40 on long-distance 2014/2015 • Full installation in LS1.5 or during winter-shutdowns / to be finished before LS2 • Assumptions: • Installation can be done without termination in UX (cables terminated on at least one end), if blown, fibres can be blown from bottom to top

More Related