1 / 13

Work supported by EU RP6 project JRA1 FutureDAQ RII3-CT-2004-506078

High performance data acquisition with InfiniBand J.Adamczewski, H.G.Essel, N.Kurz, S.Linev GSI, Experiment Electronics, Data Processing group. CBM data acquisition Event building network InfiniBand & OFED Performance tests.

Download Presentation

Work supported by EU RP6 project JRA1 FutureDAQ RII3-CT-2004-506078

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. High performance data acquisition with InfiniBand J.Adamczewski, H.G.Essel, N.Kurz, S.LinevGSI, Experiment Electronics, Data Processing group • CBM data acquisition • Event building network • InfiniBand & OFED • Performance tests Work supported by EU RP6 project JRA1 FutureDAQ RII3-CT-2004-506078

  2. Detector Detector Detector FEE deliver time stamped data ~50000 FEE chips ~1000 collectors CNet collect data into buffers collect ~1000 active buffers data dispatcher TNet time distribution ~1000 links a 1 GB/sec switching network ~10 dispatchers →subfarm BNet sort time stamped data event dispatcher subfarm subfarm subfarm PNet process events level 1&2 selection ~100 subfarms ~100 nodes per subfarm processing HNet high level selection to high level computing and archiving ~1 GB/sec Output CBM data acquisition W.F.J.Müller, 2004 S.Linev High performance data acquisition with InfiniBand

  3. Detector Detector Detector FEE deliver time stamped data ~50000 FEE chips ~1000 collectors CNet collect data into buffers collect ~1000 active buffers data dispatcher TNet time distribution ~1000 links a 1 GB/sec switching network ~10 dispatchers →subfarm BNet sort time stamped data event dispatcher subfarm subfarm subfarm PNet process events level 1&2 selection ~100 subfarms ~100 nodes per subfarm processing HNet high level selection to high level computing and archiving ~1 GB/sec Output Event building network - BNet W.F.J.Müller, 2004 S.Linev High performance data acquisition with InfiniBand

  4. BNet: SystemC simulations 100x100 traffic shaping no traffic shaping events data meta data S.Linev High performance data acquisition with InfiniBand

  5. InfiniBand • High-performance interconnect technology • switched fabric architecture • up to 20 GBit/s bidirectional serial link • Remote Direct Memory Access (RDMA) • quality of service • zero-copy data transfer • low latency (few microseconds) • Main problem for a long time – lack of common API for different HW vendors • now there is OpenFabrics (OFED) package solves this problem • see www.openfabrics.org for more information S.Linev High performance data acquisition with InfiniBand

  6. OpenFabrics software stack* IP Based App Access Sockets BasedAccess Various MPIs Block Storage Access Clustered DB Access Access to File Systems Application Level Diag Tools Open SM User Level MAD API UDAPL User APIs InfiniBand OpenFabrics User Level Verbs / API iWARP R-NIC User Space SDP Lib Kernel Space Upper Layer Protocol IPoIB SDP SRP iSER RDS NFS-RDMA RPC Cluster File Sys Connection Manager Abstraction (CMA) Mid-Layer SA Client MAD SMA Connection Manager ConnectionManager Kernel bypass Kernel bypass InfiniBand OpenFabrics Kernel Level Verbs / API iWARP R-NIC Provider Hardware Specific Driver Hardware Specific Driver Key Common Apps & AccessMethodsfor usingOF Stack InfiniBand Hardware InfiniBand HCA iWARP R-NIC iWARP * slide from www.openfabrics.org S.Linev High performance data acquisition with InfiniBand

  7. MPI evaluation • MPI – Message Passing Interface, was designed for high performance on both massively parallel machines and on workstation clusters • MVAPICH – MPI & MPI2 over InfiniBand project. Supports: • non-blocking send/receive operation • true hardware-based multicast, but only with blocking API • Tests of data throughput and multicast performance were done. Good results for big (larger than 32K) packets, but difficulty to combine normal and multicast traffic. • Can be as option in DAQ system, while it is supported on majority of modern massively parallel machines, where different interconnect technologies are used. S.Linev High performance data acquisition with InfiniBand

  8. VERBS versus DAPL • Both have very similar functionality and API: • memory, queues, completion events; • message and RDMA transport. • But, VERBS provides extra functionality: • reliable/unreliable data transfer; • multicast support. • We decide to switch to use OFED VERBS while: • it supports full InfiniBand functionality; • it is new official development line for Mellanox products. • The only significant verbs problem – lack of good documentation. S.Linev High performance data acquisition with InfiniBand

  9. Aim:Prove of principal as event building network candidate for CBM Tests last year: GSI cluster - 4 nodes, SDR Forschungszentrum Karlsruhe* (March 2007) – 23 nodes, DDR UNI Mainz** (August 2007) - 110 nodes, DDR InfiniBand basic tests Point-to-point tests Multicast tests * thanks to Frank Schmitz, Ivan Kondov and Project CampusGrid in FZK ** thanks to Klaus Merle and Markus Tacke at the Zentrum für Datenverarbeitung in Uni Mainz S.Linev High performance data acquisition with InfiniBand

  10. Scaling of asynchronous traffic S.Linev High performance data acquisition with InfiniBand

  11. Chaotic (async.) versus scheduled (sync.) S.Linev High performance data acquisition with InfiniBand

  12. DABC and InfiniBand • DABC – Data Acquisition Backbone Core, new software development for general-purpose DAQ in GSI • DABC fully supports InfiniBand as data transport between nodes • connection establishing • memory management for zero-copy transfer • back-pressure • errors detection S.Linev High performance data acquisition with InfiniBand

  13. Conclusion • InfiniBand is a good candidate for CBM event building network • Up to 700 MB/s bidirectional data rate is achieved on 110 nodes cluster • Mixture of point-to-point and multicast traffic is possible • Further investigation of scheduling mechanisms for InfiniBand is required S.Linev High performance data acquisition with InfiniBand

More Related