1 / 24

ATLAS & CMS Online Clouds

ATLAS & CMS Online Clouds. exploit the Experiments ’ online farms for offline activities during the LS1 & beyond. Olivier Chaze (CERN-PH-CMD) & Alessandro Di Girolamo (CERN IT-SDC-OL). ~ 100 m. The L arge H adron C ollider. The experiment data flow. 40 MHz ( 1000 TB/sec).

tawana
Download Presentation

ATLAS & CMS Online Clouds

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. ATLAS & CMS Online Clouds • exploit the Experiments’ online farms • for offline activities • during the LS1 & beyond Olivier Chaze (CERN-PH-CMD) & Alessandro Di Girolamo (CERN IT-SDC-OL)

  2. ~ 100 m The Large Hadron Collider Alessandro Di Girolamo

  3. The experiment data flow 40 MHz (1000 TB/sec) Trigger Level 1 - Special Hardware 75 kHz (75 GB/sec) Trigger Level 2 - Embedded Processors 5 kHz(5 GB/sec) Trigger Level 3 - Farm of commodity CPUs …similar for each experiment... 400 Hz (400 MB/sec) Tier0 (CERN Computing Centre) Data Recording & Offline Analysis Alessandro Di Girolamo

  4. Resources overview • High Level Trigger Experiment farms • ATLAS P1: • 15k cores (28k Hyper Threading, 25% reserved for TDAQ) • CMS P5: • 13k cores (21k Hyper Threading) • when available: • ~50% bigger than the Tier0, • doubling the capacity of biggest Tier1 of the Experiments • Network connectivity to the IT Computing Centre (Tier0) Alessandro Di Girolamo

  5. Why • Experiments always resource hungry: • ATLAS + CMS: more than 250k jobs running in parallel • … exploit all the available resources! Alessandro Di Girolamo

  6. Teamwork • Experts from the Trigger & Data Acquisition teams of the Experiments • Experts from other institutes • BNL RACF, Imperial College … • Experts of WLCG (Worldwide LHC Computing Grid) Alessandro Di Girolamo

  7. Why Cloud? • Cloud as an overlay infrastructure • provides necessary management of VM resources • support & control of physical hosts remain with TDAQ • delegate Grid support •  easy to quickly switch from HLT ↔ Grid • during LS1: periodic full-scale test of TDAQ sw upgrade • can be used in the future also during short LHC stop • OpenStack: common solution, big community! • CMS, ATLAS, BNL, CERN IT…. • sharing experiences • …and support if needed Alessandro Di Girolamo

  8. OpenStack • Glance: VM base image storage and management • Central image repository (and distribution) for Nova • Nova: Central operations controller for hypervisors and VMs • CLI tools, VM scheduler, Compute node client • Network in multi-host mode for CMS • Horizon/ High level control tools • WebUI for Openstack infrastructure/project/VM control (limited use) • RabbitMQ • ATLAS: OpenStack version currently used: Folsom • CMS: OpenStackversion currently used: Grizzly Alessandro Di Girolamo

  9. Network challenges • Avoid any interference with: • Detector Control System operations • internal Control network • Each compute node is connected to two networks • One subnet per rack per network • Routers allow traffic to registered machines only • ATLAS • A new dedicated VLAN has been setup • VMs are registered on this network • CMS • VMs aren’t registered • SNAT rules defined on the hypervisors to bypass network limitations. Alessandro Di Girolamo

  10. CMS online Cloud GRID services CVMFS, Glideins, Condor, Castor/EOS network to IT Computing Centre 20Gb Data network (10.179.0.0/22) Nova network (10.29.0.0/16) br928 br928 SNAT NAT Controller (x4) Compute Node (x 1.3k) Corosync/Pacemaker Corosync/Pacemaker Corosync/Pacemaker Corosync/Pacemaker VM VM MySQL Cluster Dashboard api VM Dashboard api Dashboard api Horizon Nova APIs CMS Site group1 db1 db2 conductor Keystone conductor Keystone Keystone conductor Keystone Nova Scheduler Libvirt/KVM group2 db3 db4 Gateways 10.29.0.1 and 10.29.0.2 Nova Network Glance server Nova Compute mgmt1 mgmt2 NAT NAT NAT Nova Metadata Rabbit MQ ethX ethX ethX GPN network Control network (10.176.0.0/25) Alessandro Di Girolamo

  11. VM image • SL5 x86_64 based (KVM HyperVisor) • Post-boot contextualization method: script injected into the base image (Puppet in the future) • Pre-caching of images on HyperVisors • bzip2 compressed QCOW2 images that are about 350 MB • Experiment specific SW distribution with CVMFS • CVMFS: network file system based on HTTP and optimized to deliver experiment software. Alessandro Di Girolamo

  12. ATLAS: Sim@P1 running jobs: 1 June – 20 Oct Last month dedicated to interventions on the infrastructure 17 k LHC P1 Cooling Intervention ATLAS TDAQ TR August ATLAS TDAQ TR July Sim@P1 Deployment Overall: ~ 55% of time available for Sim@P1 Alessandro Di Girolamo

  13. ATLAS: Sim@P1: Start & Stop 45 min, 6 Hz 17.1k job slots • Restoring Sim@P1: • VMs all up and running within 45min (6Hz) • MC Production jobs flow 0.8 Hz • now improved to almost 1.5 Hz • Shutdown: • 10min (29Hz) the infrastructure is back to TDAQ CPUs Aug 27, 2013 Job flow:0.8 Hz Restoring the VM group from the dormant state Sep 2, 2013 10 min, 29 Hz Ungraceful shutdown of the VM group (rack by rack) 13 21 October 2013 Alessandro Di Girolamo

  14. ATLAS: Sim@P1 completed jobs: 1 June – 20 Oct Overall: ~ 55% of time available for Sim@P1 Total successful jobs: 1.65M Efficiency: 80% Total WallClock: 63.8 G seconds WallClock failed jobs: 10.3% 78% Lost Heartbeat: Intrinsic to the opportunistic nature of resources Comparison with CERN-PROD Total WallClock: 83.3 Gsec WallClock failed jobs: 6% Alessandro Di Girolamo

  15. Conclusions • Experiments online Clouds are a reality • Cloud solution: no impact on data taking, easy switch of activity: • Quick onto them, quick out from them, e.g.: • from 0 to 17k Sim@P1 jobs running in 3.5hours • from 17k Sim@P1 jobs running to TDAQ ready in10 mins • contributing to computing as one big Tier1 or CERN-PROD! • Operations: still a lot of (small) things to do • Integrate OpenStack with the online control to allow dynamic allocation of resources to the Cloud • open questions not unique to the experiments’ online clouds: opportunity to unify solutions to minimize manpower! Alessandro Di Girolamo

  16. BackUp Alessandro Di Girolamo

  17. Sim@P1: Dedicated Network Infrastructure • Sim@P1 VMs will use a dedicated 1 Gbps physical network connecting the P1 rack data switches to the “Castor router” Data Core ATLAS HLT SubFarm Output Nodes (SFOs) 10 Gbps Racks in SDX1 ATCN VLAN isolation P1 Castor Router 1 Gbps CN Ctrl Core 1 Gbps ACLs CN ATLAS: Puppet, REPO GPN Services Control switch Data switch CN 20-80 Gbps ... 1 Gbps 1 Gbps IT Castor Router IT GRID: Condor, Panda EOS/Castor, CvmFS CN Alessandro Di Girolamo

  18. Cloud Infrastructure of Point1 (SDX1) KeystoneHorizon RabbitMQ Cluster Alessandro Di Girolamo

  19. Keystone (2012.2.4) Slide from Alex Zaytsev (BNL RACF) • No issues with stability/performance observed of any scale • Initial configuration of tenant / users / services / endpoints might deserve some higher level automation • Some automatic configuration scripts were already available in 2013Q1 from the third parties, but we found that using the Keystone CLI directly is more convenient & transparent • Simple replication of the keystone MysQL DB works fine for maintaining redundant Keystone instances 6 Dec 2013 Alessandro Di Girolamo

  20. Nova (2012.2.4) Slide from Alex Zaytsev (BNL RACF) • Once bug fix needed to be applied • Back port from Grizzly release:“Handle compute node records with no timestamp”:https://github.com/openstack/nova/commit/fad69df25ffcea2a44cbf3ef636a68863a2d64d9 • The prefix for the VMs’ MAC addresses had to be changed in order to match the range pre-allocated for Sim@P1 project • No configuration option for this, direct patch to Python code was needed • Configuring the server environment for Nova Controller supporting more than 1k hypervisors / 1k of VMs requires rising the default limits for maximum number of open files per several system users • Not documented / handled automatically by Openstack recommended configuration procedures, but pretty straightforward to figure out • RabbitMQ cluster consisting of minimum two nodes was required in order to scale beyond 1k hypervisors per single Nova Controller • RabbitMQ configuration procedure / stability is version sensitive • We had to try several version (currently v3.1.3-1) before achieving a stable cluster configuration • Overall: stable long term operations with only one Cloud controller(plus one hot spare backup instance) for the entire Point 1 Alessandro Di Girolamo

  21. Slide from Alex Zaytsev (BNL RACF) Glance (2012.2.4) • Single Glance instance (provided with single 1 Gbps uplink) works nicely as an central distribution point up to the scale of about 100 hypervisors / 100 VM instances • Scaling beyond that (1.3k hypervisors, 2.1k VM instances) requires either • A dedicated group of cache servers between Glance and hypervisors • Custom made mechanism for pre-deployment of the base images on all compute nodes (multi-level replication) • Since we operate with only one base image at the time which changes rarely (approximately once a month) we built a custom image deployment mechanism, living the central Glance instances with functionality of image repositories, but not the image central distribution points • No additional cache servers needed • We distribute bzip2 compressed QCOW2 images that re only about 350 MB in size • Pre-placement of the new image to all the hypervisors take in total only about 15 minutes despite 1 Gbps network limitations on both Glance instances and on the level of every rack of compute nodes • Snapshot functionality of Glance is used only for making persistent changes in the base image • No changes are saved for VM instances during production operations Alessandro Di Girolamo

  22. Horizon (2012.2.4) Slide from Alex Zaytsev (BNL RACF) • Very early version of the web interface • Many security features are missing • Such as no native HTTPS support • Not currently used for production at Sim@P1 • Several configuration / stability issues encountered • Such as debug mode must be enabled in order for Horizon to function properly • Limited feature set of the web interface • No way to perform non-trivial network configuration purely via web-interface • No way to handle large groups of VMs (1-2k+) in a conveniently, such as to display VM instances in a tree structured according to the configuration of the availability zones / instance names • No convenient way to perform bulk operations on large subgroups of VMs (hundreds) within the production group of VMs consisting of 1-2k • All of these problems, presumably, already addressed in the resent Openstack releases Alessandro Di Girolamo

  23. • … • … • … • … • .. • … • .. • … Alessandro Di Girolamo

  24. • … • … • … • … • .. • … • .. • … Alessandro Di Girolamo

More Related