1 / 22

WLCG Service Schedule

WLCG Service Schedule. Jamie.Shiers@cern.ch June 2007. Agenda. The machine The experiments The service. 12. 23. 34. 45. 56. 67. 78. 81. Consolidation. Operation testing of available sectors.

koen
Download Presentation

WLCG Service Schedule

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. WLCG Service Schedule Jamie.Shiers@cern.ch June 2007

  2. Agenda • The machine • The experiments • The service

  3. 12 23 34 45 56 67 78 81 Consolidation Operationtesting of availablesectors Machine Checkout BeamCommissioning to 7 TeV Interconnection of the continuous cryostat Global pressure test &Consolidation Warm up Leak tests of the last sub-sectors Flushing Powering Tests Inner Triplets repairs & interconnections Cool-down LHC Schedule Mar. Mar. Apr. Apr. May May Jun. Jun. Jul. Jul. Aug. Aug. Sep. Sep. Oct. Oct. Nov. Nov. Dec. Dec. . Jan. Jan. Feb. Feb. Mar. Mar. Apr. Apr. May May Jun. Jun. Jul. Jul. Aug. Aug. Sep. Sep. Oct. Oct. Nov. Nov. Dec. Dec. . LHC commissioning - CMS June 07

  4. 2008 LHC Accelerator schedule LHC commissioning - CMS June 07

  5. 2008 LHC Accelerator schedule LHC commissioning - CMS June 07

  6. Machine Summary • No engineering run in 2007 • Startup in May 2008 • “…we aim to be seeing high energy collisions by the summer.”

  7. Experiments • Continue preparations for ‘Full Dress Rehearsals’ • Schedule from CMS is very clear: • CSA07 runs September 10 for 30 days • Ready for cosmics run in November • Another such run in March • ALICE have stated FDR from November • Bottom line: continuous activity – post CHEP likely to be (very) busy

  8. Event sizes • We already needed more hardware in the T0 because • In the TDR there was no full ESD copy to BNL included • Transfers require more disk servers than expected  10% less disk space in CAF • From TDR: RAW=1.6 MB, ESD=0.5 MB, AOD=0.1 MB • 5-day buffer at CERN  127 TByte • Currently 50 disk servers  300 TByte OK for buffer • For Release 13: RAW=1.6 MB, ESD=1.5 MB, AOD=0.23 MB (incl. trigger&truth) • 2.2  3.3 MB = 50% more at T0 • 3 ESD, 10 AOD copies: 4.1  8.4 MB = factor 2 more for exports • More disk servers needed for T0 internal and exports  40% less disk in CAF • Extra tapes and drives  25% cost increase • Have to be taken away from CAF again • Also implications for T1/2 sites • Can store 50% less data • Goal: run this summer 2 weeks uninterrupted at nominal rates with all T1 sites Software & Computing Workshop

  9. ATLAS T0T1 Exportssituation at May 28/29 2007 Software & Computing Workshop

  10. Services • Q: What do you (CMS) need for CSA07? • A: Nothing – would like FTS 2.0 at Tier1s (and not too late) but not required for CSA07 to succeed • Other major ‘residual service’: SRM v2.2 • Windows of opportunity: post CSA07, early 2008 • No long shutdown end 2008

  11. S.W.O.T. Analysis of WLCG Services

  12. 4x 6x Steep ramp-up still needed before first physics run Evolution of installed capacity from April 06 to June 07 Target capacity from MoU pledges for 2007 (due July07) and 2008 (due April 08)

  13. WLCG Service: S / M / L vision • Short-term: ready for Full Dress Rehearsals – now expected to fully ramp-up ~mid-September (>CHEP) • The only thing I see as realistic on this time-frame is FTS 2.0 services at WLCG Tier0 & Tier1s • Schedule: June 18th at CERN; available mid-July for Tier1s • Medium-term: what is needed & possible for 2008 LHC data taking & processing • The remaining ‘residual services’ must be in full production mode early Q1 2008 at all WLCG sites! • Significant improvements in monitoring, reporting, logging  more timely error response  service improvements • Long-term: anything else • The famous ‘sustainable e-Infrastructure’… ? WLCG Service Deployment – Lessons Learnt

  14. WLCG Service Deployment – Lessons Learnt

  15. Types of Intervention 0. (Transparent) – load balanced servers / (ices) 1. Infrastructure: power, cooling, network 2. Storage services: CASTOR, dCache 3. Interaction with backend DB: LFC, FTS, VOMS, SAM etc.

  16. Transparent Interventions - Definition • Have reached agreement with the LCG VOs that the combination of hardware / middleware / experiment-ware should be resilient to service “glitches” • A glitch is defined as a short interruption of (one component of) the service that can be hidden – at least to batch – behind some retry mechanism(s) • How long is a glitch? • All central CERN services are covered for power ‘glitches’ of up to 10 minutes • Some are also covered for longer by diesel UPS but any non-trivial service seen by the users is only covered for 10’ • Can we implement the services so that ~all interventions are ‘transparent’? • YES – with some provisos to be continued… EGI Preparation Meeting, Munich, March 19 2007 - Jamie.Shiers@cern.ch

  17. More Transparent Interventions • I am preparing to restart our SRM server here at IN2P3-CC so I have closed the IN2P3 channel on prod-fts-ws in order to drain current transfer queues. • I will open them in 1 hour or 2. • Is this a transparent intervention or an unscheduled one? • A: technically unscheduled, since it's SRM downtime. • An EGEE broadcast was made, but this is just an example… • But if the channel was first paused – which would mean that no files will fail – it becomes instead transparent – at least to the FTS – which is explicitly listed as a separate service in the WLCG MoU, both for T0 & T1! • i.e. if we can trivially limit the impact of an intervention, we should(c.f. WLCG MoU services at Tier0/Tier1s/Tier2s) WLCG Service Deployment – Lessons Learnt

  18. WLCG Service Deployment – Lessons Learnt

  19. Summary • 2008 / 2009 LHC running will be lower than design luminosity (but same data rate?) • Work has (re-)started with CMS to jointly address ‘critical services’ • Realistically, it will take quite some effort and time to get services up to ‘design luminosity’

  20. Service Progress Summary

  21. Service Progress Summary

More Related