1 / 11

CMS Computing and Core-Software

CMS Computing and Core-Software. USCMS CB Riverside, May 19, 2001 David Stickland, Princeton University CMS Computing and Core-Software Deputy PM. CCS Core Computing & Software. PRS Physics Reconstruction and Selection. TriDAS Online Software. 1. Computing Centres. 9. Tracker / b-tau.

tacey
Download Presentation

CMS Computing and Core-Software

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CMS Computing and Core-Software USCMS CB Riverside, May 19, 2001 David Stickland, Princeton University CMS Computing and Core-Software Deputy PM

  2. CCS Core Computing & Software PRS Physics Reconstruction and Selection TriDAS Online Software 1. Computing Centres 9. Tracker / b-tau 2. General CMS Computing Services 10. E-gamma / ECAL 7. Online Filter Software Framework 3. Architecture, Frameworks / Toolkits 11. Jets, Etmiss/HCAL 8. Online Farms 4. Software Users and Developers Environment 12. Muons 5. Software Process and Quality 6. Production Processing & Data Management RPROM (Reconstruction Project Management) SPROM (Simulation Project Management) CPROM (Calibration Project Management)…to be created Cafe (CMS Architectural Forum and Evaluation) GPI (Group for Process Improvement)…recently created CPT Project DPS May/15/2001 LHC52

  3. Developing a CCS Project Plan • Build a common planning base for all CPT tasks • Clarify responsibilities • Coordinate milestones • March 2001 planning: (http://cmsdoc.cern.ch/cms/cpt/april01-rrb) • Task Breakdown, Deliverables, Cross-projects • Next: Milestone study • Top Down • Starting from major deliverables • Bottom up • Starting from current project understanding • External Constraints • DAQ TDR, Physics TDR, CCS TDR, Data Challenges, LHC timetable etc Without this it is impossible to measure performance, assign limited resources effectively, identify conflicting constraints etc DPS May/15/2001 LHC52

  4. Computing and Software: Critical Dates • Technical Design Reports • End 2002: DAQ TDR 7M events now, +5M Y2001, +10M Y2002 • End 2003: CCS TDR. Describe system to be implemented • Mid 2004: Physics TDR: GEANT4, All Luminosities, 20+M Events (?) • A Primary Goal: Prepare the collaboration for LHC analysis, shake-down the tools, computing systems, software • End 2005: ~ 20% Computing in place ready for Pilot Run Spring 2006 • Computing milestones • End 2004: 20% Data challenge. Final test before purchase of production systems. • Test of offline, post-DAQ, (Level-2 trigger? Calibrations? Alignments…) • 20Hz for One month (reconstructed, distributed, analyzed) (40M Events) DPS May/15/2001 LHC52

  5. Shown at Nov 2000, LHCC Comprehensive Review Currently reviewing CCS Milestones • Milestone waves • Not easily reviewable • Need more detail • Not tied to deliverables • The work required to satisfy the milestone is typically not described by the milestone so may not be properly monitored or tracked DPS May/15/2001 LHC52

  6. TDR’s and Challenges (Preliminary) DPS May/15/2001 LHC52

  7. Current Computing Activity • Spring 2001: • CERN: 200-300 CPU’s, new Objectivity version, new Tape/MSS system, new Data-servers • Currently (best) 70MB/s out of Objectivity • Testing to determine where next bottleneck is: Disk access, Network, Federation locks… • 1TB output data in 3 days – to be used by ECAL-e/ PRS group • Currently running Calo+Tracker digitization at 10**34 • Will write about 6TB • 200 CPU nodes in single federation • Integrated with CASTOR • Though not as transparently as we plan for next round • Testing ATA/3Ware EIDE Disk systems for data servers (input and output) • Sustained productions achieved • FNAL has responsibility for the JetMET datasets, INFN for the Muon. Continuing to ramp productions, consolidate tools, more automation etc.. DPS May/15/2001 LHC52

  8. 20% Data Challenge Physics TDR CCS TDR DAQ TDR CERN prototype is a time-shared facility available for ~30% of the time at full power for CMS Common Prototypes: CMS Computing, 2002-2004 • Double the complexity (number of boxes) each year to reach 50% of final complexity of a single expt. in 2004, before production system purchasing • Match Computing Challenges with CMS Physics and Detector Milestones DPS May/15/2001 LHC52 Some (~50%) of current T2 prototypes primarily for GRID related R&D. Prototype and final size/cost document: http://cmsdoc.cern.ch/cms/cpt/april01-rrb

  9. Long Term Plan: Computing Ramp-up • Ramp Production systems 05-07 (30%,+30%,+40% of cost each year) • Match Computing power available with LHC luminosity 2007 300M Reco ev/mo 200M Re-Reco ev/mo 50k ev/s Analysis 2006 200M Reco ev/mo 100M Re-Reco ev/mo 30k ev/s Analysis DPS May/15/2001 LHC52

  10. Current most significant risk to the project is insufficient SW manpower • We are making good use of the resources we have and making progress: • OO code is deployed and is the standard for CMS • Worldwide productions • Full use of prototype facilities • Leading to improved code and understanding of limitations • A solid SW Infrastructure base is in place • But there are many things we are unable to cover adequately: • No Calibration infrastructure • No Alignment infrastructure • Detector Description Database only just getting underway • Analysis infrastructure not yet deployed • Slow progress with our GEANT4 implementation • Unable (time!) to answer all the (good) questions the GRID projects are asking us • “Spotty” user-support • Best effort, when time permits • Most of the tasks in SW Quality Assurance and Control are unmanned • Unacceptably high exposure to loss of key people • No backups in any role • Etc etc…. DPS May/15/2001 LHC52

  11. Next Steps • We continue to build a project plan for CCS • We continue to put in place an IMoU for the SW Manpower • In the meantime we focus action to actually get the manpower • We clearly define our prototype requirements • Those Prototypes may be supplied within an IMoU context, or within a broader context of collaboration towards LHC Computing • We try to work with CERN to ensure the experiments and the Regional centers are the driving partners in any new projects and that our real needs are addressed DPS May/15/2001 LHC52

More Related