1 / 19

Computing in CMS

Computing in CMS. May 24, 2002 NorduGrid Helsinki workshop Veikko Karimäki/HIP. Outline. CMS software overview SIMULATION RECONSTRUCTION VISUALISATION ANALYSIS Production activities Schedules of CMS Data Challenge. Simulation-Reconstruction-Analysis Chain (now). CMSIM (soon OSCAR).

ofira
Download Presentation

Computing in CMS

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Computing in CMS May 24, 2002 NorduGrid Helsinki workshop Veikko Karimäki/HIP

  2. Outline • CMS software overview • SIMULATION • RECONSTRUCTION • VISUALISATION • ANALYSIS • Production activities • Schedules of CMS Data Challenge NordUGrid, Helsinki May 23-24, 2002

  3. Simulation-Reconstruction-Analysis Chain (now) CMSIM (soon OSCAR) CMKIN Simulation Generation FZsignal HEPEVTNtuple MC generator Production ORCA Digitization Reconstruction RecReader SimReader G3Reader FZminbias OODBminbias Ntuple OODB Digis FZsignal OODBsignal OODB Tags NordUGrid, Helsinki May 23-24, 2002 User Production

  4. Software Projects • CMSIM(The original GEANT3 Simulation of CMS) • CARF (CMS Analysis and Reconstruction Framework (deprecated) • In 2001 was split out of ORCA repository to become: • COBRA (Coherent Object-oriented Base for simulation, Reconstruction and Analysis) • OSCAR (Object oriented Simulation for CMS Analysis and Reconstruction) • The GEANT4 Simulation framework for CMS • ORCA (Object Reconstruction for CMS Analysis) • The OO reconstruction program • IGUANA (Interactive Graphical User ANAlysis) • Toolkits for Interactive Analysis • FAMOS (Fast Monte-Carlo Simulation) • “Smearing” Monte-Carlo • DDD (needs an Acronym!) • The Detector Description Database See PRS talks for use of CMSIM, CARF/COBRA, ORCA New Packages, IGUANA, OSCAR, FAMOS, DDD, not yet in Physicist use, see following slides NordUGrid, Helsinki May 23-24, 2002

  5. SCRAMToolBox Product specification and versioning, Product break-down structure Configuration definition and versioning SCRAM Configuration specification Assembly break-down Mapping of local resources onto configuration requirements Build, Source-code distribution DAR Self-contained binary distribution (prod) cvs Source code management CVSpm Standardized directory structures and repositories, responsibility and access right information and control. BugsRS Error reporting Insure, workshop Memory leaks performance monitoring, etc.. CodeWizard Coding rule checking Ignominy General dependencies some project metrics McCabe,and risk pages Measurement (OO and procedural metrics) Risk evaluation DepUty Dependencies style checking UML package diagrams Oval Regression testing unit, cluster, acceptance, verification, validation Software Process Components NordUGrid, Helsinki May 23-24, 2002

  6. ORCA Project Relations FAMOS Fast Reco/Simu OSCAR (replaces CMSIM soon) Simulation Reconstruction Visualisation Framework COBRA NordUGrid, Helsinki May 23-24, 2002 Objectivity, Anaphe, Geant4, ...

  7. Tracker: Track Reconstruction Generation of seeds(Seed Generator) • Construction of trajectories for a given seed(Trajectory Builder) • Ambiguity resolution(Trajectory Cleaner) • Final fit of trajectories(Trajectory Smoother) Each component has one or more implementation. Three different algorithms are currently fully implemented: ( Combinatorial Track Finding, Connection Machine, Deterministic Annealing Filter) NordUGrid, Helsinki May 23-24, 2002

  8. IGUANA Detector and Event Display NordUGrid, Helsinki May 23-24, 2002

  9. Tracker - analysis example Resolutions for 100 GeV muons Number of hits used NordUGrid, Helsinki May 23-24, 2002

  10. CMS Distributed Production (~2001) Location 11 RC (Regional Centres) 23 Sites >30 involved persons ~1100 available CPU Europe CERN 1 4 200 Helsinki 1 2 10 IN2P3 1 3 96 INFN 7 10 150 Bristol/RAL 2 1 50 Russia Moscow 4 2 60 USA Caltech 3 2 340 FNAL 1 5 80 Florida 1 2 120 UCSD 1 1 40 Wisconsin 1 2 ? NordUGrid, Helsinki May 23-24, 2002

  11. Production 2002, Complexity Number of Regional Centers 11 Number of Computing Centers 21 Number of CPU’s ~1000 Largest Local Center 176 CPUs Number of Production Passes for each Dataset(including analysis group processing done by production) 6-8 Number of Files ~11,000 Data Size (Not including fz files from Simulation) 15 TB File Transfer by GDMP and by perl Scripts over scp/bbcp NordUGrid, Helsinki May 23-24, 2002

  12. CMS Produced Data in 2001 Simulated Events Reconstructed with pileup Caltech 2.5M FNAL 1.65M CERN 14TB Bristol/RAL 1.27M FNAL 12TB CERN 1.10M Caltech 0.60TB INFN 0.76M Moscow 0.45TB Moscow 0.43M INFN 0.40TB IN2P3 0.31M Bristol/RAL 0.22TB Helsinki 0.13M UCSD 0.20TB Wisconsin 0.07M IN2P3 0.10TB UCSD 0.06M Florida 0.08TB Florida 0.05M Wisconsin 0.05TB • TYPICAL EVENT SIZES • Simulated • 1 CMSIM event= 1 OOHit event = 1.4 MB Reconstructed • 1 “1033” event = 1.2 MB • 1 “2x1033” event = 1.6 MB • 1 “1034” event = 5.6 MB NordUGrid, Helsinki May 23-24, 2002

  13. Production Status 2002 Bristol/RAL Caltech CERN Florida FNAL Imperial College INFN IN2P3 Moscow UCSD Wisconsin On schedule for June 1 deadline NordUGrid, Helsinki May 23-24, 2002

  14. Data Transfers HIP Caltech Bristol/RAL IN2P3 FNAL Moscow Min.Bias Objy/DB Wisconsin UCSD CERN .fz files Objy/DB UFL RC archiving data INFN RC publishing data NordUGrid, Helsinki May 23-24, 2002

  15. CMS and the GRID • CMS Grid Implementation plan for 2002 published (CMS NOTE-2002/015) • Close collaboration with EDG and Griphyn/iVDGL,PPDG • Upcoming CMS GRID/Production Workshop (June CMSweek) • File Transfers • Production File Transfer Software Experiences • Production File Transfer Hardware Status & Reports • Future Evolution of File Transfer Tools • Production Tools • Monte Carlo Production System Architecture • Experiences with Tools • Monitoring / Deployment Planning • Experiences with Grid Monitoring Tools • Towards a Rational System for Tool Deployment NordUGrid, Helsinki May 23-24, 2002

  16. CMS - Schedule for Challenge Ramp Up • All CMS work to date with Objectivity,Now being phased out to be replaced with LCG Software • Enforced lull in production challenges • No point to do work to optimize a solution being replaced • (But much learnt in past challenges to influence new design) • Use Challenge time in 2002 to benchmark current performance • Aim to start testing new system as it becomes available • Target early 2003 for first realistic tests • Thereafter return to roughly exponential complexity ramp up to reach 50% complexity in 2005 • 20% Data Challenge NordUGrid, Helsinki May 23-24, 2002

  17. Objectivity Issues • Bleak • CERN has not renewed the Objectivity Maintenance • Old licenses are still applicable, but cannot be migrated to new hardware • Our understanding is that we can continue to use the product as before, clearly without support any longer • But cannot be used on newer RedHat OS’s (7…) (or other Linux OS’s) • Will become increasingly difficult during this year to find sufficient resources correctly configured for our Objectivity usage. • We are preparing for the demise of our Objectivity-based code by the end of this year • CMS already contributing to the new LCG Software • Aiming to have first prototypes for catalog layer by July • Initial release of CMS prototype ROOT+LCG, September NordUGrid, Helsinki May 23-24, 2002

  18. Planning - CMS Computing • 2002: DAQ Technical Design Report • 2003: GEANT4 validation, 5% Data Challenge start • 2004 beg: 5% Data Challenge Complete • 2004 end: Computing and Core Software (CCS) TDR submitted • 2004-2005: Physics TDR • 2005: 20% Data Challenge • 2006 beg: 20% Data Challenge Complete • 2006: CCS commissioning • 2007 beg: fully operational computing systems (20% capacity) • 2007-2008: CCS systems ramp-up • 2009 beg: CCS systems 100% operational Note: The new LHC schdule caused 9-15 months adjustements in CMS computing planning NordUGrid, Helsinki May 23-24, 2002

  19. Summary • CMSIM/Geant3 (Fortran) to be replaced by OSCAR/Geant • Then the full chain will be in C++ • ODBS: Objectivity --> customized ROOT under work • CMS simulation mass productions well under way • 11 Regional Centres, >1000 CPU’s being used • ~30 TB of data in 2001, 15 TB in 2002 so far • ~8 M events in 2001 + ~25 M MinBias events for pile-up • Active participation in LCG • 5% Data Challenge planned for beg of 2004 • 20% Data Challenge for 2006 • From 9 to 15 months delays due to new LHC schedule NordUGrid, Helsinki May 23-24, 2002

More Related