1 / 21

CERN

CERN. The Grid Project. Fabrizio Gagliardi. CERN. Information Technology Division. May, 2000. F.Gagliardi@cern.ch. TERENA Lisbon. Summary. CERN. High Energy Physics and CERN computing problem. . An excellent computing model: the GRID. . The Data Grid Initiative. .

luciam
Download Presentation

CERN

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CERN The Grid Project Fabrizio Gagliardi CERN Information Technology Division May, 2000 F.Gagliardi@cern.ch TERENA Lisbon

  2. Summary CERN High Energy Physics and CERN computing problem  An excellent computing model: the GRID  The Data Grid Initiative   (http://www.cern.ch/grid/) TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 2

  3. CERN organization CERN European International Center for Particle Physics Research Largest Particle Physics lab in the world Budget: 1020 M CHF 2700 staff 7000 physicist users TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 3

  4. The LHC Detectors CMS CERN ATLAS 3.5 PetaBytes / year ~108events/year LHCb TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 4

  5. The HEP Problem - Part I CERN The scale... TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 5

  6. CERN Estimated CPU Capacity at CERN 2,500 2,000 1,500 K SI95 Non-LHC 1,000 ~10K SI95 1200 processors LHC 500 0 1998 1999 2000 2001 2002 2003 2004 2005 2006 year technology-price curve (40% annual price improvement) Capacity that can purchased for the value of the equipment present in 2000 TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 6

  7. CERN Estimated DISK Capacity ay CERN 1800 1600 1400 1200 TeraBytes 1000 Non-LHC 800 LHC 600 400 200 0 1998 1999 2000 2001 2002 2003 2004 2005 2006 year technology-price curve (40% annual price improvement) TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 7

  8. Long Term Tape Storage Estimates CERN 14'000 12'000 10'000 TeraBytes 8'000 LHC 6'000 4'000 Current Experiments COMPASS 2'000 0 1996 2006 2000 2003 2005 1995 1997 1998 1999 2001 2002 2004 Year TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 8

  9. HPC or HTC CERN High Throughput Computing  mass of modest, independent problems  computing in parallel – not parallel computing  throughput rather than single-program performance  resilience rather than total system reliability Have learned to exploit inexpensive mass market components But we need to marry these with inexpensive highly scalable management tools Much in common with other sciences (see EU-US Annapolis Workshop at www.cacr.caltech.edu/euus): Astronomy, Earth Observation, Bioinformatics, and commercial/industrial: data mining, Internet computing, e-commerce facilities, …… TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 9

  10. Generic component model of a computing farm network servers application servers tape servers disk servers

  11. The HEP Problem - Part II CERN Geography, Sociology, Funding and Politics... TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 11

  12. World Wide Collaboration distributed computing & storage capacity CERN CMS: 1800 physicists 150 institutes 32 countries TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 12

  13. Regional Centres - a Multi-Tier Model CERN CERN – Tier 0 IN2P3 RAL Tier 1 FNAL Uni n Lab a Tier2 Uni b Lab c Department Desktop MONARC report: http://home.cern.ch/~barone/monarc/RCArchitecture.html TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 13

  14. Are Grids a solution? CERN Change of orientation of US Meta-computing activity  From inter-connected super-computers … .. towards a more general concept of a computational Grid (The Grid – Ian Foster, Carl Kesselman)  Has initiated a flurry of activity in HEP  US – Particle Physics Data Grid (PPDG)  GriPhyN – data grid proposal submitted to NSF  Grid technology evaluation project in INFN  UK proposal for funding for a prototype grid  NASA Information Processing Grid  TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 14

  15. The Grid CERN “Dependable, consistent, pervasive access to [high-end] resources” • Dependable: • provides performance and functionality guarantees • Consistent: • uniform interfaces to a wide variety of resources • Pervasive: • ability to “plug in” from anywhere TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 15

  16. R&D required CERN Local fabric  Management of giant computing fabrics  auto-installation, configuration management, resilience, self- healing  Mass storage management  multi-PetaByte data storage, “real-time” data recording requirement, active tape layer – 1,000s of users Wide-area - building on an existing framework & RN (e.g.Globus, Geant)  workload management  no central status  local access policies  data management  caching, replication, synchronisation  object database model  application monitoring TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 16

  17. HEP Data Grid Initiative CERN European level coordination of national initiatives & projects Principal goals:  Middleware for fabric & Grid management  Large scale testbed - major fraction of one LHC experiment  Production quality HEP demonstrations  “mock data”, simulation analysis, current experiments  Other science demonstrations Three year phased developments & demos Complementary to other GRID projects  EuroGrid: Uniform access to parallel supercomputing resources Synergy to be developed (GRID Forum, Industry and Research Forum)      TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 17

  18. Participants CERN Main partners: CERN, INFN(I), CNRS(F), PPARC(UK), NIKEF(NL), ESA-Earth Observation Other sciences: KNMI(NL), Biology, Medicine Industrial participation: CS SI/F, DataMat/I, IBM/UK Associated partners: Czech Republic, Finland, Germany, Hungary, Spain, Sweden (mostly computer scientists) Formal collaboration with USA Industry and Research Project Forum with representatives from:  Denmark, Greece, Israel, Japan, Norway, Poland, Portugal, Russia       TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 18

  19. Status CERN Prototype work already started at CERN and in most of collaborating institutes  Proposal to RN2 submitted  Network requirements discussed with Dante/Geant  TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 19

  20. WAN Requirements CERN  High bandwidth from CERN to Tier 1 centres (5-6)  VPN, Quality of Service  Guaranteed performance during limited test periods and at the end of the project for production quality services  Target requirements (2003) 2.5 Gb/s + 622 Mb/s + 155 Mb/s  Could saturate for limited amount of test time 2.5 Gb/s (100 MB/s out from a 100 PC farm, we plan for 1000’s PC farm)  Reliability is an important factor:  from WEB client-server model to  GRID peer distributed computing model TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 20

  21. Conclusions CERN This project, motivated by HEP and other high data and computing demanding sciences, will contribute to develop and implement a new world-wide distributed computing model: The GRID  An ideal computing model for the next generation Internet  An excellent test case for the next generation of high- performance research networks  TERENA Lisbon F. Gagliardi - CERN/IT-May-2000 21

More Related