1 / 16

GriPhyN/PPDG Virtual Data Grid Architecture

GriPhyN/PPDG Virtual Data Grid Architecture. Ian Foster Mathematics and Computer Science Division Argonne National Laboratory Department of Computer Science The University of Chicago. Overview. Recap of motivation for architecture effort Progress since April 2001

moya
Download Presentation

GriPhyN/PPDG Virtual Data Grid Architecture

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. GriPhyN/PPDGVirtual Data Grid Architecture Ian Foster Mathematics and Computer Science Division Argonne National Laboratory Department of Computer Science The University of Chicago

  2. Overview • Recap of motivation for architecture effort • Progress since April 2001 • Architecture as a framework for collaboration • Web services and databases

  3. Motivation • We need an architecture so that we can • Coordinate our own activities • Coordinate with other Data Grid projects • Explain to others (experiments, NSF, CS community) what we are doing • An architecture must: • Facilitate CS research activities by simplifying evaluation of alternatives • Not preclude experimentation with (radically) alternative approaches

  4. Data Grid Architecture • Goal: Define requirements and principal elements of a Data Grid system • Working bottom up, identify elements for which we have solid base of experience and code • Working top down, identify and prioritize higher-level elements

  5. Production Team Individual Investigator Other Users Interactive User Tools Request Planning and Request Execution Virtual Data Tools Scheduling Tools Management Tools Performance Estimation and Evaluation Resource Security and Other Grid Resource Security and Other Grid Management Policy Services Management Policy Services Services Services Services Services Transforms Raw data Distributed resources source (code, storage, computers, and network) Primary GriPhyN R&D Components

  6. MCAT; GriPhyN catalogs MDS MDS GDMP DAGMAN, Kangaroo GSI, CAS Globus GRAM GridFTP; GRAM; SRM Data Grid Architecture Application DAG Catalog Services Monitoring Planner Info Services DAG Repl. Mgmt. Executor Policy/Security Reliable Transfer Service Compute Resource Storage Resource = some [perhaps partial] solution exists

  7. Progress Since April 2001 • Major architecture revisions and extensions • E.g., virtual data catalog & scheduling tools • Establishment of collaboration with PPDG • Architecture document now joint • Adoption by expts of DGA/VDT components • E.g., DAGMAN, GridFTP • New projects instantiating missing pieces • E.g., Giggle RLS, DAGMAN extensions, VDL • v2.0 released; significant progress on v3.0

  8. Transparency wrt materialization Derived Metadata Catalog App-specific-attr id … App specific attr . id … … … i2,i10 i2,i10 id id … … … … Update upon materialization Derived Data Catalog Id Trans F Id Trans Param Param Name … Name … GCMS i1 F X F.X … i1 F X F.X … i2 F Y F.Y … i2 F Y F.Y … i10 G Y P i10 G Y P G(P).Y … G(P).Y … Trans. name Trans. name Transformation Catalog Trans Trans Prog Prog Cost … Cost … F URL:f 10 … F URL:f 10 … G URL:g 20 … G URL:g 20 … URLs for program location URLs for program location Program storage Program storage Catalog Architecture Transparency wrt location Metadata Catalog Metadata Catalog Name Name LObjN LObjN … … … X logO1 … … Y logO2 … … F.X F.X logO3 … … logO3 … … G(1).Y logO4 … … Object Name Object Name GCMS GCMS Logical Container Name Replica Catalog Replica Catalog LCN LCN PFNs PFNs … … logC1 URL1 logC1 URL1 logC2 URL2 URL3 logC2 URL2 URL3 logC3 URL4 logC3 URL4 logC4 URL5 URL6 logC4 URL5 URL6 URLs for physical file location Physical file storage

  9. Executor • Requirements • Reliable management of the execution of a set of computations and data movements • Current status • UW DAGMan, which executes a supplied directed acyclic graph (DAG) • Future directions • Error handling and recovery • Ability to express alternative strategies • Beyond DAGs

  10. Adoption by Experiments: E.g.,CMS Data Reconstruction 2) Launch secondary job on WI pool; input files via Globus GASS Master Condor job running at Caltech Secondary Condor job on WI pool 5) Secondary reports complete to master Caltech workstation 6) Master starts reconstruction jobs via Globus jobmanager on cluster 3) 100 Monte Carlo jobs on Wisconsin Condor pool 9) Reconstruction job reports complete to master 4) 100 data files transferred via GridFTP, ~ 1 GB each 7) GridFTP fetches data from UniTree NCSA Linux cluster NCSA UniTree - GridFTP-enabled FTP server 8) Processed objectivity database stored to UniTree Scott Koranda, Miron Livny, others

  11. Pre / Simulation Jobs / Post (UW Condor) ooDigis at NCSA ooHits at NCSA Delay due to script error Trace of a Condor-G Physics Run

  12. New Projects; Architecture as a Roadmap for Collaboration • “Giggle” replica location service • Joint design, implementation, & evaluation effort with CERN EU DataGrid team • Replica management, other higher-level data management tools • Joint work with elements of PPDG and EU DataGrid • Monitoring and discovery requirements • Joint with PPDG, iVDGL • DAGMAN extensions and next steps

  13. Web Services and Databases • Q: How does DGA relate to • Web services? • WSDL as interface definition language, SOAP as transport • Other services for discovery, workflow, etc. • Databases? • Importance for HEP a topic of debate, but clearly critical for many disciplines • In both cases, significant industrial adoption • A: We have a plan, and partnerships in place to execute this plan during 2002 • Only partially GriPhyN/DGA-specific

  14. The Basis for the Plan:Open Grid Services Architecture • Grid Service = Web service with specified codified behaviors and interfaces, e.g. • Global naming; reliable, secure creation/mgmt of transient, stateful remote service instances • Standard factory, discovery, etc., interfaces • Roll-out planned during 2002 • Evolution of existing protocols, e.g. • GRAM-2: Remote management protocol • GSI: Restricted delegation, etc. • New protocols & services: e.g., databases

  15. OGSA and DGA • We are relying on various OGSA components for future VDT components • E.g., GRAM-2 for reservation, restricted delegation for security • These will arrive during 2002 • In the meantime, we are starting to package selected services as Grid services • E.g., reliable data transfer (with LBNL), replica location (with CERN), etc.

  16. Next Steps • V3 to be released first quarter 2002 • After extensive review by GriPhyN, PPDG • Integrate all latest thinking • Use this • To drive R&D for 2002-2003 • As a basis for discussions with EU DataGrid, LHC Grid, and TeraGrid projects • Initiate new efforts • Discovery and monitoring • Virtual data catalogs and tools

More Related