1 / 16

Workload Management Workpackage

Workload Management Workpackage. Massimo Sgaravatto INFN Padova. Overview. Goal: define and implement a suitable architecture for distributed scheduling and resource management in a GRID environment Large heterogeneous environment PC farms and not supercomputers used in HEP

peigi
Download Presentation

Workload Management Workpackage

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Workload ManagementWorkpackage Massimo Sgaravatto INFN Padova

  2. Overview • Goal: define and implement a suitable architecture for distributed scheduling and resource management in a GRID environment • Large heterogeneous environment • PC farms and not supercomputers used in HEP • Large numbers (thousands) of independent users in many different sites • Different applications with different requirements • HEP Monte Carlo productions, reconstructions and production analyses • “Scheduled” activities • Goal: throughput maximization • HEP individual physics analyses • “Chaotic”, non-predictable activities • Goal: latency minimization • …

  3. Overview • Many challenging issues : • Optimizing the choice of execution location based on the availability of data, computation and network resources • Optimal co-allocation and advance reservation of CPU, data, network • Uniform interface to different local resource management systems • Priorities, policies on resource usage • Reliability • Fault tolerance • Scalability • … • INFN responsibility in DataGrid

  4. Tasks • Job resource specification and job description • Method to define and publish the resources required by a job • Job control language (command line tool, API, GUI) • Partitioning programs for parallel execution • “Decomposition” of single jobs in multiple, “smaller” jobs that can be executed in parallel • Exploitation of task and data parallelism

  5. Tasks • Scheduling • Definition and implementation of scheduling policies to find the best match between job requirements and available resources • Co-allocation and advance reservation • Resource management • Services • Authentication, authorization, bookkeeping, accounting, logging,

  6. Effort breakdown (mm)

  7. Italian participation • Bologna (INFN) • Catania (INFN and University) • CNAF (INFN and University of Bologna) • Lecce (INFN and University) • Milano (INFN) • Padova (INFN and University of Venezia) • Roma (INFN) • Torino (INFN and University)

  8. Workload Management in the INFN-GRID project • Integration, adaptation and deployment of middleware developed within the DataGrid project • GRID software must enable physicists to run their jobs using all the available GRID resources in a “transparent” way • HEP applications classified in 3 different “classes”, with incremental level of complexity • Workload management system for Monte Carlo productions • Goal: throughput maximization • Implementation strategy: code migration (moving the application where the processing will be performed) • Workload management system for data reconstruction and production analysis • Goal: throughput maximization • Implementation strategy: code migration + data migration (moving the data where the processing will be performed, and collecting the outputs in a central repository) • Workload management system for individual physics analysis • “Chaotic” processing • Goal: latency minimization • Implementation strategy: code migration + data migration + remote data access (accessing data remotely) for client/server applications

  9. First Activities and Results • CMS-HLT use case (Monte Carlo production and reconstruction) analyzed in terms of GRID requirements and GRID tools availability • Discussions with Globus team and Condor team • Good and productive collaborations already in place • Definition of a possible high throughput workload management system architecture • Use of Globus and Condor mechanisms • But major developments needed

  10. High throughput workload management system architecture (simplified design) Other info Resource Discovery Submit jobs (using Class-Ads) Master Grid Information Service (GIS) condor_submit (Globus Universe) Master chooses in which Globus resources the jobs must be submitted Information on characteristics and status of local resources Condor-G Condor-G able to provide reliability Use of Condor tools for job monitoring, logging, … globusrun Globus GRAM as uniform interface to different local resource management systems Globus GRAM Globus GRAM Globus GRAM Local Resource Management Systems CONDOR LSF PBS Site1 Farms Site2 Site3

  11. First Activities and Results • On going activities in putting together the various building blocks • Globus deployment • INFNGRID distribution toolkit to make Globus deployment easier and more automatic • INFN customizations • Evaluation of Globus GRAM • Tests with job submissions on remote resources • Globus GRAM as uniform interface to different underlying resource management systems (LSF, Condor, PBS) • Evaluation of Globus RSL as uniform language to describe resources • “Cooperation” between GRAM and GIS

  12. First Activities and Results • Evaluation of Condor-G • It works, but some problems must be fixed: • Very difficult to understand about errors • Problems with log files • Problems with scalability in the submitting machine • Condor-G is not able to provide fault tolerance and robustness (because Globus doesn’t provide these features) • Fault tolerance only in the submitting side • Condor team is already working to fix some of these problems • They are also implementing a new Globus jobmanager

  13. First activities and results • Tests with a real CMS MC production • Real applications (Pythia) • Real production environments • Jobs submitted from Padova using Condor-G and executed in Bologna and Pisa • Many many memory leaks found in the Globus jobmanager !!! • Fixes provided by Francesco Prelz (INFN Milano)

  14. Test layout for CMS production Submit jobs Production manager (Ivano Lippi – Padova) condor_submit (Globus Universe) Condor-G Padova globusrun Globus GRAM Globus GRAM Local Resource Management Systems CONDOR LSF Farms Bologna Pisa

  15. Some next steps • Evaluation of the new Globus jobmanager and the new Condor-G implementations (when ready) • Master development !!!

  16. Other info • http://www.infn.it/grid

More Related