1 / 7

Data management demonstrators

WLCG MB 18 th January 2011. Data management demonstrators. Ian Bird; . Reminder. March 2010 Experiments express concern over Data management and access First brainstorming … agree on Jamboree June 2010 Amsterdam Jamboree ~ 15 demonstrator projects proposed Process:

noah
Download Presentation

Data management demonstrators

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. WLCG MB 18th January 2011 Data management demonstrators Ian Bird;

  2. Reminder • March 2010 • Experiments express concern over Data management and access • First brainstorming … agree on Jamboree • June 2010 • Amsterdam Jamboree • ~15 demonstrator projects proposed • Process: • Follow up at WLCG meeting – ensure projects had effort and interest • Follow up in GDB’s • By end of year – decide which would continue based on demonstrated usefulness/feasibility • 2nd half 2010 • Initial follow up at WLCG workshop • GDB status report • Jan 2011 • GDB status report (last week) • Close of process started in Amsterdam (today) Ian.Bird@cern.ch

  3. My Summary of demonstrators • 12 projects scheduled at GDB • 2 had no progress reported (CDN + Cassandra/Fuse) • 10 either driven, or interest expressed, by experiments • Assume these 10 will progress and be regularly reported on in GDB • Scope for collaboration between several • To be encouraged/pushed • Several using xrootd technology • Must ensure we arrange adequate support • Which (and how) to be wrapped into WLCG sw distributions? • Process and initiatives have been very useful • MB endorses continued progress on these 10 projects Ian.Bird@cern.ch

  4. Summary … 1 • ATLAS PD2P • In use; linked to LST demonstrator • Implementation is ATLAS-specific, but ideas can be re-used with other central task-queues • ARC Caching • Use to improve ATLAS use of ARC sites, could also help others use ARC • More general use of cache – needs input from developers (and interest/need from elsewhere) • Speed-up of SRM getturl (make sync) • Essentially done, but important mainly if lots of small files (then other measures should be taken too) • Cat/SE sync + ACL propagation with MSG • Prototype exists • Interest in testing from ATLAS • Ideas for other uses • CHIRP • Seems to work well – use case of personal SE (grid home dir) • Used by ATLAS, tested by CMS Ian.Bird@cern.ch

  5. Summary … 2 • Xrootd-related • Xrootd (EOS, LST) • Well advanced, tested by ATLAS and CMS • Strategy for Castor evolution at CERN • Xrootd – ATLAS • Augments DDM • Commonality with CMS • Xrootd-global – CMS • Global xrootd federation, integrates with local SE’s and FS’s • Many commonalities – can we converge on a common set of tools?? • Proxy-caches in root • Requires validation before production • Continue to study file caching in experiment frameworks • NFS4.1 • Lot of progress in implementations and testing • Needs some xrootd support (CMSD) • Should MB push for pNFS kernel in SL? Ian.Bird@cern.ch

  6. Post – Amsterdam • Amsterdam was not only the demonstrators • Was a recognition that the network was a resource • Could use remote access • Should not rely on 100% accuracy of catalogues, etc. • Can use network to access remote services • Network planning group set up • Work with NRENs etc is ongoing (they got our message) • Also understood where data management model should change • Separate tape and disk caches (logically at least) • Access to disk caches does not need SRM • SRM for “tape” can be minimal functionality • Disk to be treated as a cache – move away from data placement for analysis at least • Re-think “FTS” • Etc. Ian.Bird@cern.ch

  7. Conclusions • We have changed direction • There are a number of very active efforts – driven by experiment need/interests for future • Not just what was in the demonstrators • Should continue to monitor and support – and look for commonalities: opportunity to reduce duplication and improve support efforts Ian.Bird@cern.ch

More Related