1 / 30

Application of the EDG Testbed

International HEP DataGrid Workshop 9 Nov 2002. Application of the EDG Testbed. Bockjoo Kim *, Soo-Bong Kim Seoul National University (SNU) Kihyeon Cho, Youngdo Oh, Dongchul Son Center for High Energy Physics (CHEP) Kyungpook National University Jysoo Lee KISTI, Supercomputing Center.

shilah
Download Presentation

Application of the EDG Testbed

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. International HEP DataGrid Workshop 9 Nov 2002 Application of the EDG Testbed Bockjoo Kim*, Soo-Bong Kim Seoul National University (SNU) Kihyeon Cho, Youngdo Oh, Dongchul Son Center for High Energy Physics (CHEP) Kyungpook National University Jysoo Lee KISTI, Supercomputing Center

  2. Outline • LHC Computing Challenge • European DataGrid • European DataGrid (EDG) testbed in Korea • The EDG testbed and tests • The GDMP (Datamanagement) and tests • Possible Immediate Application of the EDG Testbed • Summary

  3. LHC HEP computing, key parameters • All LHC experiments at CERN: • 10 Peta Bytes/yr data storage; disk: 2 P Byte • Multi-experiment Tier 1: • 3 Peta Byte/yr; disk: 0.5 P Byte • Tier 0 & 1 at CERN: 2 M SI 95 (PC today ~ 20SI95) • Multi-experiment Tier 1: 0.9 M SI 95 • Networking Tier 0 --> Tier 1: 622 Mbps (4 Gbps) (black fibre: 1 Tbps today)

  4. Online Data Rate vs Size High No. ChannelsHigh Bandwidth(500 Gbit/s) Level 1 Rate (Hz) 106 LHCB ATLAS CMS 105 HERA-B “How can this data reach the end user?” KLOE CDF II 104 High Data Archive(PetaByte) CDF It doesn’t… Factor O(1000) Online data reduction via trigger selection 103 H1ZEUS ALICE NA49 UA1 102 104 106 105 107 LEP Event Size (bytes)

  5. European DataGrid (EDG) • To cope with the data challenges in LHC era • Participants • EC+(Korea+U.S.+Japan+China…) • Operational at EC • WP1: workload • WP2: data management • WP3: grid monitoring services • WP4: fabric management • WP5: mass storage management • WP6: Integration and testbed • WP7: network services • WP8: HEP application • WP9: Earth Observation • WP10:Biology • WP11:Dissemination • WP12:Management Specific app layer LHCb Other apps ALICE ATLAS CMS VO common application layer LHC Other apps High level GRID middleware EDG middleware Basic Services (Globus) GLOBUS 2.0 OS & Net services Concept of EDG

  6. RPMs repository Profile repository The EDG Fabric and Testbed • LCFG : EDG software + Profile + CE, WN, SE, RB, UI installation&conf • UI : User Interface : Job submission interface for authorized users • CE : Farms of commodity PCs • Globus Resource Access Management (GRAM) service+GSI • RB : Resource Broker : Core of job scheduling • SE : Any storage system: accessible via ftp, nfs, etc & GSI • WN : Worker CPU hidden underneath CE ~ clusters An EDGTestbed CE/WN (PC Cluster) Listens update LCFG Server Profiles UI SE (GDMP) RB All the machines are linux red hat

  7. EDG Important Ingredients VO Management Security • X.509 Authentication + Users certs on LDAP • Grid Information Index Service (GIIS): not fully used. RB does the job • Virtual Organization(VO) : CMS, Atlas, etc • Virtual Users and Groups for VO • LDAP services for VO, authentication list with edg schema • Replica Catalog (RC) uses also LDAP server with edg schema • Replica Management uses GDMP (grid data mirroring pkg.) Resource Locating Data Management

  8. EDG Testbed in Korea RB UI Real user In operation 빅 팻 디스크 Big Fat Disk NFS GSIFTP GDMP client (with new VO) KNU SE VOuser CE 디스크 VOuser LDAP Server SNU NFS GSIFTP GDMP server (with new VO) NFS GDMP client (with new VO) GSIFTP MAP on disk With maximum security In preparation NFS WN grid-security VOuser GSIFTP SKKU SNU In operation . . . • The EDG testbeds are installed at KNU and at SNU • The globus simple CA is managed at KNU and at SNU to sign certs • In addition to the default VO’s in EDG, a cdf VO is constructed • Authentication and VO LDAP server is running at SNU • A replica catalog (RC) LDAP server is running at SNU • See the Korean EDG testbed description web site at • http://neutrino.snu.ac.kr/~bockjoo/EDG_testbed/ and EDG web

  9. o=xyz,dc=edg, dc=org o=testbed,dc=edg, dc=org ou=People ou=People ou=tb1 ou=Admin CN=John Smith CN=Mario Rossi CN=John Smith Authentication Certificate Authentication Certificate Authentication Certificate CN=Franz Elmer CN=Franz Elmer mkgridmap grid-mapfile ban list local users Authorization and grid-mapfile • X.509 certs by CA • CA published on LDAP • Check Auth. dir and VO VODirectory “AuthorizationDirectory”

  10. LDAP Services for Korean HEP DG Auth LDAP CDF VO LDAP There are also the CMS and the CDF RC More Info: http://neutrino.snu.ac.kr/~bockjoo/EDG_testbed/ LDAP for RB job CMS VO LDAP

  11. How to Use the EDG Testbed • Certificate should be obtained • Account on a UI must exist • UI Commands Basic Globus Commands grid-proxy-init  To Access Resource Broker globus-url-copy ………………. EDG Commands(python) dg-job-cancel dg-job-get-output dg-job-get-logging-info dg-job-id-info dg-job-list-match dg-job-status dg-job-submit • More Info: http://marianne.in2p3.fr/datagrid/documentation

  12. Input “sandbox” UI JDL Input “sandbox” Output “sandbox” Job Submit Job Query Brokerinfo Job Status Output “sandbox” Job Status dg-job-cancel dg-job-get-output dg-job-get-logging-info dg-job-id-info dg-job-list-match dg-job-status dg-job-submit Data Management Services An Illustration of Job Submission Information Service LFN->PFN Resource Broker Author. &Authen. Storage Element Job Submission Service Logging & Book-keeping Compute Element

  13. Job Submission - UI Job submission command Job Id

  14. Job Submission - RB Initial job submission phase Job execution phase

  15. Job Submission - JSS JSS Server Log JSS parsed

  16. Job Submission - CondorG CondorG just located the resource (CE)

  17. Job Submission – the CE=QUEUE

  18. Job Submission – RB ftp+WN RB wuftpd WN pbs mom

  19. GDMP:Datamanagement More Info: http://project-gdmp.web.cern.ch/project-gdmp/

  20. How to use GDMP and RC? • Setup environment, e.g., EDG_VO=cms • Subscribe to other site’s SE gdmp_host_subscribe –S $SE1 –P $P1 –r $SE2 –p $P2 • Create file(s) on SE by hand or by JDL (The file should belong tothe group $EDG_VO) • On UI, register the file locally gdmp_register_local_file -R -p /flatfiles/SE00/cms/dummy –S $SE –P $P –V $EDG_VO • On UI, publish the file and register it on the RC gdmp_publish_catalogue -V cms -C -L log_ID • On UI, replicate file if necessary (gsiftp+ldapadd) gdmp_replicate_get –r $SE –p $P –V $EDG_VO More Info: http://neutrino.snu.ac.kr/~bockjoo/EDG_testbed/contents/gdmp_test.html

  21. SE Subscription Data Management using GDMP Check catalog on se1 File Creation on SE LDAP :RC for se1 Check File on SE Replicate from se1 to se2 Register and Publish LDAP: RC for se1 and se2 Check catalog on se2

  22. Example Application of EDG RB UI Real user 빅 팻 디스크 Big Fat Disk NFS Que GSIFTP GDMP client (with new VO) K2K Software • Currently running Experiment • Existing Resources • Matter of adding ques and hacking : examples exist Que SE VOuser CE 디스크 VOuser NFS GSIFTP GDMP server (with new VO) NFS GDMP client (with new VO) GSIFTP MAP on disk With maximum security Que NFS WN CDF Software grid-security VOuser GSIFTP EDG testbed at SNU • Some works are in progress to include the existing node as a WN for different VO’s within a testbed site

  23. Summary • At the moment, the EDG testbed is operated at EC • EDG testbeds are constructed and working at KNU and SNU in Korea • For Data Management, an LDAP for RC is constructed and the GDMP is used within Korea. This is working. • The testbed is also operational betweenKNU and SNU • A possible HEP application for currently running experiments is illustrated and we are moving toward that direction • The EDG testbed will be constructed for other institutions in Korea • New VO will be tested extensively for a few months • Hopefully, we would like to learn other testbeds and test crossgrid functionalities

  24. US FNAL APII-TransPac EuropeCERN TEIN 현해 Network to the world 한-일: 8Mbps 한-미: 45Mbps • 10Mbps • 45Mbps by • the end of 2002 2Gbps ( soon )

  25. File transfer test on the network • parallel ftp used. • KNU – SNU : ~50Mbps (155Mbps) • KNU – SKKU : ~50Mbps (155Mbps) • SKKU – SNU : ~50Mbps (155Mbps) • CHEP – KEK : ~7Mbps (8Mbps) • CHEP – Fermilab : ~17 Mbps (45 Mbps) • CHEP – CERN : ~8Mbps (10Mbps) • FERMILAB – CERN : ~2Gbps ( DataTAG project of iVDGL )

  26. File transfer test on the network (KNU-SKKU,SNU) file size for one ftp= 30MB

  27. File transfer test on the network KNU-US ftp test US-CERN DataTAG link test

  28. Plans • Route to Europe • 1st TEIN : 45Mbps from next year • 2nd : via USA Korea  FERMILAB  CERN % We will use Fermilab’s storage as buffer. • 3rd : via Japan and USA Korea  KEK  Fermilab  CERN 2Gbps >1Gbps 2 Gbps Participation to iVDGL anables 2nd and 3rd route. • Contribution : 2003 CMS summer production We will provide significant CPU power and storage.

  29. Future bandwidth requirement on the network • Korea – Japan – USA Connection • for experiments in Japan ~ 400Mbps by 2006 ~ 650Mbps after 2006 • CDF experiment ~40Mbps by 2004 ~230Mbps after 2004 • CMS experiment ~800Mbps ( 1PB/year ) • Including other traftic and Japan’s, at least 2Gbps required. • Should evaluate to Lambda (10Gbps) connection to USA, Japan, Europe

  30. 622 Mbps Future bandwidth requirement on the network +10 Gbps l

More Related