1 / 39

Kihyeon Cho & Soonwook Hwang (KISTI)

Super Belle with FKPPL VO & AMGA Data Handling. Kihyeon Cho & Soonwook Hwang (KISTI). Contents. KISTI Super computing Center FKPPL VO Farm Super Belle Data Handling Summary. FKPPL VO Grid Testbed. Goal. Background

dillon
Download Presentation

Kihyeon Cho & Soonwook Hwang (KISTI)

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Super Belle with FKPPL VO & AMGA Data Handling Kihyeon Cho & Soonwook Hwang (KISTI)

  2. Contents • KISTI Super computing Center • FKPPL VO Farm • Super Belle Data Handling • Summary

  3. FKPPL VO Grid Testbed

  4. Goal Background Collaborative work between KISTI and CC-IN2P3 in the area of Grid computing under the framework of FKPPL Objective (short-term) to provide a Grid testbed to the e-Science summer school participants in order to keep drawing their attention to Grid computing and e-Science by allowing them to submit jobs and access data on the Grid (long-term) to support the other FKPPL projects by providing a production-level Grid testbed for the development and deployment of their applications on the Grid Target Users FKPPL members 2008 Seoul e-Science summer school Participants

  5. FKPPL Testbed VOMS WIKI CE UI CE LFC WMS SE SE FKPPL VO IN2P3 KISTI

  6. VO Registration Detail Official VO Name fkppl.kisti.re.kr Description VO dedicated to joint research projects of the FKPPL(France Korea Particle Physics Laboraroty), under a scientific research programme in the fields of high energy physics (notably LHC and ILC) and e-Science including Bioinformatics and related technologies Information about the VO https://cic.gridops.org/index.php?section=vo

  7. FKPPL VO Usage Application porting support on FKPPL VO Geant4 Detector Simulation Toolkit Working with National Cancer Center WISDOM MD part of the WISDOM drug discovery pipeline Working with the WISDOM Team Support for FKPPL Member Projects Grid Testbed for e-Science School Seoul e-Science summer school

  8. How to access resources in FKPPL VO Testbed • Get your certificate issued by KISTI CA • http://ca.gridcenter.or.kr/request/certificte_request.php • Join a FKPPL VO membership • https://palpatine.kisti.re.kr:8443/voms/fkppl.kisti.re.kr • Get a user account on the UI node for FKPPL Vo • Send an email to the system administrator at ssgyu@kisti.re.kr

  9. User Support FKPPL VO Wiki site http://anakin.kisti.re.kr/mediawiki/index.php/FKPPL_VO User Accounts on UI machine 17 User accounts have been created FKPPL VO Registration 4 users have been registered as of now

  10. Contact Infomation Soonwook Hwang (KISTI), Dominique Boutigny (CC-IN2P3) responsible person hwang@kisti.re.kr, boutigny@in2p3.fr Sunil Ahn (KISTI), Yonny Cardenas (CC-IN2P3) Technical contact person, siahn@kisti.re.kr, cardenas@cc.in2p3.fr Namgyu Kim Site administrator ssgyu@kisti.re.kr Sehoon Lee User Support sehooi@kisti.re.kr

  11. Configuration • KISTI site VOMS, WMS, CE+WN*, UI, Wiki * Infrastructure installation in progress ( a cluster with 128 cores has been purchased) • CC-IN2P3 site CE+WN, SE, LFC Yonny CARDENAS Monday, December 2, 2008 11

  12. Configuration • VO Registration procedure • VO name: fkppl.kisti.re.kr • VO manager: Sunil Ahn • Status: Active Yonny CARDENAS Monday, December 2, 2008 12

  13. Status (Operational Services)‏ KISTI site: • VOMS OK • WMS OK • CE OK • WIKI OK • SE OK • WN* in Progress CC-IN2P3 site: • SE OK • dCache/SRM OK • WN OK • CE OK • LFC OK Yonny CARDENAS Monday , December 2 , 2008 13

  14. Available Services Job Submission • Available since October 1, 2008 • Resource allocation: 5 millions hours CPU SI2K • CC-IN2P3 Job monitoring • Quality of Service • Operation team Yonny CARDENAS Monday, December 2 , 2008 14

  15. Available Services Data storage dCache SE/SRM • System for storing and retrieving data, distributed among a large number of heterogeneous server nodes. • Implements the SRM v2.2 interface • required EGEE/LCG • Resource allocation: 0.5 Terabytes Yonny CARDENAS Monday , December 2 , 2008 15

  16. Available Services Data storage • AFS (Andrew File System) • Network file system for personal and group files, experiment software, system tools (compilers, libraries , ... )‏ • Indirect use (jobs)‏ • Resource allocation: 2 Gigabytes Yonny CARDENAS Monday, December 2, 2008 16

  17. Available Services Data storage LFC - LCG File Catalog • Maintains mappings between logical file names (LFN) and SRM file identifiers. • Supports references to SRM files in several storage elements. Yonny CARDENAS Monday, December 2 , 2008 17

  18. Utilisation - Services Jobs Submission • October • 34 jobs for 150 hours CPU SI2K • November • 1690 jobs for 48250 hours CPU SI2K Yonny CARDENAS Monday, December 2, 2008 18

  19. Utilisation - Services • Data Storage • 7193 files for 60 G bytes of used space • 440 G bytes available. Yonny CARDENAS Monday, December 2, 2008 19

  20. User Support • FKPPL VO Wiki site • http://anakin.kisti.re.kr/mediawiki/index.php/FKPPL_VO • User Accounts on UI • 20 User accounts has been created • FKPPL VO Membership Registration • 7 Users have been registered at FKPPL VO membership

  21. FKPPL VO Usage • Deployment of Geant4 applications on FKPPL VO • Detector Simulation Toolkit • Working with Jungwook Shin at National Cancer Center • Grid Interoperability Testbed

  22. Geant4 Application: GTR2_com Application name: GTR2_com (G4 app for proton therapy sim s/w by developed by NCC) -> GTR2 : Gantry Treatment Room #2, com: commissioning (now GTR2 simulation code is under commissioning phase) -> libraries: Geant4 , root (root.cern.ch) as simulation output library user macro output /user/io/OpenFile root B6_1_1_0.root /GTR2/SNT/type 250 /GTR2/SNT/aperture/rectangle open #Geant4 kernel initialize /run/initialize /GTR2/FS/lollipops 9 5 /GTR2/SS/select 3 /GTR2/RM/track 5 /GTR2/RM/angle 80.26 /GTR2/VC/setVxVy cm 14.2 15.2 /beam/particle proton /beam/energy E MeV 181.8 1.2 /beam/geometry mm 3 5 /beam/emittance G mm 1.5 /beam/current n 3000000
#SOBP /beam/bcm TR2_B6_1 164 /beam/juseyo /user/io/CloseFile GTR2_com GTR2_com 의 input은 nozzle의 configuration 이며, 이 configuration 이 명시된 macro 파일을 읽어서 최종 양성자 빔에 의한 선량분포를 3D-histogram 의 root 파일로 출력

  23. Distribution of the completion time of 1000 GTR2_com jobs on FKPPL VO The submission of 1000 GTR2_com jobs was done around 18:05

  24. BC408 M.C study • Purpose : the accurate simulation study will help to design and construct a dosimetry device utilizing the BC408 scintilator

  25. From 1 file, ~3.5 hrs on WN of FKPPL From 589 files, 589 files sucessfully generated among total 99*7 (693) jobs BC408 M.C on FKPPL resolution : 2mm in X,Y and 1mm thicknes I immediately submit parametric job after initializing the proxy

  26. Super Belle Data Handling

  27. Super Belle Computing • Conveners: T. Hara, T. Kuhr • Distributed Computing (Martin Seviour) • Data Handling (Kihyeon Cho) • Data Base (Vacant)

  28. Super Belle Data Handling • Data Handling depends on distribution computing. • Cloud Computing? • Grid farm? • Data Handling Suggestions (2/17) • SAM (Sequential Access through Metadata) • CDF by Thomas Kuhr • AMGA • KISTI by Soonwook Hwang

  29. Objectives Build large-scale, production-quality grid infrastructure for e-Science Available to scientists 24/7 EGEE grid Infrastructure 300 sites in 50 countries 80,000 CPU cores 20 PBytes 10,000 User EGEE (Enabling Grids for E-SciencE) • the largest multi-disciplinary grid infrastructure in the world

  30. Overview of AMGA (1/2) • Metadata is data about data • AMGA provides: • Access to Metadata for files stored on the Grid • A simplified general access to relational data stored in database systems. • 2004 – the ARDA project evaluated existing Metadata Services from HEP experiments • AMI (ATLAS), RefDB (CMS), Alien Metadata Catalogue (ALICE) • Similar goals, similar concepts • Each designed for a particular application domain • Reuse outside intended domain difficult • Several technical limitations: large answers, scalability, speed, lack of flexibility • ARDA proposed an interface for Metadata access on the GRID • Based on requirements of LHC experiments • But generic - not bound to a particular application domain • Designed jointly with the gLite/EGEE team

  31. Overview of AMGA (2/2) What is AMGA ? (ARDA Metadata Grid Application) • Began as prototype to evaluate the Metadata Interface • Evaluated by community since the beginning: • Matured quickly thanks to users feedback • Now part of gLite middleware : EGEE’s gLite 3.1 MW • Requirements from HEP community • Millions of files, 6000+ users, 200+ computing centres • Mainly (real-only) file metadata • Main concerns :scalability, performance, fault-tolerance, Support for Hierarchical Collection • Requirements from Biomed community • Smaller scale than HEP • Main concerns : Security ARDA Project (A Realisation of Distributed Analysis for LHC)

  32. Metadata user requirements I want to store some information about files In a structured way query a system about those information keep information about jobs I want my jobs to have read/write access to those information have easy access to structured data using the grid proxy certificate NOT use a database

  33. Metadata Concepts • Schema (table, think directory) • Has hierarchical name and list of attributes /prod/events • Attributes (columns) • Have name and storage type • Interface handles all types as strings • Entry (row) • Live in a schema, assign values to attributes • Collections • A set of entries associated with schema • Query • SELECT ... WHERE ... clause in SQL-like or SQL query language

  34. AMGA data organization Relational schema AMGA(hierarchy) Schema/Directory /HOSPITAL/ TABLE: HOSPITAL Schema/Directory PATIENTS/ DOCTORS/ TABLE: PATIENTS Entries john george Collection Attributes

  35. Importing existing data Suppose that you have the data A reasonable question would be: Can I use my existing database data?? The answer is YES Importing data to AMGA Pretty simple Connect a database to AMGA Execute the import command import table directory Ready to go!

  36. AMGA Use Cases

  37. AMGA Features in One Slide • Official gLite middleware component for metadata catalogue • Metadata is relationally structured • Schema (aka table, think directory)Has hierarchical name and list of attributes /prod/events • Attributes (aka columns)Have name and storage type, Interface handles types as strings • Entry (aka row)Live in a schema, assign values to attributes • Query: SELECT ... WHERE ... clause in SQL • Fine grained access control (ACL) support • Table level and entry level • Tight integration into the Virtual Organization Management System (VOMS) • X509 Grid certificate • Native SQL support in AMGA 1.9 • Direct DB access to existing databases on the Grid via SQL • OGF WS-DAIR compatible interface support in AMGA 2.0 • Uniform interface to Heterogeneous Database backend • Oracle, Postgres SQL, MySQL, etc. • Support for Many Programming APIs • Diverse user community requested/provided • e.g., C/C++, Java, Python, Perl, PHP • Replication Support • Full replication, Partial replication, Federation • Support for “Import of pre-existing databases”

  38. AMGA Website http://amga.web.cern.ch/amga

  39. To do • Prototype of AMGA • Using Belle flat form, we will use AMGA (Namkyu Kim and Dr. Junghyun Kim) • Belle flat form => flag • Using Super Belle flat form (Grid or Cloud computing), we will use AMGA.

More Related