1 / 22

Grid activities in the Czech Republic

Grid activities in the Czech Republic. Jiří Kosina, Miloš Lokajíček, Jan Švec Institute of Physics of the Academy of Sciences of the Czech Republic http://www.fzu.cz/. HEP Experiments in the Czech Republic (1). D0, STAR, ATLAS and ALICE (main experiments only) computing activities:

wenda
Download Presentation

Grid activities in the Czech Republic

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Grid activities in the Czech Republic Jiří Kosina, Miloš Lokajíček, Jan Švec Institute of Physics of the Academy of Sciences of the Czech Republic http://www.fzu.cz/

  2. HEP Experiments in the Czech Republic (1) • D0,STAR,ATLAS and ALICE (main experiments only) computing activities: • detector simulations, reconstruction • data analysis in running projects - main participants are • Institutions • Institute of Physics AS CR, Institute of Nuclear Physics AS CR • Charles University in Prague • Czech Technical University in Prague • Facilities • EDG test-bed now being converted to LCG

  3. Institutions involved in the Grid activities • CESNET(http://www.cesnet.cz/) • Charles University in Prague(http://www.cuni.cz/) • Czech Technical University in Prague(http://www.cvut.cz/) • Institute of Physics AS CR(http://www.fzu.cz/)

  4. CESNET(1) CESNET z.s.p.o. • association of legal entities created by Czech universities and Academy of Sciences of the CR • Czech scientific network provider • shareholder of the Dante • member of TERENA • Internet2 international partner • Computing and networking strategic research projects • Optical networks and their development • IPv6 implementation in the CESNET2 network • Multimedia transfers • MetaCenter • Voice services in the CESNET2 network • QoS in high-speed networks

  5. CESNET(2) • International projects • GEANT • DataGrid -> EGEE • SCAMPI • 6NET • Other projects • Infrastructure and technology for on-line education , Distributed contact center , Smart NetFlow analyser , Storage over IP , Presentation , RT system in CESNET, Securing CESNET2 local networks , Time synchronization and NTP servers, Platforms for video transmission and production, MEDIMED

  6. CESNET(3) • 1.2Gbps connection to GEANT (over HW 10 Gbps) • 1+1Gbps connection to NIX (Peering with Czech ISPs) • 800Mbps connection to USA (Telia) • 2.5 Gbps CzechLight connection (now used for other tests) • Reserved optical connection 1 GbpsInstitute of Physics <-> CERN • Connected to our linux router, switching to “backup” line through Internet managed by BGP protocol • Possibility to connect to StarLight in the future

  7. CESNET(4) • European DataGrid project • WP1 workload management • WP7 network monitoring • Certification authority established for EDG • issuing certificates to Czech academic subjects

  8. Institute of Physics AS CR - FZU • D0 experiment participation • ATLAS – main contribution to the hadronic calorimeter TILECAL and Inner Detector (pixel sensors production and test of strip and pixel detectors), power supplies design and production • Computing • Mass simulation for D0, reconstruction, data analysis • ATLAS – participation to Data Challenges, Atlas-LCG

  9. Institute of Physics AS CR – FZU (2) • Computing projects • EDG from 1 Jan 2001 • WP6 – testbed • CESNET network support important • Plan to continue in EGEE • LCG – GDB (LHC Computing Grid – Grid Deployment Board) • LCG deployed in October 2003 – CERN press release

  10. Institute of Physics AS CR – FZU (3) • Dedicated server farm for HEP and Grid computing • 34x dual 1.13Ghz PIII, 1TB disk array, currently experimenting with new 10 TB disk array

  11. Institute of Physics AS CR – FZU (4) • Current status • Construction of the new computing room in the institute • Designed for 150 kW electric power • UPS, cooling, engine-generator • Construction finished by end 2003 with 50% capacity • Full capacity next year • Application for triple capacity upgrade of the current farm (30 double Xeon units, 20 TB disk space) for the 2004. Hope in positive result.

  12. Job Management PBS Pro 5.2 LCG1 -> OpenPBS (will merge soon with PBS Pro) queues shortq normalq longq • hightest • d0 • atlas • alice golias:~$ qsub -q atlas run.sh

  13. D0 MC simulations SAM stationSequential Access to data via Metadata UPS, UPDmanagement of software products on local systems (UPS) downloading products from product distribution servers (UPD) D0-RunIIsoftware rel. p14.02.00, p14.05.01

  14. EDG, LCG • We have installed LCG1 software and currently running LCG1-1_1_1 version of the software. • LCG – middleware software to provide environment for distributed computing for LHC • Member of WP6 work package – farm is running EDG 1.4 software, for testbed purposes

  15. LCG (1) • LCG – LHC (Large Hadron Collider) Computing Grid • The accelerator will start operation in 2007 • 12-14 PetaBytes of data will be generated every year (20 milion Cds). It is assumed that analyzing this will require approximately 70.000 CPUs

  16. LCG(2) • Based on EDG software – project funded by EU, finishing by the end of this year • GRID infrastructure allows every member to submit his job (along with JDL file) to the GRID, and after the computing is finished (no matter where), get the output • In JDL you can specify requirements specific for task (SW versions, # CPUs, etc) • Distributed storage of data.

  17. LCG(3) • Installation of LCG software is done through LCFGng toolkit – useful to manage wide variety of configurations for different hosts • Joining sites are provided with pregenerated skeleton of configuration files in order to simplify installation procedure (it's necessary to modify them) • Installation is done simply by enabling PXE boot and rebooting the nodes

  18. 10TB disk array and Linux • We have 10 TB disk array. Problem with Linux 2.4 – only 2TB block devices supported • Patch for LBD exists for 2.4 kernel, but it collides with another patch needed for LVM.

  19. 10TB disk array and Linux • Kernel 2.6 (not yet stable) supports both LBD and device mapper (needed for LVM). After some hacking of NFS code (our patches are incorporated in 2.6.0-test9 kernel) we have successfully created 6 TB partition (XFS, ext2)

  20. Institutions involved – contacts • Jan Svec <svecj@fzu.cz> - network and system management, D0 experiment (grid) • Jiri Kosina <kosina@fzu.cz> - installation of grid software at FZU, ALICE experiment, network and system management • Jiri Chudoba <Jiri.Chudoba@cern.ch> - ATLAS experiment • <pcclust@cesnet.cz> - CESNET's PC clusters

  21. Summary – Grid infrastructure • Basic Grid infrastructure has been established during last year and further expansion due to research or commercial needs in future shouldn't be problematic • CESNET provides good international network connectivity, plan tests of CzechLight optical network • Preparing for Data Challenges of LHC experiments in 2004

More Related