1 / 23

SitE Report

SitE Report. University of Johannesburg South Africa Stavros Lambropoulos Network Engineer I.C.S Department. Overview. History of the UJ Research Cluster User Groups Hardware South African Compute Grid (SA Grid) Status Applications Issues Future Links Contributions.

mika
Download Presentation

SitE Report

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. SitE Report University of Johannesburg South Africa Stavros Lambropoulos Network Engineer I.C.S Department

  2. Overview • History of the UJ Research Cluster • User Groups • Hardware • South African Compute Grid (SA Grid) • Status • Applications • Issues • Future • Links • Contributions

  3. History of UJ Research Cluster • UJRC started as an initiative of the High Energy Physics Group • March 2009 - The UJ-OSG Compute Element passes validation and is registered on VORS (Resource Selector) • March 2009 – 56 CPU Cores available • April 2009 – UJ Hosted Grid School

  4. UseR Groups • High Energy Physics (Physics) • Astrophysics (Physics) • Molecular Dynamics (Chemistry) • Quantum Chemistry (Chemistry) • Applied Mathematics • Numerical Studies (Engineering)

  5. HARDWARE • 1 Head Node comprising : Dell 2950, 2 x 4 Core Xeon Processors 16 GB RAM 900 GB – RAID5 Scientific Linux 4/64 Bit hosts : NFSv4, Accounts, Torque, Ganglia

  6. HARDWARE (Contd) • Separate Virtual Machines (VMWare Server) for : OSG CE (1 GB RAM) OSG UI (submit node) – (2 GB RAM) gLite CE (1 GB RAM) gLite UI (submit node) – (1 GB RAM)

  7. HARDWARE (Contd) • 7 Worker Nodes comprising : Dell 1425, 2 x 4 Core Opteron Processors 16 GB RAM Scientific Linux 4/64 Bit gLite sw locally installed OSG sw from NFS • Alcatel 6400 Gig Switch

  8. HARDWARE (Contd)

  9. SA Grid • The South African National Grid is a project to provide a national grid computing infrastructure to support scientific computing and collaboration. This project is managed by a consortium of universities, national laboratories and the Meraka Institute, under the cyber infrastructure programme, based on the gLite middleware .

  10. Status • OSG is operational on SL4 base

  11. Status (contd) • Started Discussion on the choice of referent Tier1/Tier2 for ATLAS and ALICE • WN’s, CE’s and UI’s to be updated to SL5 as requested by LHC Computing Grid • Cobbler and Puppet to be used for the new SL5 node installation and management • Updating of Head Node from SL4/VMWare to SL5/Xen is planned

  12. Applications • 2 Commercial Applications Running Locally : ANSYS FLUENT – Flow modeling Software Star-CCM+ - Computational Fluid Dynamics • Other Local Applications : Geant4 for NA63, MineralPET NA63 dedicated simulation code Diamond Lattice Deformation

  13. Applications • On OSG : • Full ATLAS VO Support • ENGAGE VO runs a few jobs • local ATLAS users submit remote jobs from local UI • Initial discussions have started to allow DOSAR VO • On SAGrid : • Will allow SAGrid VO’s • ALICE VO • ATLAS • e-NMR VO • WISDOM VO • GILDA

  14. The UJ Research Cluster and the OSG GRID UJ – Physics‏ High Energy Physics, ATLAS experiment at CERN Ketevi Assamagan, Simon Connell, Sergio Ballestrero, Claire Lee, Neil Koch, Phineas Ntsoele ATHENA installed, using Pythia event generator to study various Higgs scenarios.

  15. UJ – Physics‏‏ Diamond Ore Sorting (Mineral-PET)‏ Sergio Ballestrero, Simon Connell, Norman Ives, Martin Cook, Winile Sibande GEANT4 MonteCarlo Online diamond detection Monte Carlo simulation Online diamond detection

  16. Issues • Limited International Bandwidth Currently Using 11Mb/s To be Upgraded early next year with the SEACOM cable

  17. Issues (contd) • Research Funding – for HW and training • Additional complexity to manage both OSG and gLite • Lack of caching by OSG installer, partially solved with local Squid cache • No automated install & config system yet, starting to work on Cobbler and Puppet • NFSv4 problematic on SL4 • Monitoring, need to add detailed job monitoring/stats for Torque

  18. Issues (contd) • Manpower – Grid services not primary job/role for the 3 people – Addressing problem with single national Operations Team • Low Usage – Marketing of services and availability has been done but researchers are slow to start • No experience gathered on utilization of resource in terms of constraints on memory, disks, CPU and network • Final VO acceptance policy required

  19. Future • Hardware Upgrade : • Additional 4 x WN’s being configured • 1 x Dell MD1000 Storage shelf (6TB raw) will be connected to the Head Node. – Ordered • 16 x WN’s (Dell M605 blade chassis, with 2 x 6 Cores, 32GB Ram) – Ordered • 224 Cores will be available • DOSAR Workshop in South Africa in 2010

  20. Links • University of Johannesburg http://www.uj.ac.za • UJ Physics http://physics.uj.ac.za/cluster • South African Grid http://www.sagrid.ac.za

  21. Contributions • Prof. S. Connell – UJ Physics Department • Sergio Ballestrero – UJ Physics & CERN ATLAS TDAQ • Bruce Becker – SA Grid Co-ordinator • Francois Mynhardt – UJ I.C.S Department

  22. Questions

More Related