230 likes | 399 Views
SitE Report. University of Johannesburg South Africa Stavros Lambropoulos Network Engineer I.C.S Department. Overview. History of the UJ Research Cluster User Groups Hardware South African Compute Grid (SA Grid) Status Applications Issues Future Links Contributions.
E N D
SitE Report University of Johannesburg South Africa Stavros Lambropoulos Network Engineer I.C.S Department
Overview • History of the UJ Research Cluster • User Groups • Hardware • South African Compute Grid (SA Grid) • Status • Applications • Issues • Future • Links • Contributions
History of UJ Research Cluster • UJRC started as an initiative of the High Energy Physics Group • March 2009 - The UJ-OSG Compute Element passes validation and is registered on VORS (Resource Selector) • March 2009 – 56 CPU Cores available • April 2009 – UJ Hosted Grid School
UseR Groups • High Energy Physics (Physics) • Astrophysics (Physics) • Molecular Dynamics (Chemistry) • Quantum Chemistry (Chemistry) • Applied Mathematics • Numerical Studies (Engineering)
HARDWARE • 1 Head Node comprising : Dell 2950, 2 x 4 Core Xeon Processors 16 GB RAM 900 GB – RAID5 Scientific Linux 4/64 Bit hosts : NFSv4, Accounts, Torque, Ganglia
HARDWARE (Contd) • Separate Virtual Machines (VMWare Server) for : OSG CE (1 GB RAM) OSG UI (submit node) – (2 GB RAM) gLite CE (1 GB RAM) gLite UI (submit node) – (1 GB RAM)
HARDWARE (Contd) • 7 Worker Nodes comprising : Dell 1425, 2 x 4 Core Opteron Processors 16 GB RAM Scientific Linux 4/64 Bit gLite sw locally installed OSG sw from NFS • Alcatel 6400 Gig Switch
SA Grid • The South African National Grid is a project to provide a national grid computing infrastructure to support scientific computing and collaboration. This project is managed by a consortium of universities, national laboratories and the Meraka Institute, under the cyber infrastructure programme, based on the gLite middleware .
Status • OSG is operational on SL4 base
Status (contd) • Started Discussion on the choice of referent Tier1/Tier2 for ATLAS and ALICE • WN’s, CE’s and UI’s to be updated to SL5 as requested by LHC Computing Grid • Cobbler and Puppet to be used for the new SL5 node installation and management • Updating of Head Node from SL4/VMWare to SL5/Xen is planned
Applications • 2 Commercial Applications Running Locally : ANSYS FLUENT – Flow modeling Software Star-CCM+ - Computational Fluid Dynamics • Other Local Applications : Geant4 for NA63, MineralPET NA63 dedicated simulation code Diamond Lattice Deformation
Applications • On OSG : • Full ATLAS VO Support • ENGAGE VO runs a few jobs • local ATLAS users submit remote jobs from local UI • Initial discussions have started to allow DOSAR VO • On SAGrid : • Will allow SAGrid VO’s • ALICE VO • ATLAS • e-NMR VO • WISDOM VO • GILDA
The UJ Research Cluster and the OSG GRID UJ – Physics High Energy Physics, ATLAS experiment at CERN Ketevi Assamagan, Simon Connell, Sergio Ballestrero, Claire Lee, Neil Koch, Phineas Ntsoele ATHENA installed, using Pythia event generator to study various Higgs scenarios.
UJ – Physics Diamond Ore Sorting (Mineral-PET) Sergio Ballestrero, Simon Connell, Norman Ives, Martin Cook, Winile Sibande GEANT4 MonteCarlo Online diamond detection Monte Carlo simulation Online diamond detection
Issues • Limited International Bandwidth Currently Using 11Mb/s To be Upgraded early next year with the SEACOM cable
Issues (contd) • Research Funding – for HW and training • Additional complexity to manage both OSG and gLite • Lack of caching by OSG installer, partially solved with local Squid cache • No automated install & config system yet, starting to work on Cobbler and Puppet • NFSv4 problematic on SL4 • Monitoring, need to add detailed job monitoring/stats for Torque
Issues (contd) • Manpower – Grid services not primary job/role for the 3 people – Addressing problem with single national Operations Team • Low Usage – Marketing of services and availability has been done but researchers are slow to start • No experience gathered on utilization of resource in terms of constraints on memory, disks, CPU and network • Final VO acceptance policy required
Future • Hardware Upgrade : • Additional 4 x WN’s being configured • 1 x Dell MD1000 Storage shelf (6TB raw) will be connected to the Head Node. – Ordered • 16 x WN’s (Dell M605 blade chassis, with 2 x 6 Cores, 32GB Ram) – Ordered • 224 Cores will be available • DOSAR Workshop in South Africa in 2010
Links • University of Johannesburg http://www.uj.ac.za • UJ Physics http://physics.uj.ac.za/cluster • South African Grid http://www.sagrid.ac.za
Contributions • Prof. S. Connell – UJ Physics Department • Sergio Ballestrero – UJ Physics & CERN ATLAS TDAQ • Bruce Becker – SA Grid Co-ordinator • Francois Mynhardt – UJ I.C.S Department