1 / 16

Northgrid Status

Alessandra Forti Gridpp25 Ambleside 25 August 2010. Northgrid Status. Apel pies Lancaster status Liverpool status Manchester status Sheffield status Conclusions. Outline. Apel pie (1). Apel pie (2). Apel pie (3). Lancaster. The new shared HPC facility at Lancaster.

zada
Download Presentation

Northgrid Status

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Alessandra Forti Gridpp25 Ambleside 25 August 2010 Northgrid Status

  2. Apel pies • Lancaster status • Liverpool status • Manchester status • Sheffield status • Conclusions Outline

  3. Apel pie (1)

  4. Apel pie (2)

  5. Apel pie (3)

  6. Lancaster The new shared HPC facility at Lancaster. 1760 cores, delivering over 20k HEPSPEC2k6 Behind an LSF batch system. In a proper machine room (although that's had teething troubles). Added bonus of also providing a home for future GridPP kit. Access to a “Panasus Shelf” for a high performance tarball and VO software areas. 260 TB Grid-only storage included in purchase. Roger Jones is managing director and we have admin access to the cluster, so we have a strong voice in this joint venture. But with root access comes root responsibility.

  7. Lancaster New ``HEC'' facility almost ready. Final stages of acceptance testing. Number of gotchas and niggles causing some delays, but this is expected due to the facilities scale. Underspecced fuses, overspecced UPS, overheating distribution boards, misbehaving switches, oh my! Site otherwise chugging along nicely (with some exceptions). New hardware to replace older servers Site needs a downtime in the near future to get these online and have a general clean up. Tendering for additional +400TB storage

  8. Liverpool: Addition to Cluster 32 bit cluster replaced by 9 x Quad chassis (4 worker nodes in each), boosts 64bit capability, reduces power consumption Each node has 2 x X5620 CPUs (4 core), 2 x hot swap SATA, IPMI, redundant power supply Runs 64bit SL5, glite 3.2 Entire Cluster CPU: 6 x Quad core worker nodes (56 cores) 16 x Quad chassis, 4 WNs each (512 cores) 3 GB RAM per core 8214.44 HEPSPEC06 Storage: 286 Terabytes Plans: Migrate from VMWare to KVM (lcg-CE, CREAM-CE, Site BDII, MON Box ...)

  9. Liverpool: problem https://gus.fzk.de/ws/ticket_info.php?ticket=61224 Site looks as not “not available” in gstat because GSTAT/Nagios doesn't recognise RedHat Enterprise Server as OS OS was correctly advertised following the procedure published How_to_publish_the_OS_name The ticket has been reassigned to gstat SU but the problem has affected the site nagios tests.

  10. Manchester: new hardware Procurement at commissioning stage New hardware will replace half of the current cluster Computing power 20 quad chassis 2x6 core per motherboard 13.7 HEPSPEC06 per core 4GB memory per core 125 GB disk space per core 2 Gb/s bonded link per mother board Total: 960 cores = 13.15k HEPSPEC06 Storage 9x36bay units 30+2+1x2TB for the raid (30 data disks + 2 parity disks + 1 hot spare) 2x250GB for the OS Total: 540 TB usable

  11. Manchester: new hardware Computing nodes arrived and currently running on site soak tests Testing opportuninty to use ext4 as file system rather than ext3 Initial test with iozone on old nodes show a visible improvement especially with multithreaded tests Single thread better in writing only Need to start testing new nodes Storage will be delivered this week Raid card had a disk detection problem. Issue is now solved by a new version of the firmware. Rack switches undergoing reconfiguration to allow 4x2Gb/s bonded links from quad chassis 1x4Gb/s bonded link from storage units 8xGb/s uplink to the main cisco

  12. Manchester: general news Current stability Consistently delivered a high number of CPU hours Current storage is stable although slowed down by SL5(.3?)/XFS problem that causes bursts of load on the machine Evaluating how to tackle the problem for the new storage. Management Stefan Soldner Rembold, Mike Seymour and Un-ki Yang will take over from Roger Barlow the management of the Tier2.

  13. Sheffield: major cluster upgrade We have built a new cluster in Physics Department with 24/7 access. It is a join cluster gridpp/local group with a common torque server Storage Storage Nodes running DPM 1.7.4, recently upgraded to SL5.5. Tests show latest SL5 kernels no longer show xfs bug DPM head node 8 cores and 16 GB RAM 200 TB of disk space have been deployed – covering the pledge for 2010 according to the new GRIDPP requirements SW RAID5 (no raid controllers) All disk pools assembled in Sheffield 2 TB disks seagate barracuda and one 2TB Western digital 5x16bay unit with 2 fs, 2x24 bay unit with 2 fs Cold spare unit on standby near each server 95% isk space reserved for Atlas Plan to add 50 TB late in the autumn

  14. Sheffield: major cluster upgrade Infrastructure Additional 32A ring mains added to the machine room in the physics department Fiber links connecting servers in physics to WN in CICS Old 5kW aircon replaced with a new 10kW (3x10kW aircon units) Dedicated WAN link to cluster Torque Server Accepts jobs from grid CE and local cluster Sends jobs in to all Wns Hosts DNS server CE SL4.8 4 single AMD opteron processor 850, 8GB if RAM, redundant power supply, 72 GB scsi in a raid1 MONBOX and BDII

  15. Sheffield: major cluster upgrade WN 50 WNs in Physics Department (32 from local hep cluster + 18 new), 5 Gb backbone Phenom 3200 MHz x86_64, 8 GB of RAM, 140GB/4cores, 11.96 HepSpec/core 102 old WNs in CICS, 1 Gb backbone 204 single Core 2.4 GHz Opterons (2GB), 4GB of RAM , 72 GB local disk per2Cores; 7.9 HepSpec/core; connected to the servers via fibre link Jobs requiring greater network bandwidth directed to WNs with better backbone software server with 1 TB disk (RAID1) and squid server were moved from CICS Cluster availability and reliability in July is 100% Sheffield is active in atlas production and user analysis

  16. Northgrid has been pretty stable and steadily crunching CPU hours for the past 4 months and data. • Sites have got or are in the process of getting new hardware rejuvenating the CPUs and increasing the storage • Both the CPUs and storage MoU requirements for Northgrid should be satisfied in excess in the near future. Conclusion

More Related