1 / 24

A Distributed Tier-1

A Distributed Tier-1. An example based on the Nordic Scientific Computing Infrastructure. GDB meeting – NIKHEF/SARA 13th October 2004 John Renner Hansen – Niels Bohr Institute. With contributions from Oxana Smirnova, Peter Villemoes and Brian Vinter.

Download Presentation

A Distributed Tier-1

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. A Distributed Tier-1 An example based on the Nordic Scientific Computing Infrastructure GDB meeting – NIKHEF/SARA 13th October 2004 John Renner Hansen – Niels Bohr Institute With contributions from Oxana Smirnova, Peter Villemoes and Brian Vinter

  2. Basis for a distributed Tier-1 structure • External connectivity • Internal connectivity • Computer and Storage capacity • Maintenance and operation • Long term stability

  3. NORDUnet network in 2003 General Internet 5G NETNOD 3.5G 155M RUNNet 622M General Internet 2.5G NASK 12M GÉANT 10G (Oct ’03)

  4. NorthernLight Stockholm Oslo Helsinki Dec 2003 2.5G links connected to “ONS boxes” giving 2 GE channels between endpoints Aug 2003 Copenhagen NetherLight Amsterdam

  5. NORDUNet was represented at the NREN-TIER1 Meeting Paris, Roissy Hilton, 12:00-17:00, 22 July 2004 by Peter Villemoes

  6. Denmark / Forskningsnet • Upgrading from 622Mbit/s to a 2.5 Gbit/s ring structure finished: • Copenhagen-Odense-Århus-Aalborg- and back via Göteborg to Copenhagen • Network research • setting up a Danish national IPv6 activity • dark fibre through the country • experimental equipment for 10GE channels

  7. Finland / FUNET • Upgraded to 2.5G already in 2002 • Upgrading backbone routers to 10G capability

  8. Norway / UNINETT • Network upgraded to 2.5G between major universities, • UNINETT is expanding to more services and organisations:

  9. Sweden / SUNET • 10G resilient nationwide network since Nov 2002 • all 32 universities have 2.5G access • Active participation in SweGrid, Swedish Grid Initiative

  10. Computer and Storage capacity at a Nordic Tier-1

  11. Who • Denmark Danish Center for Grid Computing - DCGC Danish Center for Scientific Computing -DCSC

  12. Who • Denmark • Finland - CSC

  13. Who • Denmark • Finland • Norway - NorGrid

  14. Who • Denmark • Finland • Norway • Sweden - SweGrid

  15. Denmark • Two collaborating Grid projects • Danish Centre for Scientific Computing Grid • DCSC-Grid spans the four DCSC sites and thus unify the resources, PC-clusters, IBM-Regatta, SGI-Enterprise, … within DCSC • Danish Centre for Grid Computing • Is the the national Grid project • DCSC Grid is a partner in DCGC

  16. Finland • Remains centred about CSC • CSC participates in NDGF and NGC • A Finnish Grid will probably be created • This Grid will focus more on accessing CSC resources with local machines

  17. Norway • NOTUR Emerging Technologies on Grid Computing is the main mover • Oslo-Bergen “mini-Grid” in place • Trondheim and Tromsø should be joining Curently under reorganization

  18. Sweden • SweGrid is a very ambitious project • 6 clusters have been created for the purpose of SweGrid each equipped with a 100 PCs and a large disk system • Large support and education organization is integrated in the plans

  19. NOS-N Research Councils DK SF S N Nordic Data Grid Facility Nordic Project Core Group Project Director 4 Post Doc.s Steering Group 3 members per country 1 R.C. Civil Servant 2 Scientists Nordic Data Grid Facility • Create the basis for a common Nordic Data Grid Facility 2. Coordinate Nordic GridActivities

  20. Services provided by the Tier-1 Regional Centres • acceptance of raw and processed data from the Tier-0 centre, keeping up with data acquisition; • recording and maintenance of raw and processed data on permanent mass storage; • provision of managed disk storage providing permanent and temporary data storage for files and databases; • operation of a data-intensive analysis facility; • provision of other services according to agreed experiment requirements • provision of high capacity network services for data exchange with the Tier-0 centre, as part of an overall plan agreed between the experiments, Tier-1 and Tier-0 centres; • provision of network services for data exchange with Tier-1 and selected Tier-2 centres, as part of an overall plan agreed between the experiments, Tier-1 and Tier-2 centres; • administration of databases required by experiments at Tier-1 centres;

  21. ARC-connected resources for DC2 Totals at peak: • 7 countries • 22 sites • ~3000 CPUs • dedicated ~700 • 7 Storage Services (in RLS) • few more storage facilities • ~12TB • ~1FTE (1-3 persons) in charge of production • At most 2 executor instances simultaneously

  22. ARC performance in ATLAS DC2 • Total # of successful jobs: 42202(as of September 25, 2004) • Failure rate before ATLAS ProdSys manipulations: 20% • ~1/3 of failed jobs did not waste resources • Failure rate after: 35% • Possible reasons: • Dulcinea failing to add DQ attributes in RLS • DQ renaming • Windmill re-submitting good jobs

  23. Failure analysis • Dominant problem: hardware accidents

More Related