240 likes | 326 Views
A Distributed Tier-1. An example based on the Nordic Scientific Computing Infrastructure. GDB meeting – NIKHEF/SARA 13th October 2004 John Renner Hansen – Niels Bohr Institute. With contributions from Oxana Smirnova, Peter Villemoes and Brian Vinter.
E N D
A Distributed Tier-1 An example based on the Nordic Scientific Computing Infrastructure GDB meeting – NIKHEF/SARA 13th October 2004 John Renner Hansen – Niels Bohr Institute With contributions from Oxana Smirnova, Peter Villemoes and Brian Vinter
Basis for a distributed Tier-1 structure • External connectivity • Internal connectivity • Computer and Storage capacity • Maintenance and operation • Long term stability
NORDUnet network in 2003 General Internet 5G NETNOD 3.5G 155M RUNNet 622M General Internet 2.5G NASK 12M GÉANT 10G (Oct ’03)
NorthernLight Stockholm Oslo Helsinki Dec 2003 2.5G links connected to “ONS boxes” giving 2 GE channels between endpoints Aug 2003 Copenhagen NetherLight Amsterdam
NORDUNet was represented at the NREN-TIER1 Meeting Paris, Roissy Hilton, 12:00-17:00, 22 July 2004 by Peter Villemoes
Denmark / Forskningsnet • Upgrading from 622Mbit/s to a 2.5 Gbit/s ring structure finished: • Copenhagen-Odense-Århus-Aalborg- and back via Göteborg to Copenhagen • Network research • setting up a Danish national IPv6 activity • dark fibre through the country • experimental equipment for 10GE channels
Finland / FUNET • Upgraded to 2.5G already in 2002 • Upgrading backbone routers to 10G capability
Norway / UNINETT • Network upgraded to 2.5G between major universities, • UNINETT is expanding to more services and organisations:
Sweden / SUNET • 10G resilient nationwide network since Nov 2002 • all 32 universities have 2.5G access • Active participation in SweGrid, Swedish Grid Initiative
Who • Denmark Danish Center for Grid Computing - DCGC Danish Center for Scientific Computing -DCSC
Who • Denmark • Finland - CSC
Who • Denmark • Finland • Norway - NorGrid
Who • Denmark • Finland • Norway • Sweden - SweGrid
Denmark • Two collaborating Grid projects • Danish Centre for Scientific Computing Grid • DCSC-Grid spans the four DCSC sites and thus unify the resources, PC-clusters, IBM-Regatta, SGI-Enterprise, … within DCSC • Danish Centre for Grid Computing • Is the the national Grid project • DCSC Grid is a partner in DCGC
Finland • Remains centred about CSC • CSC participates in NDGF and NGC • A Finnish Grid will probably be created • This Grid will focus more on accessing CSC resources with local machines
Norway • NOTUR Emerging Technologies on Grid Computing is the main mover • Oslo-Bergen “mini-Grid” in place • Trondheim and Tromsø should be joining Curently under reorganization
Sweden • SweGrid is a very ambitious project • 6 clusters have been created for the purpose of SweGrid each equipped with a 100 PCs and a large disk system • Large support and education organization is integrated in the plans
NOS-N Research Councils DK SF S N Nordic Data Grid Facility Nordic Project Core Group Project Director 4 Post Doc.s Steering Group 3 members per country 1 R.C. Civil Servant 2 Scientists Nordic Data Grid Facility • Create the basis for a common Nordic Data Grid Facility 2. Coordinate Nordic GridActivities
Services provided by the Tier-1 Regional Centres • acceptance of raw and processed data from the Tier-0 centre, keeping up with data acquisition; • recording and maintenance of raw and processed data on permanent mass storage; • provision of managed disk storage providing permanent and temporary data storage for files and databases; • operation of a data-intensive analysis facility; • provision of other services according to agreed experiment requirements • provision of high capacity network services for data exchange with the Tier-0 centre, as part of an overall plan agreed between the experiments, Tier-1 and Tier-0 centres; • provision of network services for data exchange with Tier-1 and selected Tier-2 centres, as part of an overall plan agreed between the experiments, Tier-1 and Tier-2 centres; • administration of databases required by experiments at Tier-1 centres;
ARC-connected resources for DC2 Totals at peak: • 7 countries • 22 sites • ~3000 CPUs • dedicated ~700 • 7 Storage Services (in RLS) • few more storage facilities • ~12TB • ~1FTE (1-3 persons) in charge of production • At most 2 executor instances simultaneously
ARC performance in ATLAS DC2 • Total # of successful jobs: 42202(as of September 25, 2004) • Failure rate before ATLAS ProdSys manipulations: 20% • ~1/3 of failed jobs did not waste resources • Failure rate after: 35% • Possible reasons: • Dulcinea failing to add DQ attributes in RLS • DQ renaming • Windmill re-submitting good jobs
Failure analysis • Dominant problem: hardware accidents