210 likes | 329 Views
Tier 2 Prague Institute of Physics AS CR. Status and Outlook J. Chudoba , M. Elias, L. Fiala, J. Horky, T . Kouba, J. Kundrat , M. Lokajicek , J . Svec , P. Tylka. Outline. Institute of Physics AS CR ( FZU ) Computing Cluster Networking LHCONE Looking for new resources
E N D
Tier 2 PragueInstitute of Physics AS CR Status and Outlook J. Chudoba, M. Elias, L. Fiala, J. Horky, T. Kouba, J. Kundrat, M. Lokajicek, J. Svec, P. Tylka NEC2013 Varna M. Lokajicek
Outline • Institute ofPhysics AS CR (FZU) • Computing Cluster • Networking • LHCONE • Looking for new resources • CESNET National Storage Facility • IT4I supercomputing project • Outlook
Institute of Physics AS CR (FZU) • Institute ofPhysicsofthe Academy of the Czech Republic • 2 locations in Prague, 1 in Olomouc • In 2012: 786 employees (281 researchers + 78 doctoral students) • 6 Divisions • Division of Elementary Particle Physics • Division of Condensed Matter Physics • Division of Solid State Physics • Division of Optics • Division of High Power Systems • ELI Beamlines Project Division • Department of Networking and Computing Techniques (SAVT)
FZU - SAVT • Institute’s networking and computing service department • Several server rooms • Computing clusters Golias – Particle physics, Tier2 • Few nodes from already before EDG • WLCG iMoU 4 July 2003 (interim) • New server room 1 November 2004 • WLCG MoU from 28 April 2008 • Dorje – solid state, condensed matterLuna, Thsun, smaller group clusters
Main server room • Main server room (in FZU, Na Slovance) • 62 m2, ~20 racks, 350 kVA motor generator, 200 + 2 x 100 kVA UPS, 108 kW air cooling, 176 kW water cooling • continuous changes • hosts computing servers and central services
Cluster Golias • Upgraded every yearseveral (9) sub-clusters of the identical HW • 3800 cores, 30 700 HS06 • 2 PB disk space • Tapes used only for local backups (125 LTO4, max 500 cassettes) • Serving: ATLAS, ALICE, D0 (NOvA), Auger, STAR, … • WLCG Tier2Golias@FZU + xrootdservers@REZ (NPI)
Utilization • Very high average utilization • Several different projects, different tools for production • D0 – production submitted locally by 1 user • ATLAS – panda, ganga, local users; DPM • ALICE – VO box; xrootd 3.5 k D0 ATLAS ALICE
2012 D0, ATLAS and ALICE usage • D0 • 290 M tasks • 90 MHEPSPEC06 hours • 13% contribution to D0 2012 Data transfers inside farm - month means to and from working nodes in TB • ATLAS • 2,2 M tasks • 90 MHEPSEPC06 hours,1,9 PB disk space • Data transfer 1,2 PB tofarm0,9 PB fromfarm • 2% contribution to ATLAS • ALICE • 2 M simulation tasks • 60 MHEPSEC06 hours • Data transfer 4,7 PB tofarmand 0,5 PB fromfarm • 5% our contribution to ALICE tasks processing • 140TB disk space in INF (Tier3)
Network -CESNET, z. s. p. o. • 1 Gbps FNAL, BNL, Taipei • 10 Gbps to commodity network • 1-10 Gbps to Tier3 collaborating institutes • FZU Tier2 Network connections • 10 Gbps LHCONE (GEANT), 18 July 2013 • 10 Gbps KIT from 1st Sept 2013 http://netreport.cesnet.cz/netreport/hep-cesnet-experimental-facility2/
LHCONE - Network transition • Link to KIT saturated at 1 Gbps E2E line • LHCONE from 18 July 2013 over 10 Gbps infrastructure • Relieves also the commodity network 10 Gbps
Atlas tests • Testing upload speed of files > 1 GB to all Tier1 centra • After LHCONE connection only 2 sites with < 5MB/s • Prague Tier2 ready for validation as T2D 30 60
LHCONE – trying to understand monitoring • L Prague – DESY Very asymmetric throughput LHCONE line cut DESY – Prague LHCONE optical line cut At 4:00One way latency improved
International contribution of Prague center toATLAS + ALICE centra T2 LCG • http://accounting.egi.eu/ • Grid + localtasks • Long term slide down until we received regular financing in 2008 • Original 3% target is not achievable with current financial resources • Necessary to look for other resources
Remote storage • CESNET - Czech NREN + other services • New project: National storage facility • Three distributed HSM based storage sites • Designed for research and science community • 100TB forboth ATLAS and Auger experiments offered • Implemented as remote Storage Element with dCache • disk <-> tape migration FZU Tier-2 in Prague 100km FZU<->Pilsen - 10Gbit link with ~3.5ms latency CESNET storage site in Pilsen
Remotestorage Influence of distributing a Tier-2 data storage on physics analysis • TTreeCache in ROOT helps a lot – both for local and for remote transfers • TTreeCachedremote jobs faster than local ones without the cache
Outlook • In 2015 after LHC start up • Higher data production • Flat financing not sufficient • Computingcan become an item of M&O A (Maintenance &Operations cat. A) • Search for new financial resources or new unpaid capacities necessary • CESNET • Crucial free delivery of network infrastructure • Unpaid External storage, how long? • IT4I, Czech supercomputing project search for computing capacities (free cycles), relying on other project to find the way how to use them
16th International workshop on Advanced Computing and Analysis Techniques in physics (ACAT) • http://www.particle.cz/acat2014 • Topics • Computing Technology for Physics Research • Data Analysis - Algorithms and Tools • Computations in Theoretical Physics: Techniques and Methods