310 likes | 431 Views
Ian Bird LHCC Referees Meeting; CERN, 12 th March 2013. WLCG Update. ASGC - Taipei. On Feb 21, Academia Sinica , Taipei informed CERN DG that they intended to withdraw support for CMS as a Tier 1 and as a Tier 2 at ASGC CMS have no physicists in Academia Sinica
E N D
Ian Bird LHCC Referees Meeting; CERN, 12thMarch 2013 WLCG Update Ian.Bird@cern.ch
ASGC - Taipei Ian.Bird@cern.ch • On Feb 21, Academia Sinica, Taipei informed CERN DG that they intended to withdraw support for CMS as a Tier 1 and as a Tier 2 at ASGC • CMS have no physicists in Academia Sinica • ASGC has lost part of its funding from the National Science Council • Response sent March 6 • Asking for clarification of intended timescales and to inform WLCG OB and RRB
Status of new Tier 1s Ian.Bird@cern.ch
Update on KISTI-GSDC • 1512 cores (63 WNs) with 3GB memory/core • 1.042 PB Disks: old SE(100TB) will be removed; new storage will be delivered • 1 Gbps Network to be upgraded to 2Gbps in coming April • EMI-2 Migration done and on production (better behaviour than gLite-3.2) • Tape • 1 PB capacity (expandable up to 3 PB, 2GB/s throughput) with 200 TB xrootd pool + 275 TB tape buffer • Functional tests done, Policy tuning on-going for xrootd, GPFS • Periodic test by ALICE shows good -> p-Pb data transfer scheduled to start on 7th March
JINR CMS Tier-1 progress • Disk & server installation and tests: done • Tape system installation: done • Organization of network infrastructure and connectivity to CERN via GEANT (2 Gbps):done • LHC OPN integration: postponed* • Registration in GOC DB and APEL: done • Tests of WLCG services via Nagios: done • CMS-specific tests: in process * restructurising local network to ease integration with LHC OPN
RRC-KI Tier-1 progress • Disk & server installation and tests: done • Tape system installation: done • Network connectivity to CERN via GEANT (2 Gbit/sec): done • LHC OPN integration: postponed* • Registration in GOC DB and APEL: done • Tests of WLCG services via Nagios: done • ATLAS-specific tests: in process, have preliminary results that look good • * restructurising our network to ease integration with LHC OPN
ATLAS-specific activity • Prepared services and data storage for the reprocessing of 2011 2.76 TeV reprocessing • Currently commissioning Tier-1 resource for ATLAS: • Stress tests via HammerCloud • Reconstruction tests submitted by ATLAS experts • Functional tests via Panda and DDM • Transferred 54 TB of input data to our storage element with transfer efficiency around 90%
Overall status Ian.Bird@cern.ch
Data accumulated Data written into Castor per week Volume of CERN archive Ian.Bird@cern.ch 2012/13 Data written 2012/13 Data read
CPU usage Ian.Bird@cern.ch
Operations Ian.Bird@cern.ch
Resource requests 2013-15 Ian.Bird@cern.ch
Preparation for April RRB Ian.Bird@cern.ch • Met with RSG chair + LHCC referees in February • Agreed an assumed running scenario for 2015, in order to produce requirements • 5.10^6 s of operation; less pileup in first part, increased in 2nd part of year • Discussed how to present 2015 needs • Potentially significantly increased trigger rates • Aggressive work during LS1 to enable the needs to be met with reasonable resource increases • The estimates for 2015 are now somewhat less than the x2 feared; except for tape needs; • Ramp-up during 2015 – clearly early in the year less will be needed • [Also agreed to set the disk usage “efficiency” to 1 to avoid confusion]
2013 2015 resources CPU Disk Tape 2013: pledge OR actual installed capacity if higher Ian.Bird@cern.ch
Use of Tier 0 and HLT during LS1 Ian.Bird@cern.ch • During LS1; the Tier 0 resources will be used for analysis as well as re-processing • This is a different use pattern; requires more disk that Tier 0 functions • Also the HLT farms will be extensively used to supplement the grid resources
Tier 0 Upgrades Ian.Bird@cern.ch
WIGNER Data Center After full refurbishment, hosting CERN Tier-0 From 1 january 2013 IT Department Meeting
Data Center Layout & ramp-up IT Department Meeting
Connectivity (100 Gbps) IT Department Meeting
513 Consolidation Project: Goals IT Department Meeting • Solve the cooling issue for the critical UPS room • New UPS systems in a different location • Increase critical UPS capacity to 600kW • Restore N+1 redundancy for both critical and physics UPS systems • Secure cooling for critical equipment when running on UPS and extend stored cooling capacity for physics when on diesel • Decouple the A/C for CC from the adjacent office building • Increase overall power capacity to 3.5MW
Planning for the future Ian.Bird@cern.ch
EC projects Ian.Bird@cern.ch • EMI (middleware) ends April 2013 • EGI-SA3 (support for Heavy User communities) – ends April 2013 • Although EGI-Inspire continues for 1 more year • These have impact on CERN groups supporting the experiments, as well as NGI support • Consequences: • Re-prioritisation of functions is needed • Need to take action now if we anticipate attracting EC money in the future • But there is likely to be a gap of ~1 year or more
Short term: Consolidation of activities at CERN • WLCG operations, service coordination, support • Consolidate related efforts (daily ops, integration, deployment, problem follow-up etc) • Broader than just CERN – encourage other labs to participate • Common solutions • Set of activities benefitting several experiments. Coordinates experiment work as well as IT-driven work. Experiments see this as strategic for the future; beneficial for long term sustainability • Grid monitoring • Must be consolidated (SAM/Dashboards). Infrastructure becoming more common; focus on commonalities, less on experiment-specifics • Grid swdevelopment+support • WLCG DM tools (FTS, DPM/LFC, Coral/COOL, etc), information system; Simplification of build, packaging, etc. open source community processes; (See WLCG doc)
Longer term Ian.Bird@cern.ch • Need to consider how to engage with EC and other potential funding sources • However, in future boundary conditions will be more complex: (e.g. for EC) • Must demonstrate how we benefit other sciences and society at large • Must engage with Industry (e.g. via PPP) • HEP-only proposals unlikely to succeed • Also it is essential that any future proposal is fully engaged in by CERN (IT+PH) and experiments and other partners
Background Ian.Bird@cern.ch • Requested by the LHCC in December: need to see updated computing models before Run 2 starts • A single document to: • Describe changes since the original TDRs (2005) in • Assumptions, models, technology, etc. • Emphasise what is being done to adapt to new technologies, to improve efficiency, to be able to adapt to new architectures, etc. • Describe work that still needs to be done • Use common formats, tables, assumptions, etc • 1 document rather than 5
Timescales Ian.Bird@cern.ch • Document should describe the period from LS1 – LS2 • Estimates of evolving resource needs • In order to prepare for 2015, a good draft needs to be available in time for the Autumn 2013 RRB, so needs to be discussed at the LHCC in September: • Solid draft by end of summer 2013 (!) • Work has started • Informed by all of the existing work from the last 2 years (Technical Evolution groups, Concurrency forum, Technology review of 2012)
Opportunities Ian.Bird@cern.ch • This document gives a framework to: • Describe significant changes and improvements already made • Stress commonalities between experiments – and drive strongly in that direction • Significant willingness to do this • Describe the models in a common way – calling out differences • Make a statement about the needs of WLCG in the next 5 years (technical, infrastructure, resources) • Potentially review the organisational structure of the collaboration • Review the implementation: scale, quality of service of sites/Tiers; archiving vs processing vs analysis activities • Raise concerns: • E.g. staffing issues; missing skills;
Draft ToC Ian.Bird@cern.ch Preamble/introduction Experiment computing models Technology review and outlook Challenges – the problem being addressed Distributed computing Computing services Software activities and strategies Resource needs and expected evolution Collaboration organisation and management
Summary Ian.Bird@cern.ch • WLCG operations in good shape, • Use of computing system by experiments regularly fills available resources • Concern over resources vs requirements in the future • Tier 0 consolidation work close to complete, deployment beginning • Important to take concrete steps now for future planning for support