70 likes | 197 Views
management. Grid-Computing Infrastructure & Services (GIS, H.Marten). 1.2. 2.6. FZK tape archive. F Z K b a c k b o n e. software server interactive login ALICE. 1.5.1. 1.6. software installation service. certification authority. compute nodes CPU intensive. Data
E N D
management Grid-Computing Infrastructure & Services (GIS, H.Marten) 1.2 2.6 FZK tape archive F Z K b a c k b o n e software server interactive login ALICE 1.5.1 1.6 software installation service certification authority compute nodes CPU intensive Data tape archive PBytes G R I D b a c k b o n e 1.3.1 1.8 1.12 1.1 software server interactive login ATLAS licence service firewall 1.4.1 1.7.1 Data server 1 compute nodes I/O intensive . . . . 8x 1.3.2 infrastructure S A N f a b r i c 1.9 1.13 2.1 ALICE Grid services Data server 2 ATLAS .... 1.7.2 Data Pool Volume Management user support . . . 1.4.2 1.14 compute nodes communication intensive ?? 2.2 Data server n training and education software server interactive login other science system monitoring & management 1.4.3 2.3 1.3.9 1.10 compute nodes TESTING; R&D documentation network address translation 1.7 2.4 Data server R&D software server interactive login TEST; R&D 1.11 1.4.4 R & D batch and background; invisible for users private network ?? 1.3.10 2.5 Data Import Data Export Tier-0/1/2 RDCCG – Regional Computing and Data Center Germany WAN 23-nov-2001
RDCCG Evolution (available capacity) 30% rolling upgrade each year after 2007 Networking Evolution 2002 - 2005 1) RDCCG to CERN/FermiLab/SLAC (Permanent point to point): 1 GBit/s – 10 GBit/s => 2 GBit/s could be arranged on a very short timescale 2) RDCCG to general Internet: 34 MBit/s – 100 MBit/s => Current situation, generally less affordable than 1) FTE Evolution 2002 - 2005 Support: 5 - 30 Development: 8 - 10 New office building to accommodate 130 FTE in 2005
Installation and Update of OSThe idea is to set up a dedicated installation server for each rack of CPUsCurrently evaluating the Rocks-Toolkit MonitoringNo use of dedicated monitoring hardware and/or management busTivoli management suite in broad use for already existing operations-> Further study and deployment of tools necessary Management Tools
Certification Authority • Hardware for CA available • Delivery of certificates tested • CP and CPS draft document available • “FZK-Grid CA, Certificate Policy and Certification Practice Statement” • Organization of German virtual organizations • FZK-Grid CA known to Globus and accepted by DataGrid • Policy for acceptance of foreign certificates needs to be discussed Authorization and Registration Procedure • One account manager for each experiment • WWW form to be filled out, signed by experiment account manager and send to FZK by FAX • One experiment specific account per user • One system administrator account per experiment • No super user privileges, sorry
GSI DarmstadtTier-2 will perform e.g. interactive analysis on subsamples (PROOF) Offer support for countries w/o dedicated Tier-1e.g. Russia (MSU), … Tier-2 business needs to be further discussed and developed within Germany Tier-2 Centers LHC Data Challenges • Commitment to fully take part in all data challenges • MoU to integrate non-LHC hardware upon actual needcurrently exercised with ALICE: use BaBar farm for MC challenge-> Synergetic effect until 2004
Department for Grid-Computing and e-Science (GES, M.Kunze): • Middleware and Application Development • Special interest in ROOT core development and support • Meta-Data catalogue • Definition and implementation of interfaces to Globus and .NET • Evolution of object service • Transaction based processing • Interactive analysis (PROOF) • Special interest in database technology • Data mining techniques (Agents and Neural Networks) • Deployment of intelligent server side procedures • Deployment of XML and SOAP interface technology • Involvement into • BaBar MC production and distributed analysis (GanaTools) • LHC test beds (currently ALICE, ATLAS coming up) • CrossGrid • National Grid projects (e.g. DAS-Grid w/ ZIB, UniDo, PC2, FZ Jülich)
SAN SANergy, qfs, gfs, ... Running SAN for Grid Compute Nodes:I/O clients for nfs, cifs, root, scp, gass, ftp, s-ftp, http, https, ... Grid backbone 128 GBit/s I/O server for nfs, cifs, root, scp, gass, ftp, s-ftp, http, https, ... MDC High Throughput Cluster: Direct SAN access (>300 MB/s) Good for interactive Data-Mining: 1 TB in a few minutes tape robotics disk subsystems