200 likes | 215 Views
Steve Lloyd Tony Doyle John Gordon. GridPP Presentation to PPARC e-Science Committee 26 July 2001. Outline. Component Model Resource Allocation and Funding Scenarios Int l Financial Comparisons Int l Grid Collaborations Grid Architecture(s) Links with Industry Summary Addendum:
E N D
Steve Lloyd Tony Doyle John Gordon GridPP Presentation to PPARC e-Science Committee 26 July 2001
Outline • Component Model • Resource Allocation and Funding Scenarios • Intl Financial Comparisons • Intl Grid Collaborations • Grid Architecture(s) • Links with Industry • Summary • Addendum: 1. VISTA and GridPP 2. GridPP monitoring page e-Science Presentation
GridPP Proposal • GridPP = Vertically integrated programme • = component model... • Input to development of £15-20M funding scenarios e-Science Presentation
GridPP Workgroups Technical work broken down into several workgroups - broad overlap with EU DataGrid A - Workload Management Provision of software that schedule application processing requests amongst resources F - Networking Network fabric provision through to integration of network services into middleware G - Prototype Grid Implementation of a UK Grid prototype tying together new and existing facilities B - Information Services and Data Management Provision of software tools to provide flexible transparent and reliable access to the data H - Software Support Provide services to enable the development, testing and deployment of middleware and applications at institutes C - Monitoring Services All aspects of monitoring Grid services I - Experimental Objectives Responsible for ensuring development of GridPP is driven by needs of UK PP experiments D - Fabric Management and Mass Storage Integration of heterogeneous resources into common Grid framework J - Dissemination Ensure good dissemination of developments arising from GridPP into other communities and vice versa E - Security Security mechanisms from Certification Authorities to low level components e-Science Presentation
J 2.6% I: 11.9% Experiment CERN Staff Objectives 27.0% H*: 5.4% Software Support H: 3.2% G: Prototype Grid 9.7% CERN Hardware 6.8% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 15.3% 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A Components 1-4: £21M e-Science Presentation
J 2.6% I: 11.9% Experiment CERN Objectives H*: 5.4% Software Support 96.3% G: Prototype Grid 9.7% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A £20M Project £7.1m £6.7m H: 3.2% £3.2 £2.9 e-Science Presentation
J 2.6% CERN H*: 5.4% Software Support 90.0% G: Prototype Grid 9.7% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A £17M Project I: £2.49m £1.2m Experiment £7.1m £6.7m £6.0m Objectives H: 3.2% £3.2 £2.9 £2.45m e-Science Presentation
Experiment Objectives 50% reduction? 23 SY • Vertically integrated programme? • Broken component model… • Specific experiments or overall reduction? • To be determined by Experiments Board e-Science Presentation
CERN (Component 3) 10% reduction? 3.1 SY Basic Grid functionality: UK-CERN integrated programme - synergies, but cuts here will impact… e-Science Presentation
CERN (Component 4) 50% reduction? 11 SY Experiments support: similar conclusions to UK-based programme Non-UK funding dependencies? 15% reduction? £0.2M +HARDWARE Pro-rata reduction on disk, tape, CPU... e-Science Presentation
Workload/Data Management 10% reduction? 1.2 SY Reduced long-term programme? e.g. scheduler optimisation (WG A) query optimisation (WG B) … or overall reduction? e-Science Presentation
J 2.6% CERN H*: 5.4% Software Support 90.0% G: Prototype Grid 9.7% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A £15M Project I: £2.49m £0 Experiment £5m Objectives H: 3.2% £3.2 £2.9 £2.45m e-Science Presentation
£15M Project Summary • Even a £21-20M reduction is not trivial.. • EU DataGrid commitments are built in • Focus on CERN and UK Capital as largest single items, then reduce workgroup allocations • £17M budget cuts hard into the project • Examples are based on original Component Model • £15M budget is impossible within the Component Model • A fixed allocation help in planning the start-up phase e-Science Presentation
Tier-1 at FNAL and 5 Tier-2 centres • Prototype built during 2000-04, with full deployment during 2005-7 • Staff estimates for the Tier-1 centre are 14 FTE by 2003, reaching 35 FTE in 2007. • Integrated costs to 2006 are $54.7M • excluding, GriPhyN and PPDG • Tier-1 RC for all 4 LHC experiments at CC-IN2P3 in Lyon • BaBar TierA • an LHC prototype starting now • National Core Grid (2M€/year) • INFN National Grid based round EU-DataGrid • Tier-1 RC and a prototype starting now in CNAF, Bologna • 15.9M€ is allocated during 2001-3 for Tier-1 hardware alone • Tier1 staff rising to 25 FTE by 2003 • 10 Tier2 centres at 1M€/year • Tier1 starting up at Karlsruhe • BaBar TierB at Karlsruhe • Tier2 for ALICE at Darmstadt • No national Grid - project led • Atlas plans very similar to CMS with costs foreseen to be the same • Tier1 at Brookhaven International Comparisons PP Grids under development • France • Germany • Italy • US • CMS • Atlas e-Science Presentation
International Comparisons Summary - different countries, different models • France & Germany budget for hardware, assume staff • Italy - lots of hardware and staff • US - funds split between Tier1/2, Universities, infrastructure, and R&D • Italy > UK ~ France (EU) ~US (GriPhyN, PPDG and iVDGL characteristics within GridPP: single UK programme) ~ e-Science Presentation
GridPP Architecture The DataGrid Architecture Version 2 German Cancio, CERN Steve M. Fisher, RAL Tim Folkes, RAL Francesco Giacomini, INFN Wolfgang Hoschek, CERN Dave Kelsey, RAL Brian L. Tierney, LBL/CERN July 2, 2001 • Based on EU DataGrid developments feeding into GGF • Status: Version 2 (2/7/01) • Key elements: • Evolutionary capability • Service via Protocols and Client APIs • Representation using UML (TogetherSoft) • Defines responsibilities of Work Packages • Built from Infrastructure • Based on PP Use Cases (applies to GridPP) e-Science Presentation
The Grid and Industry • Help us develop the Grid: • Supply hardware - PCs, Disks, Mass Storage, Networking etc • Supply software, middleware, management systems, databases etc • Use the Grid for themselves: • Collaborative Engineering • Massive simulation • Federating their own worldwide databases • Sell or develop the Grid for others: • Computation Services, Data services etc e-Science Presentation
Summary • Balanced exploitation programme costs £21M • £20M-£17M-£15M 3-year funding scenarios examined • £20M = maintains balanced programme • £17M = reduced experimental objectives • £15M = eliminates experimental objectives • Final balance depends on funding allocation • Emphasis on vertical integration: component model • International comparisons: Italy > UK ~ France (EU) ~US (GriPhyN, PPDG and iVDGL characteristics within GridPP: single UK programme) • Contacts established with GriPhyN, PPDG and iVDGL • InterGrid Co-ordination Group in development • Architecture defined by GGF via lead in DataGrid • Industry links: emphasis on partnership ~ e-Science Presentation
GridPP and VISTA • Astrogrid will federate VISTA data with other large databases elsewhere • this requires that VISTA data has already been processed and catalogues and images are available. • VISTA have a proposal (e-VPAS) that concentrates on producing the databases on which the Astrogrid tools will work. This work has much in common with GridPP: • a similar timescale • very large data flows from one remote site • many distributed users • reprocessing of data • utilization of distributed computing resources • GridPP have started discussions with VISTA and EPCC (GenGrid) as to how we can collaborate and share expertise and middleware e-Science Presentation
GridPP Monitoring Page • Various sites now set up with UK Globus certificates • Grid Monitoring • Polls Grid test-bed sites via globus-job-run command • Runs basic script producing XML encoded status information • Load average and timestamp information retrieved • Current status and archived load information is plotted... • To be done... • JAVA CoG kit being investigated (more robust) • Simple monitoring system to verify test-bed timestamps (in case not everyone is using NTP) • Integrate with the Grid Monitoring Architecture • Incorporate current network bandwidth measurements into graphical system • Automatic notification system e-Science Presentation