1 / 20

GridPP Presentation to PPARC e-Science Committee 26 July 2001

Steve Lloyd Tony Doyle John Gordon. GridPP Presentation to PPARC e-Science Committee 26 July 2001. Outline. Component Model Resource Allocation and Funding Scenarios Int l Financial Comparisons Int l Grid Collaborations Grid Architecture(s) Links with Industry Summary Addendum:

dtoliver
Download Presentation

GridPP Presentation to PPARC e-Science Committee 26 July 2001

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Steve Lloyd Tony Doyle John Gordon GridPP Presentation to PPARC e-Science Committee 26 July 2001

  2. Outline • Component Model • Resource Allocation and Funding Scenarios • Intl Financial Comparisons • Intl Grid Collaborations • Grid Architecture(s) • Links with Industry • Summary • Addendum: 1. VISTA and GridPP 2. GridPP monitoring page e-Science Presentation

  3. GridPP Proposal • GridPP = Vertically integrated programme • = component model... • Input to development of £15-20M funding scenarios e-Science Presentation

  4. GridPP Workgroups Technical work broken down into several workgroups - broad overlap with EU DataGrid A - Workload Management Provision of software that schedule application processing requests amongst resources F - Networking Network fabric provision through to integration of network services into middleware G - Prototype Grid Implementation of a UK Grid prototype tying together new and existing facilities B - Information Services and Data Management Provision of software tools to provide flexible transparent and reliable access to the data H - Software Support Provide services to enable the development, testing and deployment of middleware and applications at institutes C - Monitoring Services All aspects of monitoring Grid services I - Experimental Objectives Responsible for ensuring development of GridPP is driven by needs of UK PP experiments D - Fabric Management and Mass Storage Integration of heterogeneous resources into common Grid framework J - Dissemination Ensure good dissemination of developments arising from GridPP into other communities and vice versa E - Security Security mechanisms from Certification Authorities to low level components e-Science Presentation

  5. J 2.6% I: 11.9% Experiment CERN Staff Objectives 27.0% H*: 5.4% Software Support H: 3.2% G: Prototype Grid 9.7% CERN Hardware 6.8% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 15.3% 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A Components 1-4: £21M e-Science Presentation

  6. J 2.6% I: 11.9% Experiment CERN Objectives H*: 5.4% Software Support 96.3% G: Prototype Grid 9.7% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A £20M Project £7.1m  £6.7m H: 3.2% £3.2  £2.9 e-Science Presentation

  7. J 2.6% CERN H*: 5.4% Software Support 90.0% G: Prototype Grid 9.7% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A £17M Project I: £2.49m  £1.2m Experiment £7.1m  £6.7m  £6.0m Objectives H: 3.2% £3.2  £2.9  £2.45m e-Science Presentation

  8. Experiment Objectives 50% reduction? 23 SY • Vertically integrated programme? • Broken component model… • Specific experiments or overall reduction? • To be determined by Experiments Board e-Science Presentation

  9. CERN (Component 3) 10% reduction? 3.1 SY Basic Grid functionality: UK-CERN integrated programme - synergies, but cuts here will impact… e-Science Presentation

  10. CERN (Component 4) 50% reduction? 11 SY Experiments support: similar conclusions to UK-based programme Non-UK funding dependencies? 15% reduction? £0.2M +HARDWARE Pro-rata reduction on disk, tape, CPU... e-Science Presentation

  11. Workload/Data Management 10% reduction? 1.2 SY Reduced long-term programme? e.g. scheduler optimisation (WG A) query optimisation (WG B) … or overall reduction? e-Science Presentation

  12. J 2.6% CERN H*: 5.4% Software Support 90.0% G: Prototype Grid 9.7% UK Managers 1.9% 1.5% 1.9% Work Groups A - F 1.7% F* UK Capital 1.5% F 2.7% E 0.6% 1.1% 1.9% D* 1.5% 0.4% 1.4% D C* C B* B A* A £15M Project I: £2.49m  £0 Experiment £5m Objectives H: 3.2% £3.2  £2.9  £2.45m e-Science Presentation

  13. £15M Project Summary • Even a £21-20M reduction is not trivial.. • EU DataGrid commitments are built in • Focus on CERN and UK Capital as largest single items, then reduce workgroup allocations • £17M budget cuts hard into the project • Examples are based on original Component Model • £15M budget is impossible within the Component Model • A fixed allocation help in planning the start-up phase e-Science Presentation

  14. Tier-1 at FNAL and 5 Tier-2 centres • Prototype built during 2000-04, with full deployment during 2005-7 • Staff estimates for the Tier-1 centre are 14 FTE by 2003, reaching 35 FTE in 2007. • Integrated costs to 2006 are $54.7M • excluding, GriPhyN and PPDG • Tier-1 RC for all 4 LHC experiments at CC-IN2P3 in Lyon • BaBar TierA • an LHC prototype starting now • National Core Grid (2M€/year) • INFN National Grid based round EU-DataGrid • Tier-1 RC and a prototype starting now in CNAF, Bologna • 15.9M€ is allocated during 2001-3 for Tier-1 hardware alone • Tier1 staff rising to 25 FTE by 2003 • 10 Tier2 centres at 1M€/year • Tier1 starting up at Karlsruhe • BaBar TierB at Karlsruhe • Tier2 for ALICE at Darmstadt • No national Grid - project led • Atlas plans very similar to CMS with costs foreseen to be the same • Tier1 at Brookhaven International Comparisons PP Grids under development • France • Germany • Italy • US • CMS • Atlas e-Science Presentation

  15. International Comparisons Summary - different countries, different models • France & Germany budget for hardware, assume staff • Italy - lots of hardware and staff • US - funds split between Tier1/2, Universities, infrastructure, and R&D • Italy > UK ~ France (EU) ~US (GriPhyN, PPDG and iVDGL characteristics within GridPP: single UK programme) ~ e-Science Presentation

  16. GridPP Architecture The DataGrid Architecture Version 2 German Cancio, CERN Steve M. Fisher, RAL Tim Folkes, RAL Francesco Giacomini, INFN Wolfgang Hoschek, CERN Dave Kelsey, RAL Brian L. Tierney, LBL/CERN July 2, 2001 • Based on EU DataGrid developments feeding into GGF • Status: Version 2 (2/7/01) • Key elements: • Evolutionary capability • Service via Protocols and Client APIs • Representation using UML (TogetherSoft) • Defines responsibilities of Work Packages • Built from Infrastructure • Based on PP Use Cases (applies to GridPP) e-Science Presentation

  17. The Grid and Industry • Help us develop the Grid: • Supply hardware - PCs, Disks, Mass Storage, Networking etc • Supply software, middleware, management systems, databases etc • Use the Grid for themselves: • Collaborative Engineering • Massive simulation • Federating their own worldwide databases • Sell or develop the Grid for others: • Computation Services, Data services etc e-Science Presentation

  18. Summary • Balanced exploitation programme costs £21M • £20M-£17M-£15M 3-year funding scenarios examined • £20M = maintains balanced programme • £17M = reduced experimental objectives • £15M = eliminates experimental objectives • Final balance depends on funding allocation • Emphasis on vertical integration: component model • International comparisons: Italy > UK ~ France (EU) ~US (GriPhyN, PPDG and iVDGL characteristics within GridPP: single UK programme) • Contacts established with GriPhyN, PPDG and iVDGL • InterGrid Co-ordination Group in development • Architecture defined by GGF via lead in DataGrid • Industry links: emphasis on partnership ~ e-Science Presentation

  19. GridPP and VISTA • Astrogrid will federate VISTA data with other large databases elsewhere • this requires that VISTA data has already been processed and catalogues and images are available. • VISTA have a proposal (e-VPAS) that concentrates on producing the databases on which the Astrogrid tools will work. This work has much in common with GridPP: • a similar timescale • very large data flows from one remote site • many distributed users • reprocessing of data • utilization of distributed computing resources • GridPP have started discussions with VISTA and EPCC (GenGrid) as to how we can collaborate and share expertise and middleware e-Science Presentation

  20. GridPP Monitoring Page • Various sites now set up with UK Globus certificates • Grid Monitoring • Polls Grid test-bed sites via globus-job-run command • Runs basic script producing XML encoded status information • Load average and timestamp information retrieved • Current status and archived load information is plotted... • To be done... • JAVA CoG kit being investigated (more robust) • Simple monitoring system to verify test-bed timestamps (in case not everyone is using NTP) • Integrate with the Grid Monitoring Architecture • Incorporate current network bandwidth measurements into graphical system • Automatic notification system e-Science Presentation

More Related