1 / 43

Paul Avery University of Florida avery@phys.ufl

Explore the landmark GridChem workshop connecting national cyberinfrastructure, scale of OSG resources, global science collaborations, LHC advancements, and lessons learned from Grid3 operations.

wwells
Download Presentation

Paul Avery University of Florida avery@phys.ufl

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Open Science Grid Linking Universities and Laboratories in National Cyberinfrastructure GridChem WorkshopUniversity of Texas, AustinAustin, TexasMarch 9, 2006 Paul Avery University of Florida avery@phys.ufl.edu Paul Avery

  2. OSG Roots: “Trillium Consortium” • Trillium = PPDG + GriPhyN + iVDGL • PPDG: $12M (DOE) (1999 – 2006) • GriPhyN: $12M (NSF) (2000 – 2005) • iVDGL: $14M (NSF) (2001 – 2006) • Large science experiments (HEP/LHC, LIGO, SDSS) • Total ~150 people with many overlaps between projects • Universities, labs, foreign partners • Historically, a strong driver for funding agency collaboration • Inter-agency (NSF – DOE) + intra-agency (Directorate – Directorate) • Coordination vital for meeting broad goals • CS research, developing/supporting Virtual Data Toolkit (VDT) • Multiple Grid deployments, using VDT-based middleware • Deployment of Grid3, a general purpose, national Grid • Unified entity when collaborating internationally Paul Avery

  3. Scale of OSG Resources & Services Set byLarge Hadron Collider (LHC) Expts. • 27 km Tunnel in Switzerland & France CMS TOTEM LHC @ CERN ALICE LHCb Search for • Origin of Mass • New fundamental forces • Supersymmetry • Other new particles • 2007 – ? ATLAS Paul Avery

  4. LHC: Beyond Moore’s Law LHC CPU Requirements Moore’s Law (2000) Paul Avery

  5. LHC: Petascale Global Science • Complexity: Millions of individual detector channels • Scale: PetaOps (CPU), 100s of Petabytes (Data) • Distribution: Global distribution of people & resources BaBar/D0 Example - 2004 700+ Physicists 100+ Institutes 35+ Countries CMS Example- 2007 5000+ Physicists 250+ Institutes 60+ Countries Paul Avery

  6. Korea Russia UK USA U Florida Caltech UCSD Iowa FIU Maryland LHC Global Data Grid (2007+) • 5000 physicists, 60 countries • 10s of Petabytes/yr by 2008 • 1000 Petabytes in < 10 yrs? CMS Experiment Online System CERN Computer Center 150 - 1500 MB/s Tier 0 10-40 Gb/s Tier 1 >10 Gb/s Tier 2 2.5-10 Gb/s Tier 3 Tier 4 Physics caches PCs Paul Avery

  7. Cardiff AEI/Golm • LIGO Grid • LIGO Grid: 6 US sites + 3 EU sites (UK & Germany) Birmingham• * LHO, LLO: LIGO observatory sites * LSC: LIGO Scientific Collaboration Paul Avery

  8. Common Middleware: Virtual Data Toolkit VDT NMI Test Sources (CVS) Build Binaries Build & Test Condor pool 22+ Op. Systems Pacman cache Package Patching RPMs Build Binaries GPT src bundles Build Binaries Test Many Contributors VDT: Package, test, deploy, support, upgrade, troubleshoot Paul Avery

  9. VDT Growth Over 4 Years (1.3.10 now) www.griphyn.org/vdt/ Paul Avery

  10. Grid3: A National Grid Infrastructure • October 2003 – July 2005 • 32 sites, 3,500 CPUs: Universities + 4 national labs • Sites in US, Korea, Brazil, Taiwan • Applications in HEP, LIGO, SDSS, Genomics, fMRI, CS Brazil www.ivdgl.org/grid3 Paul Avery

  11. Grid3 Lessons Learned • How to operate a Grid as a facility • Security, services, error recovery, procedures, docs, organization • Delegation of responsibilities (Project, VO, service, site, …) • Crucial role of Grid Operations Center (GOC) • How to support people  people relations • Face-face meetings, phone cons, 1-1 interactions, mail lists, etc. • How to test and validate Grid tools and applications • Vital role of testbeds • How to scale algorithms, software, process • Some successes, but “interesting” failure modes still occur • How to apply distributed cyberinfrastructure • Successful production runs for several applications Paul Avery

  12. Open Science Grid: July 20, 2005 • VO based: Partnership of many organizations • Production Grid: 50+ sites, 19,000 CPUs “present”(available but not at one time) • Sites in US, Korea, Brazil, Taiwan • Integration Grid: ~15 sites Taiwan, S.Korea Sao Paolo Paul Avery

  13. OSG Operations Snapshot (24 Hr) http://grid02.uits.indiana.edu:8080/reports/daily/ Paul Avery

  14. OSG Operations Snapshot (30 Day) November 7: 30 days Paul Avery

  15. Paul Avery

  16. Registered VOs (incomplete) Paul Avery

  17. Creating & Registering a VO With OSG • To form a Virtual Organization (VO) that participates in the Open Science Grid one needs the following: • a Charter statement describing the purpose of the VO. This should be short yet concise enough to scope intended usage of OSG resources. • at least one VO participating Organization that is a member or partner with the Open Science Grid Consortium. • a VO Membership Service which meets the requirements of an OSG Release. This means being able to provide a full list of members' DNs to edg-mkgridmap. The currently recommended way to do this is to deploy the VDT VOMS from the OSG software package. • a support organization (called a Support Center in OSG parlance) that will support the VO in OSG Operations. The Support Center should provide at least the following: • a written description of the registration process, • instructions for the members of the VO on how to complete the VO registration process, • instructions for the members of the VO on how to report problems and/or obtain help. • completion of the registration form located here using these instructions Paul Avery

  18. Vo Support Matrix Green: DNs are mapped to this VO and compute element [clickable]    Yellow: No DNs are supported under this VO and compute element. Black: No information Paul Avery

  19. OSG Integration Testbed • Test, validate new middleware & services • Test, validate new applications • Meets weekly (highly dynamic membership) Taiwan Brazil Korea Paul Avery

  20. Program Oversight OSG Organization Contributors UniversitiesLaboratories Sites Service Providers VOs Researchers Computer Science Grid Projects … Partners Campus Grids EGEE TeraGrid OSG Consortium OSG Council Council Chair Scientific Advisory Group OSG Users Group Applications Coordinator Executive Board Executive Director Applications Coordinator Education Coordinator Facility Coordinator Resources Manager Finance Board Resources Manager OSG FacilityFacility CoordinatorEngagement CoordinatorOperations CoordinatorMiddleware CoordinatorSecurity Officer Council Chair Engagement Coordinator Middleware Coordinator Operations Coordinator Security Officer Liaison to EU Grid Projects Liaison to TeraGrid/USGrid Projects Project Technical Managers Contributor Technical Managers Executive Team Line reporting Advisory Projects Project Mangers Resource Managers … Contributing & Interfacing, (MOUs, etc) Projects Project Managers Resource Managers … Paul Avery

  21. Process forDeploying NewOSG Service Paul Avery

  22. OSG Participating Disciplines Paul Avery

  23. OSG Grid Partners Paul Avery

  24. Example of Partnership:WLCG and EGEE Paul Avery

  25. OSG Activities Paul Avery

  26. Networks Paul Avery

  27. Evolving Science Requirements for Networks (DOE High Performance Network Workshop) See http://www.doecollaboratory.org/meetings/hpnpw/ Paul Avery

  28. UltraLight Integrating Advanced Networking in Applications http://www.ultralight.org 10 Gb/s+ network • Caltech, UF, FIU, UM, MIT • SLAC, FNAL • Int’l partners • Level(3), Cisco, NLR Paul Avery

  29. Training Outreach Communications Paul Avery

  30. Grid Summer Schools • June 2004: First US Grid Tutorial (South Padre Island, Tx) • 36 students, diverse origins and types • July 2005: Second Grid Tutorial (South Padre Island, Tx) • 42 students, simpler physical setup (laptops) • June 23-27: Third Grid Tutorial (South Padre Island, Tx) • Reaching a wider audience • Lectures, exercises, video, on web • Students, postdocs, scientists • Coordination of training activities • More tutorials, 3-4/year • Agency specific tutorials Paul Avery

  31. Grid Technology Cookbook A guide to building and using grid resources Current Timetable (2005 – 06) Acknowledgements Preface Introduction What Grids Can Do For You Grid Case Studies Technology For Grids Standards & Emerging Technologies Programming Concepts & Challenges Building Your Own Grid Installation Procedure Examples Typical Usage Examples Practical Tips Glossary Appendices • Outline Development, Vetting September-October • Assemble Writing Teams October-December • Develop Web Structure November-December • Writing Process Underway November-March • Material Edited and Entered December-April • Review of First Draft May • Edits to First Draft Entered Early June • Review of Final Draft Late June • Release of Version 1 July 2006 Paul Avery

  32. QuarkNet/GriPhyN e-Lab Project http://quarknet.uchicago.edu/elab/cosmic/home.jsp Paul Avery

  33. CHEPREO: Center for High Energy Physics Research and Educational OutreachFlorida International University www.chepreo.org • Physics Learning Center • CMS Research • Cyberinfrastructure • WHREN network (S. America) • Funded September 2003 • $MPS, CISE, EHR, INT

  34. Grids and the Digital Divide Background • World Summit on Information Society • HEP Standing Committee on Inter-regional Connectivity (SCIC) Themes • Global collaborations, Grids and addressing the Digital Divide • Focus on poorly connected regions • Brazil (2004), Korea (2005) Paul Avery

  35. Science Grid Communications Broad set of activities • (Katie Yurkewicz) • News releases, PR, etc. • Science Grid This Week • OSG Monthly Newsletter www.interactions.org/sgtw Paul Avery

  36. OSG Newsletter Monthly newsletter • (Katie Yurkewicz) • 4 issues now www.opensciencegrid.org/osgnews Paul Avery

  37. 2000 2001 2002 2003 2004 2005 2006 2007 Grid Timeline First US-LHCGrid Testbeds Grid Communications Grid Summer School 06 Grid3 operations GriPhyN, $12M UltraLight, $2M Start of LHC LIGO Grid DISUN, $10M iVDGL, $14M CHEPREO, $4M OSG operations VDT 1.0 Grid Summer Schools 04, 05 OSG funded? NSF, SciDAC PPDG, $9.5M Digital Divide Workshops Paul Avery

  38. OSG Consortium Meetings • July 20, 2006: University of Wisconsin, Milwaukee • Kickoff meeting, ~100 attendees • Focus on getting off the ground with running jobs • January 23, 2006: University of Florida (Gainesville) • ~110 people • Partnerships, organization, funding, operations, software infrastructure • August 21-24, 2006: University of Washington (Seattle) • January, 2007: TACC Taiwan, S.Korea Sao Paolo Paul Avery

  39. Jan. 23-25 OSG Meeting Paul Avery

  40. END Paul Avery

  41. Open Science Grid www.opensciencegrid.org Grid3 www.ivdgl.org/grid3 Virtual Data Toolkit www.griphyn.org/vdt GriPhyN www.griphyn.org iVDGL www.ivdgl.org PPDG www.ppdg.net CHEPREO www.chepreo.org UltraLight www.ultralight.org Globus www.globus.org Condor www.cs.wisc.edu/condor WLCG www.cern.ch/lcg EGEE www.eu-egee.org Grid Project References Paul Avery

  42. Sloan Data Galaxy cluster size distribution Sloan Digital Sky Survey (SDSS)Using Virtual Data in GriPhyN Paul Avery

  43. Cardiff AEI/Golm • The LIGO Scientific Collaboration (LSC)and the LIGO Grid • LIGO Grid: 6 US sites + 3 EU sites (Cardiff/UK, AEI/Germany) Birmingham• * LHO, LLO: LIGO observatory sites * LSC: LIGO Scientific Collaboration Paul Avery

More Related