1 / 59

Campus Bridging in XSEDE, OSG, and Beyond

This presentation provides an overview of campus bridging, including the XSEDE Campus Bridging Program, the Global Federated File System, and initiatives such as RedCloud and POD at IU. It also covers the role of the NSF Advisory Committee for CyberInfrastructure and community input in shaping campus bridging efforts.

graziano
Download Presentation

Campus Bridging in XSEDE, OSG, and Beyond

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Campus Bridging in XSEDE, OSG, and Beyond Dan Fraser, OSG Jim Ferguson, NICS Andrew Grimshaw, UVA Rich Knepper, IU David Lifka, Cornell Internet2 Spring Members’ Meeting, Arlington, VA

  2. Overview • Campus Bridging introduction • Open Science Grid and software deployment • XSEDE Campus Bridging Program • Global Federated File System • Campus Initiatives • RedCloud • POD at IU

  3. Campus Bridging: Formation • NSF Advisory Committee for CyberInfrastructure • Community input at multiple workshops • Surveys • ACCI Campus Bridging Task Force Report • http://pti.iu.edu/campusbridging

  4. Campus Bridging: Concepts • Making it easier for users to transition from their laptop to large-scale resources • Gathering best practices to deploy resources in a way that makes them familiar to users • Providing training and documentation that covers research computation at multiple scales

  5. OSG Campus Bridging (Campus High Throughput Computing Infrastructures) Dan Fraser OSG Production Coordinator Campus Infrastructure Lead Internet2 Spring Members Meeting Arlington, VA Sept 24, 2012

  6. The Open Science Grid The Open Science Grid (OSG) has focused on campuses from its inception. All OSG computing power comes from campuses. OSG has a footprint on over 100 campuses in the US and abroad. http://display.grid.iu.edu

  7. OSG Campus Bridging Focus One step at a time • Focus on the Researcher (…or Artist)

  8. Simple interfaces are good

  9. Engaging with the Campus Campus Bridging = Cultural Bridging • Campuses each have their own “culture” • Terminology • Access patterns • Security • Operational styles • Processes (autN, autZ, monitoring, accounting, data management, …) • The most fundamental issues are not technological in nature

  10. Campus Bridging Direction Submit Locally, Run Globally • Help the researcher use local resources • Run on a local cluster (on campus) • Run on several local clusters • Use/share resources with a collaborator on another campus • Access to the national cyberinfrastructure • OSG (and also XSEDE) resources • (BTW, OSG is also an XSEDE service provider)

  11. OSG Campus Bridging Today Campus LSF PBS Local User Credential Condor Local Cluster Submit Host(Bosco) OSG Cloud External Campus (could also submit to XSEDE)

  12. Summary • OSG is focused on the researcher/artist • Campus bridging = cultural bridging • A single submit model (Bosco) can be useful • OSG is exploring how best to collaborate with XSEDE on campus bridging

  13. Introduction to XSEDE and its Campus Bridging Program Jim Ferguson, XSEDE TEOS team Education, Outreach & Training Director, NICS jwf@utk.edu

  14. Acknowledgements Craig Stewart, Rich Knepper, and Therese Miller of Indiana University, and others on the XSEDE campus bridging team. John Towns, NCSA, XSEDE PI

  15. XD Solicitation/XD Program • eXtreme Digital Resources for Science and Engineering (NSF 08-571) • High-Performance Computing and Storage Services • aka Track 2 awardees • High-Performance Remote Visualization and Data Analysis Services • 2 awards; 5 years; $3M/year • proposals due November 4, 2008 • Integrating Services (5 years, $26M/year) • Coordination and Management Service (CMS) • 5 years; $12M/year • Technology Audit and Insertion Service (TAIS) • 5 years; $3M/year • Advanced User Support Service (AUSS) • 5 years; $8M/year • Training, Education and Outreach Service (TEOS) • 5 years, $3M/year

  16. XSEDE Vision The eXtreme Science and Engineering Discovery Environment (XSEDE): enhances the productivity of scientists and engineers by providing them with new and innovative capabilities and thus facilitates scientific discovery while enabling transformational science/engineering and innovative educational programs

  17. Science requires diverse digital capabilities • XSEDE is a comprehensive, expertly managed and evolving set of advanced heterogeneous high-end digital services, integrated into a general-purpose infrastructure. • XSEDE is about increased user productivity • increased productivity leads to more science • increased productivity is sometimes the difference between a feasible project and an impractical one

  18. XSEDE’s Distinguishing Characteristics - Governance • World-class leadership • partnership will be led by NCSA, NICS, PSC, TACC and SDSC • CI centers with deep experience • partners who strongly complement these CI centers with expertise in science, engineering, technology and education • Balanced governance model • strong central management provides rapid response to issues and opportunities • delegation and decentralization of decision-making authority • openness to genuine stakeholder participation • stakeholder engagement, advisory committees • improved professional project management practices • formal risk management and change control

  19. What is campus bridging? • Term originated by Ed Seidel as he charged six task forces of the NSF Advisory Committee for Cyberinfrastructure. Considerable info and final ACCI Task force report, online at pti.iu.edu/campusbridging • The Taskforce definition: Campus bridging is the seamlessly integrated use of cyberinfrastructure operated by a scientist or engineer with other cyberinfrastructure on the scientist’s campus, at other campuses, and at the regional, national, and international levels as if they were proximate to the scientist . . . • Catchy name, great ideas … interest > our ability to implement yet • Vision: • Help XSEDE create the software, tools and training that will allow excellent interoperation between XSEDE infrastructure researchers local (campus) cyberinfrastructure; • Enable excellent usability from the researcher’s standpoint for a variety of modalities and types of computing: HPC, HTC, and data intensive computing • Promote better use of local, regional and national CI resources

  20. Campus Bridging Use Cases • InCommon Authentication • Economies of scale in training and usability • Long term remote interactive graphic session • Use of data resources from campus on XSEDE, or from XSEDE at a campus • Support for distributed workflows spanning XSEDE and campus-based data, computational, and/or visualization resources • Shared use of computational facilities mediated or facilitated by XSEDE • Access to “____ as a Service” mediated or facilitated by XSEDE 20

  21. Year 2 Strategic Plan Leveraging the XSEDE process of implementing new services via the systems engineering process (Architecture & Design => Software Development and Integration => Operations), begin deploying some new services that deliver campus bridging services Communicate effectively via Campus Champions, advocates for XSEDE now located at over 100 institutions. Develop relationship and terminology with OSG, as they have been bridging between institutions for several years.

  22. Year 2 Strategic Plan • Complete planned pilot projects • GFFS Pilot Program • CUNY – PI: Paul Muzio • KU – PI: Thorbjorn Axelsson • Miami – PI: Joel Zysman • TAMU – PI: Guy Almes • Begin delivering selected campus-bridging related tools • GFFS • Documentation • ROCKS Rolls • Communicate “what is campus bridging”

  23. Example: More consistency in CI setups => economies of scale for all • In reality, the four cluster admins depicted here being in agreement are all right. • Experienced cluster admins all learned how to use what they learn when the tools were still developing, so the tool each sysadmin knows the best is the tool that lets that sysadmin do their work the best • The only way to develop consistency is to provide installers that will make their work easier • The XSEDE architecture group is developing installers for file management tools • *A la Steven Colbert, the “4 out of 5…” comment is not intended to be a factual statement 23

  24. Your Comments, Please! Do we have the right direction? What is Campus Bridging to You?

  25. Thank You! jwf@utk.edu

  26. Global Federated File SystemGFFS Andrew Grimshaw

  27. Basic idea and canonical use cases • Accessing the GFFS • Attaching (provisioning) data to the GFFS • Deployment

  28. PDB Cluster 1 APP 1 NCBI APP 2 Cluster 2 EMBL APP N Cluster N SEQ_1 Applications Processing Data Basic idea: Map resources into a global directory structure Map global directory structure into local file system SEQ_1 SEQ_3 SEQ_2 APP 1 APP 2 Biology Biochemistry Partner Institution Partner Institution Research Institution

  29. Canonical use cases Definitions • Resource is {compute | job | data | identity | …} • Access means create, read, update, delete • Access center resource from campus • Access campus resource from center • Access campus resource from another campus • Sharing file system or instrument data • Sharing clusters

  30. Basic idea and canonical use cases • Accessing the GFFS • Attaching (provisioning) data to the GFFS • Deployment

  31. Accessing the GFFS • Via a file system mount • Global directory structure mapped directly into the local operating system via FUSE mount • XSEDE resources regardless of location can be accessed via the file system • Files and directories can be accessed by programs and shell scripts as if they were local files • Jobs can be started by copying job descriptions into directories • One can see the jobs running or queued by doing an “ls”. • One can “cd” into a running job and access the working directory where the job is running directly

  32. E.g., Access a job’s running directory

  33. Accessing the GFFS • Via a command line tools, e.g., • cp local:fred.txt /home/grimshaw/fred.txt • rm /home/grimshaw/fred.txt

  34. GUI Grid Client • Typical folder based tool • Tools to define, run, manage jobs • Tools to manage “grid” queues • Tools to “export” data • Grid shell • Shell has tab completion, history, help, scripting, etc.

  35. GUI Grid Client: View Access Control • To view access control information: Browse to and highlight resource, then select Security tab

  36. Basic idea and canonical use cases • Accessing the GFFS • Attaching (provisioning) data to the GFFS • Deployment

  37. Exporting (mapping) data into the Grid Data clients Data clients • Links directories and files from source location to GFFS directory and user-specified name • Presents unified view of the data across platforms, locations, domains, etc. • Sarah controls authorization policy. Sarah’s department file server Sarah’s instrument in the lab Sarah’s TACC workspace TACC Windows Linux

  38. Exporting/sharing data • Can also export • Windows shares • Directory structures via ssh (slow – like sshFX) • User selects • Server that will perform the export • Directory path on that server • Path in GFFS to link it to

  39. Basic idea and canonical use cases • Accessing the GFFS • Attaching (provisioning) data to the GFFS • Deployment

  40. Deployment • Sites that wish to export or share resources must run a Genesis II or UNICORE 6 container • There will be an installer for the GFFS package for SPs, and a “Campus Bridging” package • There is an installer for client side access • There are training materials • Used at TG 11 • In the process of being turned into videos

  41. On-Demand Research Computing - Infrastructure as a Service - - Software as a Service - www.cac.cornell.edu/redcloud

  42. Infrastructure as a Service (IaaS) Cloud Red Cloud provides on-demand: • Computing Cycles: Virtual Servers in Cloud “Instances” • Storage: Virtual Disks in Elastic Block Storage (“EBS”) Volumes Red Cloud Virtual Servers “Cloud Instances” Cloud Management Virtual Disks “Elastic Block Storage (EBS)” Virtual Server Users www.cac.cornell.edu/redcloud

  43. GPU Chassis Dell C410x Compute Nodes Dell C6100 NVIDIA Tesla M2070s with MATLAB Software as a Service (SaaS) Cloud Head Node DDN Storage Web Server Network Interconnect SQL Server GridFTP Server MyProxy Server www.cac.cornell.edu/redcloud

  44. Motivation • Research computing means many different things… • Scientific workflows have different requirements at each step • Cloud is only part of the solution • Connecting to and from other CI resources is important • Nobody likes a bad surprise • Transparency, no hidden costs • Need a way to bound financial risk • Economies of scale • Sharing hardware and software where it makes sense • Pay for what you need, when you need it • Customized environments for various disciplines • Collaboration tools • Data storage & analysis tools • Flexibility to support different computing models (e.g. Hadoop) www.cac.cornell.edu/redcloud

  45. Provides Predictable, Reproducible, Reliable Performance We publish hardware specifications (CPU, RAM, network) and do not oversubscribe. Convenient Need system up and running yesterday. Need a big fast machine for only a few months, weeks or days. Need a small server to run continuously. No Hidden CostsNo cost for network traffic in or out of the cloud. Fast Access to Your DataFast data transfers via 10Gb Ethernet in or out of the cloud at no additional charge. Globus Online access Economies of scale IaaS: Infrastructure SaaS: Software Expert Help System, application, and programming consulting are available. Easy Budgeting with SubscriptionsNo billing surprises! IaaS is Amazon API Compatible Migrate when your requirements outgrow Red Cloud. www.cac.cornell.edu/redcloud

  46. Some Use Cases to Consider • Support for Scientific Workflows • Pre & post-processing of data and results • Data analysis • Globus Online for fast reliable data transfer • https://www.globusonline.org/ • Collaboration • Wiki hosting • Customized data analysis & computational environments • Web Portals • Science Gateways • Domain Specific Portals • Hub Zero • http://hubzero.org/pressroom • http://nanohub.org • Event-Driven Science • https://opensource.ncsa.illinois.edu/confluence/display/SGST/Semantic+Geostreaming+Toolkit • Education, Outreach & Training • Pre-configured systems & software tools providing consistent training platform • Common laboratory computing environment • Bursting • Additional software and hardware on demand www.cac.cornell.edu/redcloud

  47. Subscription-based Recovery Model with MATLAB Cornell University $500/core year* Other Academic $750/core yearInstitutions Cornell University $750/core year Other Academic $1200/core yearInstitutions *A core year is equal to 8585 hours Each subscription account includes 50GB of storage www.cac.cornell.edu/redcloud

  48. What if ??? Consulting Additional Storage Cornell Users $59.90/hour$0.91/GB/year Other Academic $85.47/hour$1.45/GB/year Institutions www.cac.cornell.edu/redcloud

  49. Penguin Computing / IU PartnershipHPC “cluster as a service” andCloud Services Internet2 Spring Members’ Meeting 2012 Rich Knepper (rich@iu.edu) Manager, Campus Bridging Indiana University

More Related