400 likes | 413 Views
Explore the hierarchical growth and collaborative advancements in the Abilene Network, focusing on the next generation of optical networking and international connectivity initiatives. Learn about the network infrastructure, partnerships, and future endeavors shaping the Internet2 backbone. Discover how technological innovations are enabling new services and applications, with a key emphasis on collaborative security approaches.
E N D
Internet2 Abilene Network and Next Generation Optical Networking Steve Corbató Director, Backbone Network Infrastructure Access NovaForum May 28, 2002
This presentation • Abilene Network today • Optical networking evolution • Next generation of Abilene • Future national optical initiatives
Networking hierarchy • Internet2 networking is a fundamentally hierarchical and collaborative activity • International networking • Ad hoc Global Terabit Research Network (GTRN) • National backbones • Regional networks • GigaPoPs advanced regional networks • Campus networks • Much activity now at the metropolitan and regional scales
Abilene focus • Goals • Enabling innovative applications and advanced services not possible over the commercial Internet • Backbone & regional infrastructure provides a vital substrate for the continuing culture of Internet advancement in the university/corporate research sector • Advanced service efforts • Multicast • IPv6 • QoS • Measurement • an open, collaborative approach • Security
Partnership approach • The Abilene Network is a UCAID project done in partnership with • Cisco Systems (routers, switches, and access) • Juniper Networks (routers) • Nortel Networks (SONET kit) • Qwest Communications (SONET & DWDM circuits) • Indiana University (network operations center) • Internet2 Test & Evaluation Centers (ITECs) • North Carolina and Ohio
Abilene – May, 2002 • IP-over-SONET backbone (OC-48c, 2.5 Gbps) 53 direct connections • 4 OC-48c connections • 1 Gigabit Ethernet trial • 23 will connect via at least OC-12c (622 Mbps) by 1Q02 • Number of ATM connections decreasing • 215 participants – research universities & labs • All 50 states, District of Columbia, & Puerto Rico • 15 regional GigaPoPs support ~70% of participants • Expanded access • 50 sponsored participants • New: Smithsonian Institution, Arecibo Radio Telescope • 23 state education networks (SEGPs)
Abilene international connectivity • Transoceanic R&E bandwidths growing! • GÉANT – 5 Gbps between Europe and New York City now • Key international exchange points facilitated by Internet2 membership and the U.S. scientific community • STARTAP & STAR LIGHT – Chicago (GigE) • AMPATH – Miami (OC-3c OC-12c) • Pacific Wave – Seattle (GigE) • MAN LAN - New York City (GigE/10GigE EP soon) • CA*NET3/4: Seattle, Chicago, and New York • CUDI: CENIC and Univ. of Texas at El Paso • International transit service • Collaboration with CA*NET3 and STARTAP
Abilene international connectivity model • Abilene is a GTRN partner • Already peering with GTRN routers in New York City and Seattle • Peering at major int’l EPs in U.S. encouraged • Chicago: Star Light (migration from STAR TAP) • Seattle: Pacific Wave • Miami: AMPATH • New York City: Manhattan Landing (MAN LAN) in progress • Los Angeles (soon?) • Direct BGP peering preferred • via Layer-2 EP media or direct connection • ATM support generally ends by Sept 2003 • No new ATM peers
09 March 2002 Sacramento Washington Los Angeles Abilene International Peering STAR TAP/Star Light APAN/TransPAC, Ca*net3, CERN, CERnet, FASTnet, GEMnet, IUCC, KOREN/KREONET2, NORDUnet, RNP2, SURFnet, SingAREN, TAnet2 Pacific Wave AARNET, APAN/TransPAC, CA*net3, TANET2 NYCM BELNET, CA*net3, GEANT*, HEANET, JANET, NORDUnet SNVA GEMNET, SINET, SingAREN, WIDE LOSA UNINET OC3-OC12 San Diego (CALREN2) CUDI AMPATH REUNA, RNP2 RETINA, ANSP, (CRNet) El Paso (UACJ-UT El Paso) CUDI * ARNES, CARNET, CESnet, DFN, GRNET, RENATER, RESTENA, SWITCH, HUNGARNET, GARR-B, POL-34, RCST, RedIRIS
Packetized raw High Definition Television (HDTV) • Raw HDTV/IP – single UDP flow of 1.5 Gbps • Project of USC/ISIe, Tektronix, & U. of Wash (DARPA) • 6 Jan 2002: Seattle to Washington DC via Abilene • Single flow utilized 60% of backbone bandwidth • 18 hours: no packets lost, 15 resequencing episodes • End-to-end network performance (includes P/NW & MAX GigaPoPs) • Loss: <0.8 ppb (90% c.l.) • Reordering: 5 ppb • Transcontinental 1-Gbps TCP requires loss of • <30 ppb (1.5 KB frames) • <1 ppm (9KB jumbo)
End-to-End Performance:‘High bandwidth is not enough’ • Bulk TCP flows (> 10 Mbytes transfer) • Current median flow rate over Abilene: 1.9 Mbps
True End-to-End Performance requires a system approach • User perception • Application • Operating system • Host IP stack • Host network card • Local Area Network • Campus backbone network • Campus link to regional network/GigaPoP • GigaPoP link to Internet2 national backbones • Internationalconnections EYEBALL APPLICATION STACK JACK NETWORK . . . . . . . . . . . .
Jumbo frames supported • Default Abilene backbone MTU has been increased from 4.5 to 9 kB • We now can support 9 kB MTUs on a per connector basis • Motivation: support for HPC computing and large TCP flows
Abilene traffic characterization information • Weekly detailed reports • http://netflow.internet2.edu/weekly/ • General analysis • http://www.itec.oar.net/abilene-netflow/
Optical networking technology drivers • Aggressive period of fiber construction on the national & metro scales in U.S. • Now rapid industry contraction and capital crisis • Many university campuses and regional GigaPoPs already use dark fiber • Dense Wave Division Multiplexing (DWDM) • Allows the provisioning of multiple channels (’s) over distinct wavelengths on the same fiber pair • Fiber pair can carry 160 channels (1.6 Tbps!) • Optical transport is the current focus • Optical switching is still in the realm of experimental networks, but may be nearing practical application
DWDM technology primer • DWDM fundamentally is an analog optical technology • Combines multiple channels (2-160+ in number) over the same fiber pair • Uses slightly displaced wavelengths (’s) of light • Generally supports 2.5 or 10 Gbps channels • Physical obstacles to long-distance transmission of light • Attenuation • Solved by amplification (OO) • Wavelength dispersion • Requires periodic signal regeneration – an electronic process (OEO)
DWDM system components • Base fiber pair (+ right of way & conduit) • Multiplexing/demultiplexing terminals • OEO equipment at each end of light path • Output: SONET or Ethernet (10G/1G) framing • Amplifiers • All optical (OO) • ~100 km spacing • Regeneration • Electrical (OEO) process – costly (~50% of capital) • ~500 km spacing (with Long Haul - LH - DWDM) • New technologies (ELH/ULH) can extend this distance • Remote huts, operations & maintenance
Telephony’s recent past (from an IP perspective in the U.S.)
National optical networking options • 1 – Provision incremental wavelengths • Obtain 10-Gbps ’s as with SONET • Exploit smaller incremental cost of additional ’s • 1st costs ~10x than subsequent ’s • 2 – Build dim fiber facility • Partner with a facilities-based provider • Acquire 1-2 fiber pairs on a national scale • Outsource operation of inter-city transmission equipment • Needs lower-cost optical transmission equipment • The classic ‘buy vs. build’ decision in Information Technology
Future of Abilene • Original UCAID/Qwest agreement amended on October 1, 2001 • Extension of MoU for another 5 years – until October, 2006 • Originally expired March, 2003 • Upgrade of Abilene backbone to optical transport capability - ’s (unprotected) • x4 increase in the core backbone bandwidth • OC-48c SONET (2.5 Gbps) to 10-Gbps DWDM
Key aspects of next generation Abilene backbone - I • Native IPv6 • Motivations • Resolving IPv4 address exhaustion issues • Preservation of the original End-to-End Architecture model • p2p collaboration tools, reverse trend to CO-centrism • International collaboration • Router and host OS capabilities • Run natively - concurrent with IPv4 • Replicate multicast deployment strategy • Close collaboration with Internet2 IPv6 Working Group on regional and campus v6 rollout • Addressing architecture
Key aspects of next generation Abilene backbone - II • Network resiliency • Abilene ’s will not be ring protected like SONET • Increasing use of videoconferencing/VoIP impose tighter restoration requirements (<100 ms) • Options: • MPLS/TE fast reroute (initially) • IP-based IGP fast convergence (preferable)
Key aspects of next generation Abilene backbone - III • New & differentiated measurement capabilities • Significant factor in NGA rack design • 4 dedicated servers at each nodes • Additional provisions for future servers • Local data collection to capture data at times of network instability • Enhance active probing • Now: latency & jitter, loss, reachability (Surveyor) • Regular TCP/UDP throughput tests – ~1 Gbps • Separate server for E2E performance beacon • Enhance passive measurement • Now: SNMP (NOC) & traffic matrix/type (Netflow) • Routing (BGP & IGP) • Optical splitter taps on backbone links at select location(s)
Abilene Observatories • Currently a program outline for better support of computer science research • Influenced by discussions with NRLC members • 1) Improved & accessible data archive • Need coherent database design • Unify & correlate 4 separate data types • SNMP, active measurement data, routing, Netflow • 2) Provision for direct network measurement and experimentation • Resources reserved for two additional servers • Power (DC), rack space (2RU), router uplink ports (GigE) • Need process for identifying meritorious projects • Need ‘rules of engagement’ (technical & policy)
Next generation router selection • Extensive router specification and test plan developed • Team effort: UCAID staff, NOC, NC and Ohio ITECs • Discussions with four router vendors • Tests focused on next gen advanced services • High performance TCP/IP throughput • High performance multicast • IPv6 functionality & throughput • Classification for QoS and measurement • 3 router platforms tested & commercial ISPs referenced • New Juniper T640 platform selected
Abilene program changes • 10-Gbps (OC-192c POS) connections • backhaul available wherever needed & possible • Only required now for 1 of 4 OC-48c connections • 3-year connectivity commitment required • Gigabit and 10-Gigabit Ethernet • Available when connector has dark fiber access into Abilene router node • Backhaul not available • ATM connection & peer support • TAC recommended ending ATM support by fall 2003 • Two major ATM-based GigaPoPs have migrated • 2 of 3 NGIXes still are ATM-based • NGIX-Chicago @ STAR LIGHT is now GigE • Urging phased migration for connectors & peers
Deployment timing • Ongoing – Backbone router procurement • Detailed deployment planning • July – Rack assembly (Indiana Univ.) • Aug/Sep – New rack deployment at all 11 nodes • Fall – First Wave ’s commissioned • Fall meeting demonstration events • iGRID 2002 (Amsterdam) – late Sep. • Internet2 Fall Member Meeting (Los Angeles) – late Oct. • SC2002 (Baltimore) – mid Nov. • Remaining ’s commissioned in 2003
Two leading national initiatives in the U.S. • Next Generation Abilene • Advanced Internet backbone • connects entire campus networks of the research universities • 10 Gbps nationally • TeraGrid • Distributed computing (Grid) backplane • connects high performance computing (HPC) machine rooms • Illinois: NCSA, Argonne • California: SDSC, Caltech • 4x10 Gbps: Chicago Los Angeles • Ongoing collaboration between both projects
TeraGrid: A National Infrastructure For more information: www.teragrid.org
OC-12 vBNS Abilene MREN OC-12 OC-3 = 32x 1GbE 32 quad-processor McKinley Servers (128p @ 4GF, 8GB memory/server) TeraGrid Architecture – 13.6 TF (Source: C. Catlett, ANL) 574p IA-32 Chiba City 32 256p HP X-Class 32 Argonne 64 Nodes 1 TF 0.25 TB Memory 25 TB disk 32 32 Caltech 32 Nodes 0.5 TF 0.4 TB Memory 86 TB disk 128p Origin 24 32 128p HP V2500 32 HR Display & VR Facilities 24 8 8 5 5 92p IA-32 HPSS 24 HPSS OC-12 ESnet HSCC MREN/Abilene Starlight Extreme Black Diamond 4 OC-48 Calren OC-48 OC-12 NTON GbE OC-12 ATM Juniper M160 NCSA 500 Nodes 8 TF, 4 TB Memory 240 TB disk SDSC 256 Nodes 4.1 TF, 2 TB Memory 225 TB disk Juniper M40 Juniper M40 OC-12 vBNS Abilene Calren ESnet OC-12 2 2 OC-12 OC-3 Myrinet Clos Spine 8 4 UniTree 8 HPSS 2 Sun Starcat Myrinet Clos Spine 4 1024p IA-32 320p IA-64 1176p IBM SP Blue Horizon 16 14 = 64x Myrinet 4 = 32x Myrinet 1500p Origin Sun E10K = 32x FibreChannel = 8x FibreChannel 10 GbE 32 quad-processor McKinley Servers (128p @ 4GF, 12GB memory/server) Fibre Channel Switch 16 quad-processor McKinley Servers (64p @ 4GF, 8GB memory/server) IA-32 nodes Router or Switch/Router
Optical networking scaling factors • 2 TeraGrid routing nodes • 11 Next Generation Abilene routers • 53 Abilene connectors • 215 Abilene participants (univs & labs) • But… • 30-60 DWDM access nodes in leading viable carriers’ U.S. networks
Regional optical fanout • Next generation architecture: Regional & state based optical networking projects are critical • Three-level hierarchy • backbone, GigaPoPs/ARNs, campuses • Leading examples • CENIC ONI (California), I-WIRE (Illinois), Indiana (I-LIGHT) • Collaboration with the Quilt GigaPoPs • Regional Optical Networking project • U.S. carrier DWDM access is now not nearly as widespread as with SONET circa 1998 • 30-60 cities for DWDM • ~120 cities for SONET
Conclusions • Backbone upgrade project underway • Partnership with Qwest extended thru 2006 • Juniper T640 routers selected for backbone • 10-Gbps backbone deployment starts this fall • Incremental, non-disruptive transition • Advanced service foci • Native, high-performance IPv6 • Enhanced, differentiated measurement • Network resiliency • NSF TeraGrid and Extended Terascale Facility • Complementary and collaborative relationship • Continue to examine prospects for a fiber optical networking facility – National Light Rail
For more information • Web: www.internet2.edu/abilene • E-mail: abilene@internet2.edu