270 likes | 283 Views
Explore the motivation, issues, and lessons learned from the extension of optical networks into the campus, focusing on the CA*net 4 IGT project. This project aims to federate research-based computing resources on campus and provide unencumbered access to end users.
E N D
The extension of optical networks into the campus • Wade Hong • Office of the Dean of Science • Carleton University
Outline • Motivation • CA*net 4 IGT • From the Carleton U Perspective • Issues • Lessons learned
Motivation • Large scale distributed scientific experiments (LHC - ATLAS, SNOLab, Polaris, NEES Grid ... ) • Access to regional distributed HPC resources (HPCVL, SharcNet, WestGrid, TRIUMF Tier 1.5, ...) • Federating growing research-based computing resources on campus • Allowing the end users to access these resources in an unencumbered way • CA*net 4 customer empowered networking last mile
CA*net 4 IGT • CANARIE funded directed research project • build a testbed to experiment with customer empowered networking, pt2pt optical networks, network performance, long haul 10 GbE, UCLP, last mile issues, etc. • participants from the HEP community from across Canada, the provincial ORANs, CERN, StarLight, SURFnet, and potentially others • setup end to end GbE and 10 GbE lightpaths between institutions in Canada and CERN
CA*net 4 IGT • interoperability testing with 10 GbE WAN PHY and OC-192 • used IXIA traffic generators to characterize the trams-atlantic link • transferred real experimental data from ATLAS FCAL beam tests (GbE and 10 GbE) • demonstrated native end-to-end 10 GbE between CERN and Ottawa for the ITU Telecom World 2003
Planned CA*net 4 IGT Activities • complete the last mile connectivity for most of the participating Canadian sites • third OC-192 across Canada being brought up using Nortel OME 6500s • continuing long haul native 10 GbE experiments (Foundry MG8s) • TRIUMF to CERN, TRIUMF to Carleton, Carleton to CERN • CERN to Tokyo via Canada • HEPix Robust Transfer Challenge - sustained disk to disk transfers between TRIUMF and CERN
Planned CA*net 4 IGT Activities • Real-time remote farms for ATLAS • CERN to U of Alberta • Data transfer of End Cap Calorimeter data from the combined beam tests to several Canadian sites • one beam test just completed (~1TB) • second test to start late August (significantly more data) • Transfer of CDF MC data from the Big Mac Cluster • establish a GbE lightpath between UofT and FermiLab
Planned CA*net 4 IGT Activities • Experimentation with bulk data transfer • investigating RDMA/IP (sourcing NICs) • establish GbE lightpaths between Canadian sites
Carleton University • located in Ottawa, the nation’s capital • at the southern end of the world’s longest outdoor skating rink • Canada’s Capital University • student population of 22,000 students, 1700 faculty and staff • over $100M in research funding in the past year • CFI contribution significant • about half to Physics • Bill St. Arnaud’s alma mater
External Network Connectivity • commodity Internet • Telecom Ottawa - was the largest metro 10 GbE deployment • R&E traffic • finally connected to ORION (Dec 2003), the new ORAN, just prior to the decommissioning of ONET • EduNet • non profit, OCRI managed dial-up and High Speed Internet for higher education institutions in Ottawa • dial-up ISP has a dedicated link back to campus
Carleton U Network Upgrade • campus has been in the process of planning a campus network upgrade for the past 3 to 4 years • several false starts • application to funding agencies based on requirements of research activities • may have missed the window of opportunity • finally proceeding with the network upgrade • RFPs currently being evaluated
Network Upgrade Proposal • original proposal • phase one (Year 1) • build the campus core network • phase two (Year 2) • build the distribution layer • phase three (Year 3) • rewire the buildings for access • not my preferred ordering!
Differing Viewpoints • debate over how to handle high capacity research traffic flows • necessity of routing traffic through the proposed high capacity campus core • on the other hand optical bypasses would simplify and reduce the complexity and cost of the campus network • 4 fibre pairs between Herzberg Laboratories and Robertson Hall cost about $4K CDN - we prevailed • reality check • current campus network cannot handle the high volume and high speed flows
Motivations Revisited • Large scale distributed scientific experiments
Motivations Revisited • Access to regional distributed HPC resources • other HPCVL sites (Queens, UofO, RMC, Ryerson U) • TRIUMF ATLAS Canada computing centre • SNOLab • shared ORION and CA*net 4 connectivity is only at GbE • high capacity flows probably dictate pt2pt optical bypass • interconnectivity can be static or dynamic • fully statically meshed or scheduled dynamic connectivity on demand - probably the latter
Motivations Revisited • Federating growing research-based computing resources into a campus grid • HPCVL Linux cluster upgrade (128+256 CPUs) • Physics research cluster upgrade (40+96+96 CPUs) • Civil Engineering (~128 CPUs) • Architecture/Psychology visualization cluster (>128 CPUs) • Systems and Computer Engineering ( 64 CPUs) • debating a condominium or distributed model • most likely a hybrid with optical fibre as the interconnecting fabric • probably static pt2pt optical bypass for ease of use and user control
Motivations Revisited • federated the Physics research computing cluster with part of the HPCVL Linux cluster last summer for about 2 months • clusters located on different floors • pt2pt link established - much easier than routing through the campus network • completed half of the MC regeneration for the third SNO paper • similar arrangement this summer to add part of the HPCVL cluster to the Carleton U Physics contribution to the LHC Computing Grid till the end of the year
Issues • control • central management and control vs end user empowerment • disruptive • network complexity • using pt2pt ethernet links for high capacity flows should simplify campus networks (reduce costs?) • security • disruptive - bypassing DMZ • for the uses considered here, the pt2pt links are inherently secure - non routed private subnets
Issues • why not copper? • it could be but with fibre • greater distances • requires less active devices along the path • management and control - device at each end under the control of the end users is ideal • consistent device characteristics - jumbo frames, port speed, duplex, etc. • inter-building connectivity is fibre and planned vertical cabling will be fibre
Issues • last mile connectivity • demarcation point • end user device (NIC) or an edge device (switch, CWDM mux) • location of the demarc • at the end user or a common shared location • technology used to extend the end to end lightpath into the campus • pt2pt GbE • optical GbE NIC - patched thru to GbE interface on ONS • media converter - copper to optical
Issues • pt2pt 10GbE • LAN PHY to WAN PHY conversion to OC192c on ONS 15454/OME 6500 • wavelength conversion • CWDM • media converters - copper to colored wavelength • colored GBICS for GbE switch • optical link charateristics • padding (attenuation), proper power budget, etc. • end user shouldn’t need to be an optical networking expert
Lessons Learned • good to be rich in fibre • provides greater flexibility • support of ORANs, national R&E network, and international partners is essential - all have been very supportive • need to convince local campus networking folks that this is not really too disruptive • will simplify and not burden the campus production network • need a more coherent way of dealing with optical access in the last mile • still lots to learn!
Thank You! Wade Hong xiong@physics.carleton.ca