340 likes | 472 Views
FAST TCP. Cheng Jin David Wei Steven Low. netlab. CALTECH .edu. GNEW, CERN, March 2004. Acknowledgments. Caltech Bunn, Choe, Doyle, Jin, Newman, Ravot, Singh, J. Wang, Wei UCLA Paganini, Z. Wang CERN/DataTAG Martin, Martin-Flatin Internet2 Almes, Shalunov SLAC Cottrell, Mount
E N D
FAST TCP Cheng Jin David Wei Steven Low netlab.CALTECH.edu GNEW, CERN, March 2004
Acknowledgments • Caltech • Bunn, Choe, Doyle, Jin, Newman, Ravot, Singh, J. Wang, Wei • UCLA • Paganini, Z. Wang • CERN/DataTAG • Martin, Martin-Flatin • Internet2 • Almes, Shalunov • SLAC • Cottrell, Mount • Cisco • Aiken, Doraiswami, Yip • Level(3) • Fernes • LANL • Wu
NSF RI (2003) HEP networks TeraGrid Abilene IETF GGF Deployment NSF STI (2002) Experiment DummyNet HEP networks Abilene PlanetL WAN in Lab UltraLight testbed Implement NSF ITR (2001) Linux TCP kernel Other platforms Monitoring Debugging Theory Performance Stability Fairness TCP/IP Noise Random-ness FAST project
Outline • Experiments • Results • Future plan • Status • Open issues • Code release mid 04 • Unified framework • Reno, FAST, HSTCP, STCP, XCP, … • Implementation issues
Aggregate throughput 88% FAST • Standard MTU • Utilization averaged over > 1hr 90% 90% Average utilization 92% 95% 1.1hr 6hr 6hr 1hr 1hr 1 flow 2 flows 7 flows 9 flows 10 flows DataTAG: CERN – StarLight – Level3/SLAC (Jin, Wei, Ravot, etc SC2002)
Dynamic sharing: 3 flows FAST Linux Dynamic sharing on Dummynet • capacity = 800Mbps • delay=120ms • 3 flows • iperf throughput • Linux 2.4.x (HSTCP: UCL)
Dynamic sharing: 3 flows FAST Linux Steady throughput HSTCP STCP
30min queue FAST Linux loss throughput Dynamic sharing on Dummynet • capacity = 800Mbps • delay=120ms • 14 flows • iperf throughput • Linux 2.4.x (HSTCP: UCL) HSTCP STCP
30min queue Room for mice ! FAST Linux loss throughput HSTCP STCP HSTCP
ideal performance Aggregate throughput Dummynet: cap = 800Mbps; delay = 50-200ms; #flows = 1-14; 29 expts
small window 800pkts large window 8000 Aggregate throughput Dummynet: cap = 800Mbps; delay = 50-200ms; #flows = 1-14; 29 expts
HSTCP ~ Reno Jain’s index Fairness Dummynet: cap = 800Mbps; delay = 50-200ms; #flows = 1-14; 29 expts
stable in diverse scenarios Stability Dummynet: cap = 800Mbps; delay = 50-200ms; #flows = 1-14; 29 expts
Outline • Experiments • Results • Future plan • Status • Open issues • Code release • Unified framework • Reno, FAST, HSTCP, STCP, XCP, … • Implementation issues
Benchmarking TCP • Not just static throughput • Dynamic sharing, what protocol does to network, … • Tests to zoom in on specific properties • Throughput, delay, loss, fairness, stability, … • Critical for basic design • Test scenarios may not be realistic • Tests with realistic scenarios • Same performance metrics • Critical for refinement for deployment • Just started • Input solicited • What’s realistic for your applications?
Open issues: well understood • baseRTT estimation • route changes, dynamic sharing • does not upset stability • Small network buffer • at least like TCP • adapt a on slow timescale, but how? • TCP-friendliness • friendly at least at small window • tunable, but how to tune? • Reverse path congestion • should react? rare for large transfer?
Status: code release • Source release mid 2004 • For any non-profit purposes • Re-implementation of FAST TCP completed • Extensive testing to complete by April 04 • Pre-release trials • CFP for high-performance sites! • Incorporate into Web100 • with Matt Mathis
Status: IPR Caltech will license royalty-free if FAST TCP becomes IETF standard • IPR covers more broadly than TCP • Leave all options open
Outline • Experiments • Results • Future plan • Status • Open issues • Code release mid 04 • Unified framework • Reno, FAST, HSTCP, STCP, XCP, … • Implementation issues
ACK: W W + 1/W Loss: W W – 0.5W • Packet level • Flow level • Equilibrium • Dynamics pkts Packet & flow level Reno TCP (Mathis formula)
Reno TCP • Packet level • Designed and implemented first • Flow level • Understood afterwards • Flow level dynamics determines • Equilibrium: performance, fairness • Stability • Design flow level equilibrium & stability • Implement flow level goals at packet level
Reno TCP • Packet level • Designed and implemented first • Flow level • Understood afterwards • Flow level dynamics determines • Equilibrium: performance, fairness • Stability Packet level design of FAST, HSTCP, STCP, H-TCP, … guided by flow level properties
ACK: W W + 1/W Loss: W W – 0.5W • Reno AIMD(1, 0.5) ACK: W W + a(w)/W Loss: W W – b(w)W • HSTCP AIMD(a(w), b(w)) ACK: W W + 0.01 Loss: W W – 0.125W • STCP MIMD(a, b) • FAST Packet level
Flow level: Reno, HSTCP, STCP, FAST • Similarflow level equilibrium MSS/sec a = 1.225 (Reno), 0.120 (HSTCP), 0.075 (STCP)
Flow level: Reno, HSTCP, STCP, FAST • Commonflow level dynamics window adjustment control gain flow level goal = • Different gain k and utility Ui • They determine equilibrium and stability • Different congestion measure pi • Loss probability (Reno, HSTCP, STCP) • Queueing delay (Vegas, FAST)
FAST TCP • Reno, HSTCP, and FAST have commonflow level dynamics window adjustment control gain flow level goal = • Equation-based • Need to estimate “price” pi(t) • pi(t) = queueing delay • Easier to estimate at large window • k(t) and U’i(t) explicitly designed for • Performance • Fairness • Stability
Window control algorithm • Full utilization • regardless of bandwidth-delay product • Globally stable • exponential convergence • Intra-protocol fairness • weighted proportional fairness • parameter a
Goal: • Less delay • Less jitter FAST stabilized FAST tunes to knee TCP oscillation
Window adjustment FAST TCP
netlab.caltech.edu/FAST • FAST TCP: motivation, architecture, algorithms, performance IEEE Infocom March 2004 • FAST TCP: from theory to experiments Submitted for publication April 2003
Metrics • Performance • Throughput, loss, delay, jitter, stability, responsiveness • Availability, reliability • Simplicity • Application • Management • Evolvability, robustness
Constraints • Scientific community • Small & fixed set of major sites • Few & large transfers • Relatively simple traffic characteristics and quality requirements • General public • Large, dynamic sets of users • Diverse set of traffic characteristics & quality requirements • Evolving/unpredictable applications
Mechanisms • Fiber infrastructure • Lightpath configuration • Resource provisioning • Traffic engineering, adm control • Congestion/flow control Months - years Mintes - days Service: sec - hrs Flow: sec - mins RTT: ms - sec • Timescale: desired, instead of feasible • Balance: cost/benefit, simplicity, evolvability