150 likes | 162 Views
Explore experiments on latency, jitter, packet loss, and traffic load in networking scenarios to optimize data delivery and enhance application performance. Evaluate the impact of background and foreground traffic loads on latency and packet loss rates, with a focus on error correction schemes for low-latency data delivery. Consider the trade-offs between redundancy and benefit, TCP vs. UDP vs. FEC/UDP, and the effects of FEC on latency and packet loss rates. Discover how FEC can improve the quality of streamed video and examine the benefit of FEC in small vs. large packets. Instrumentation and testing of teleimmersive applications aim to improve performance through QoS and network tuning. Uncover bottlenecks, monitor network characteristics, and develop techniques to optimize collaborative distributed applications.
E N D
Performance Parameters Latency= Recv Time - Send Time Note: Recv Host and Send Host are synchronized. Jitter = E[{Li - E[L]}] Note: E[ ] is the expection of data set. L is the set of 100 most recent Latency samples. Packet Loss Rate
Latency vs. Time Jitter vs. Time 0.008 0.009 Background Traffic Load 0.008 0.007 0.007 0.006 20Mbps 20Mbps 0.006 0.005 40Mbps 40Mbps 0.005 Jitter Latency 60Mbps 0.004 60Mbps 0.004 80Mbps 80Mbps 0.003 0.003 0.002 0.002 Foreground Traffic Load is 250Kbps 0.001 0.001 0 0 R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R R o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o o Time o o o o o o o o o o o o Time w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w w Packet Lost Rate vs. Background Traffic 0.5 • Note: • These experiments were based on best effort platform. • These experiments will be repeated on DiffServ platform when available. Foreground Traffic Loadis 3 Mbps 0.4 0.3 Packet Lost Rate(%) 0.2 0.1 0 20Mbps 40Mbps 45Mbps 50Mbps 60Mbps 80Mbps Background Traffic
Forward error correction scheme for low-latency delivery of error sensitive data • Ray Fang, Dan Schonfeld, Rashid Ansari • Transmit redundant data over high bandwidth networks that can be used for error correcting UDP streams to achieve lower latency than TCP. • Transmit redundant data to improve quality of streamed video by correcting for lost packets.
FEC Experiments • EVL to SARA- Amsterdam (40Mb/s 200ms RT latency) • Broader Ques: • Can FEC provide a benefit? How much? • Tradeoff between redundancy and benefit? • Specific Ques: • TCP vs UDP vs FEC/UDP • How much jitter does FEC introduce? • High thru put UDP vs FEC/UDP to observe loss & recovery • UDP vs FEC with background traffic • FEC over QoS: WFQ or WRED congestion management- hypothesis: WRED is bad for FEC
UDP Latency (ms) TCP Latency (ms) FEC over UDP Latency (ms) UDP vs TCP vs FEC/UDP with 3:1 redundancy 128 77.0 115 90.3 256B 81.7 121 95.3 512 101.0 150.8 126.0 1024 143.0 210 189.0 2048 227.3 339 314.3 Packet size (bytes)
FEC greatest benefit is in small packets. Larger packets impose greater overhead. As redundancy decreases FEC approaches UDP.
Data Rate (bits/s) Packet Size (Bytes) Packet Loss Rate in UDP (%) Packet Loss Rate in FEC over UDP (%) Packet Loss over UDP vs FEC/UDP 1M 128 0.4 0 1M 256 0.2 0 1M 1024 0.2 0 10M 128 30 4 10M 256 25 3 10M 1024 21 1.5 UDP UDP FEC
Application Level Experiments • Two possible candidates for instrumentation and testing over EMERGE: • Teleimmersive Data Explorer (TIDE) – Nikita Sawant, Chris Scharver • Collaborative Image Based Rendering Viewer (CIBR View) – Jason Leigh, Steve Lau [LBL]
Research Goal: • Hope to see improved performance with QoS and/or TCP tuning enabled. • Monitor applications and characterize their network characteristics as it stands over non-QoS enabled networks. • Idenitfy & remove bottlenecks in the application. • Monitor again to verify bottlenecks removed. • Monitor over QoS enabled networks. • End result is a collection of techniques and tools to help tune similar classes of collaborative distributed applications. • Instrumentation: Time, Info (to identify a flow), Event (to mark a special event), Inter-msg delay, 1-way latency, Read bw, Send bw, Total read, Total sent • TIME=944767519.360357 INFO=Idesk_cray_avatar EVENT=new_avatar_entered MIN_IMD=0.000254 AVG_IMD=0.218938 MAX_IMD=1.170086 INST_IMD=0.134204 MIN_LAT=0.055527 AVG_LAT=0.169372 MAX_LAT=0.377978 INST_LAT=0.114098 AVG_RBW=74.292828 INST_RBW=750.061367 AVG_SBW=429.815557 INST_SBW=704.138274 TOTAL_READ=19019 TOTAL_SENT=110033
QoSiMoto: QoS Internet Monitoring Tool • Kyoung Park • Reads Netlogger data sets from file or from netlogger daemon. • CAVE application runs on SGI and Linux • Information Visualization problem. • How to leverage 3D. • Averaging of data points over long traces. • www.evl.uic.edu/cavern/qosimoto