1 / 29

The Effects of Wide-Area Conditions on WWW Server Performance

Motivation: Benchmarking Today. . Motivation: Real Life. . Web Server Performance. Workload Generators Webstone, SpecWeb, SURGE, s-client, httperf, etc. Based on measured traffic behavior Reproducible- WAN case is ignored: no drops, delays, etc.. Live Server Analysis California elections, 96 Olympics, WAWM Capture real WAN conditions- But not reproducible.

Faraday
Download Presentation

The Effects of Wide-Area Conditions on WWW Server Performance

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


    1. The Effects of Wide-Area Conditions on WWW Server Performance Erich Nahum, Marcel Rosu, Srini Seshan, Jussara Almeida Hi. Thanks for being here, and hope you enjoy the talk! Hi. Thanks for being here, and hope you enjoy the talk!

    2. Motivation: Benchmarking Today The way people do benchmarking today: grab a 12-way SMP, add 12 gigabit Ethernet interfaces, lots of clients, and blast away to get the best SpecWeb number possible. IBM is no exception.The way people do benchmarking today: grab a 12-way SMP, add 12 gigabit Ethernet interfaces, lots of clients, and blast away to get the best SpecWeb number possible. IBM is no exception.

    3. Motivation: Real Life Unfortunately, vast majority of sites are connected over the actual Internet, or perhaps some corporate Intranet. Clients have diversity, Internet has variability, fluctuations, etc. These are manifested as packet drops, delays, Reorderings, duplications, etc.Unfortunately, vast majority of sites are connected over the actual Internet, or perhaps some corporate Intranet. Clients have diversity, Internet has variability, fluctuations, etc. These are manifested as packet drops, delays, Reorderings, duplications, etc.

    4. Web Server Performance Workload Generators Webstone, SpecWeb, SURGE, s-client, httperf, etc. + Based on measured traffic behavior + Reproducible - WAN case is ignored: no drops, delays, etc. Workload generators are derived from traces, reproducible, but ignore WAN case Live servers are hard to do, and can be difficult to reproduce What we want is the best of both worldsWorkload generators are derived from traces, reproducible, but ignore WAN case Live servers are hard to do, and can be difficult to reproduce What we want is the best of both worlds

    5. Outline Motivation and Background The WASP Environment Hardware and software Workload generators Results Effects of packet loss Effects of packet delay Effects of TCP variants Summary and Conclusions

    6. Wide-Area Server Performance What WASP is not: Doesn’t reproduce a specific web site Doesn’t reproduce a specific network topology Reproducing a specific web site is essentially impossible Instead, we vary parameters over a range of `realistic’ values. Operators can configure the system based on their own traffic measurementsReproducing a specific web site is essentially impossible Instead, we vary parameters over a range of `realistic’ values. Operators can configure the system based on their own traffic measurements

    7. Centralized Approach

    8. WASP Approach Each client acts as a ‘WAN emulator’ Use DummyNet to drop and delay packets

    9. Scaling with Load This curve compares the two approaches doing nothing (no loss, no delay). For small responses, centralized and WASP approaches both work fine. But the centralized approach starts breaking down under large transfers and large loads.This curve compares the two approaches doing nothing (no loss, no delay). For small responses, centralized and WASP approaches both work fine. But the centralized approach starts breaking down under large transfers and large loads.

    10. Packet Loss Model Two-state loss model based on work by Bolot 93, Paxson 97, Rubenstein et al. 2000 Packets forwarded in good state, dropped in bad Transitions based on desired loss rate Loss probability of a packet following a previously lost packet is much higher than average loss rate. Here we use the loss event probability; note that this is not the same as the average loss rate. Paper has table explaining the relationship.Loss probability of a packet following a previously lost packet is much higher than average loss rate. Here we use the loss event probability; note that this is not the same as the average loss rate. Paper has table explaining the relationship.

    11. Workload Generators S-client (from Rice), SURGE (from BU) WaspClient integrates the two Singleclient uses event-driven architecture to scale to large numbers of connections SURGE captures all kinds of relevant characteristics WaspClient combines the twoSingleclient uses event-driven architecture to scale to large numbers of connections SURGE captures all kinds of relevant characteristics WaspClient combines the two

    12. Putting it all together Ultimately we are a testbed, not a live server. We just play one on TV. Server is well-connected: we want all bottlenecks to be ones we introduce.Ultimately we are a testbed, not a live server. We just play one on TV. Server is well-connected: we want all bottlenecks to be ones we introduce.

    13. Experimental Methodology Performance Metrics: Server throughput, utilization, response time, capacity Sensitivity Analysis: Vary generated load in SURGE UE’s Vary loss rate from 0 to 9 % Vary RTT from 0 to 400 msec Parameters taken from Paxson97, Allman2000 Methodology: Average of 10 runs Each run lasts 10 minutes 90 % confidence intervals

    14. Outline Motivation and Background The WASP Environment Hardware and software Workload generators Results Effects of packet loss Effects of packet delay Effects of TCP variants Summary and Conclusions

    15. Throughput vs. Loss Rate

    16. Utilization vs. Loss Rate

    17. What’s going on?

    18. Latency vs. Loss Rate

    19. Capacity vs. Loss Rate Non-overlapping columnsNon-overlapping columns

    20. Outline Motivation and Background The WASP Environment Hardware and software Workload generators Results Effects of packet loss Effects of packet delay Effects of TCP variants Summary and Conclusions

    21. Throughput vs. RTT Try to make produced gif transparent Larger fonts on graphsTry to make produced gif transparent Larger fonts on graphs

    22. Utilization vs. RTT

    23. Latency vs. RTT

    24. Capacity vs. RTT

    25. Many Variants of TCP: Reno (current baseline in the Internet): Coarse-grained timeouts, fast retransmit Recovers 1 lost segment every 3 RTT’s New Reno: Uses partial acknowledgement to improve loss recovery Recovers 1 lost segment every RTT Sender-side only modification Selective Acknowledgements (SACK): Uses SACK option bit field to improve loss recovery Recovers up to 3 segments per RTT Requires modifications to both sender and receiver Other schemes exist (e.g., Vegas)

    26. TCP Variants: Latency Not exposed by benchmarks such as SpecWeb99Not exposed by benchmarks such as SpecWeb99

    27. Summary Presented the WASP environment Emulates WAN conditions in a controlled setting Scalable, reproducible, configurable Several results: Delays and losses affect performance Loss reduces capacity, increases latency Delays increase latency but not capacity SACK, New Reno can reduce response time, don’t affect capacity Other fallout: Used to find bugs in AIX, Flash, AFPA (IBM server) Convinced AIX group to deploy SACK & New Reno

    28. Future Directions HTTP 1.1 Linux Bandwidth limitations Dynamic content Other workloads: Proxies Clients SSL

    29. Apache Capacity vs. Loss

    30. Apache Capacity vs. RTT

More Related