1 / 11

On the Placement of Web Server Replicas

This paper discusses the problem of placing web server replicas in order to improve performance. It presents several placement algorithms and evaluates their performance through simulations on synthetic and real networks.

Download Presentation

On the Placement of Web Server Replicas

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. On the Placement of Web Server Replicas Yu Cai

  2. Paper • On the Placement of Web Server Replicas • Lili Qiu, Venkata N. Padmanabhan, Geoffrey M. Voelker • Infocom 2001

  3. What is the paper talking about? • Web server replicas placement problem. • A popular Web site aims to improve its performance (e.g., reducing its clients’ perceived latency) by pushing its content to some hosting services. • The problem is to choose M replicas (or hosting services) among N potential sites (N > M) such that some objective function is optimized under a given traffic pattern. • The objective function can be minimizing either its clients’ latency, or its total bandwidth consumption, or an overall cost function if each link is associated with a cost.

  4. Contribution of the paper • Present several placement algorithms • Evaluate the performance by simulating on synthetic and real network.

  5. Algorithms • Tree based algorithm: • The underlying topologies are trees, and modeled it as a dynamic programming problem. • The algorithm was originally designed for Web proxy cache placement, and it is also applicable for Web replica placement. • Unrealistic assumption and not so good performance under normal network.

  6. Algorithms • Greedy algorithm • Evaluate N potential site to determine its suitability by assuming all clients traffic converge at this site • Pick the best one • Repeat step 1 and 2 for the rest N-1 site, • Until we pick M sites.

  7. Algorithms • Random algorithm • Randomly pick M sites from N sites. • Can be improve by introducing genetic evolution. • Hotspot algorithm • Place replicas near the clients generating the greatest load. • Can be improve by using client clustering. • Super optimal algorithm to get lower bound • May not be feasible solution, only used for comparison.

  8. Simulation • Use GT-ITM to generate random network topology • Transit-stub, hierarchical graph • Use BGP routing tables to generate real Internet topology • AS hop count ???

  9. Simulation • Web workload and client generation • Use the access logs collected from real web sites, like MSNBC • Cluster the web clients who are topologically close to each other • Top 10, 100, 1000 and 3000 clusters account for 24%, 45%, 78% and 94% requests. • Map the clusters randomly to the nodes in the simulation network

  10. Simulation • Evaluate the effects of imperfect data • Salt the input data with random noise of uniform distribution. • Evaluate the effects of dynamic network • Input data change over time.

  11. Conclusion • Greedy perform the best • Error increase or network changes, the performance degrades slightly.

More Related