1 / 33

Low latency via redundancy

Low latency via redundancy. Ashish Vulimiri , P. Brighten Godfrey Radhika Mittal, Justin Sherry Sylvia Ratnasamy , Scott Shenker. Presented by Xuzi Zhou. Outline. Introduction System View Queuing Analysis Applications Individual View Conclusion. Introduction.

kaleb
Download Presentation

Low latency via redundancy

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Low latency via redundancy Ashish Vulimiri, P. Brighten Godfrey Radhika Mittal, Justin Sherry Sylvia Ratnasamy, Scott Shenker Presented by Xuzi Zhou

  2. Outline Introduction System View Queuing Analysis Applications Individual View Conclusion CS 685 Fall 2013 Paper Presentation

  3. Introduction Why do we want low latency? CS 685 Fall 2013 Paper Presentation

  4. Introduction People react to small differences in latency For a website: Higher latency = fewer visits = lower revenue Exponential distribution The tail of the distribution is critical Possible causes Server overload Network congestion Packet loss … About latency CS 685 Fall 2013 Paper Presentation

  5. Introduction Use redundancy Duplicate an operation Use diverse resources Use the first received result How to reduce latency? CS 685 Fall 2013 Paper Presentation

  6. Introduction “Less work is best” Redundancy caused higher system utilization Network bandwidth cost Computation cost Effectiveness of redundancy is unclear When redundancy improves latency? When not? What’s the gain from redundancy? Why not widely used? CS 685 Fall 2013 Paper Presentation

  7. Outline Introduction System View Queuing Analysis Applications Individual View Conclusion CS 685 Fall 2013 Paper Presentation

  8. SYSTEM VIEW Find the system threshold Under the threshold: Improve latency Above the threshold: Worsen latency Queuing Analysis CS 685 Fall 2013 Paper Presentation

  9. SYSTEM VIEW Queuing Analysis Service time distributions: Deterministic Variable (Using Pareto distribution) CS 685 Fall 2013 Paper Presentation

  10. SYSTEM VIEW If client-side cost of redundancy is negligible: Deterministic service time is the worst case Threshold is around 25% Service time distribution is varying with tail index Threshold > 30% Queuing Analysis CS 685 Fall 2013 Paper Presentation

  11. SYSTEM VIEW Effect of client-side overhead: Client-side overhead must be smaller than mean request latency to improve mean latency Queuing Analysis CS 685 Fall 2013 Paper Presentation

  12. Outline Introduction System View Queuing Analysis Applications Individual View Conclusion CS 685 Fall 2013 Paper Presentation

  13. SYSTEM VIEW Request: Download a random file from the data store Base configuration: Mean file size: 4KB File size distribution: deterministic Memory cache ratio: 0.1 # servers: 4 # clients: 10 Use Emulab nodes Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  14. SYSTEM VIEW 1. Base Configuration Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  15. SYSTEM VIEW 2. Mean file size 0.04 KB instead of 4 KB Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  16. SYSTEM VIEW 3. Pareto file size distribution instead of deterministic Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  17. SYSTEM VIEW 4. Cache:disk ratio 0.01 instead of 0.1 Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  18. SYSTEM VIEW 5. EC2 nodes instead of Emulab Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  19. SYSTEM VIEW 6. Mean file size 400 KB instead of 4 KB Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  20. SYSTEM VIEW 7. Cache:disk ratio 2 instead of 0.1 Application: disk-backed data store CS 685 Fall 2013 Paper Presentation

  21. SYSTEM VIEW Test with the memcached in-memory database Normal Version Request database directly Stub Version: Call stub instead of database Return results immediately Estimate effect of client-side latency Application: memcached CS 685 Fall 2013 Paper Presentation

  22. SYSTEM VIEW Application: memcached CS 685 Fall 2013 Paper Presentation

  23. SYSTEM VIEW Application: memcached CS 685 Fall 2013 Paper Presentation

  24. SYSTEM VIEW Simulated fat-tree data center # servers: 54 Use standard data center workload: Flow size range: 1KB to 3 MB Ratio of flows with size < 10 KB: 80% Method: Every switch replicates the first few (eight) packets of each flow along an alternate route Replicated packets have lower priority Application: replication in the network CS 685 Fall 2013 Paper Presentation

  25. SYSTEM VIEW For flows smaller than 10 KB Application: replication in the network CS 685 Fall 2013 Paper Presentation

  26. Outline Introduction System View Queuing Analysis Applications Individual View Conclusion CS 685 Fall 2013 Paper Presentation

  27. Individual VIEW Replicate TCP-handshake packets Send two back-to-back copies of a packet Reduce probability of packet loss In PlanetLab tests: Probability of Individual packet loss: 0.0048 Probability of back-to-back packet pair loss: 0.0007 0.0048 >> 0.0007 >>0.00482 Reduce average completion time of handshake by 25 ms in an idealized network with 3-second timeout ofr SYN/SYN-ACK packets Application: connection establishment CS 685 Fall 2013 Paper Presentation

  28. Individual VIEW Setup: # Clients: 15 PlanetLab nodes across the continental US # DNS servers: 10 Local DNA server and famous public DNS servers # website names: 1 million Method: Every node ranks the DNS servers according to response time Query the TOP n DNS servers with a random website name, n = 1, 2, …, 10 Application: DNS CS 685 Fall 2013 Paper Presentation

  29. Individual VIEW Application: DNS CS 685 Fall 2013 Paper Presentation

  30. Individual VIEW Application: DNS CS 685 Fall 2013 Paper Presentation

  31. Outline Introduction System View Queuing Analysis Applications Individual View Conclusion CS 685 Fall 2013 Paper Presentation

  32. Conclusion Redundancy improves latency under a certain system load threshold (normally between 25% - 50%) when client-side cost of redundancy is low Redundancy offers a significant benefit in a number of practical applications, both in the Internet and in the data center. Redundancy should be used more commonly in network systems. CS 685 Fall 2013 Paper Presentation

  33. QUESTIONS? CS 685 Fall 2013 Paper Presentation

More Related