1 / 27

Achieving Dependable Bulk Throughput in a Hybrid Network

This paper discusses the concept of using session-layer depots to improve throughput in a hybrid network. It explores the use of depots in the Phoebus experiment and presents early performance results. The Phoebus project aims to narrow the performance gap in wide-area bulk data flows by leveraging both packet infrastructure and dynamically-provisioned lambdas.

rconway
Download Presentation

Achieving Dependable Bulk Throughput in a Hybrid Network

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Achieving Dependable Bulk Throughput in a Hybrid Network Guy Almes <almes@internet2.edu> Aaron Brown <aaronmb@udel.edu> Martin Swany <swany@cis.udel.edu> Joint Techs Meeting Univ Wisconsin -- 17 July 2006

  2. Outline • Observations: • on user needs and technical opportunities • on TCP dynamics • Notion of a Session Layer • the obvious application • a stronger application • Phoebus as HOPI experiment • deployment • early performance results • Phoebus as an exemplar hybrid network

  3. On User Needs • In a variety of cyberinfrastructure-intensive applications, dependable high-speed wide-area bulk data flows are of critical value • Examples: • Terabyte data sets in HPC applications • Data-intensive TeraGrid applications • Access to Sloan Digital Sky Survey and similar very large data collections • Also, we stress ‘dependable’ rather than ‘guaranteed’ performance • As science becomes more data-intensive, these needs will be prevalent in many science disciplines

  4. On Technology Drivers • Network capacity increases, but user throughput increases more slowly • Source: DOE • The cause of this gap relates to TCP dynamics

  5. On TCP Dynamics • Consider the Mathis Equation for Reno • Focus on bulk data flows over wide areas • How can we attack it? • Reduce non-congestive packet loss (a lot!) • Raise the MTU (but only helps if end-to-end!) • Improve TCP algorithms (e.g., FAST, Bic) • RTT is still a factor • Use end-to-end circuits • Decrease RTT??

  6. Situation for running example

  7. The Transport-Layer Gateway User Space Transport Network Data Link Session Session Physical Transport Transport Network Network Data Link Data Link Physical Physical • A session is the end-to-end chain of segment-specific transport connections • In our early work, each of these transport connections is a conventional TCP connection • Each transport-level gateway (depot) receives data from one connection and pipes it to the next connection in the chain

  8. The Logistical Session Layer

  9. Obvious Application • Place a depot half-way between hosts A and B, thus cut the RTT roughly in half • Bad news: only a small factor • Good news: it actually does more

  10. Obvious Application: With one depot to reduce RTT

  11. Stronger Application • Place one depot at HOPI node near the source, and another near the destination • Observe: • Abilene Measurement Infrastructure: • 2nd percentile: 950 Mb/s median: 980 Mb/s • MTU = 9000 bytes; loss is very low • Local infrastructure: • MTU and loss are good, but not always very good • but the RTT is very small • But with HOPI we can do even better

  12. The HOPI Project • The Hybrid Optical and Packet Infrastructure Project (hopi.internet2.edu) • Leverage both the 10-Gb/s Abilene backbone and a 10-Gb/s lambda of NLR • Explore combining packet infrastructure with dynamically-provisioned lambdas

  13. Stronger application: depots near each host Backbone: large RTT 9000-byte MTU very low non-congestive loss GigaPoP / Campus: very small RTT some 1500-byte MTU some non-congestive loss

  14. Two Conjectures • Small RTT does effectively mask moderate imperfections in MTU and loss • End-to-end session throughput is (only a little less than) the minimum of component connection throughputs

  15. Phoebus • Phoebus aims to narrow the performance gap by bringing revolutionary networks like HOPI to users • Phoebus is another name for the mythical Apollo in his role as the “sun god” • Phoebus stresses the ‘session’ concept to enable multiple network/transport infrastructures to be catenated • Phoebus builds on an earlier project called the Logistical Session Layer (LSL)

  16. Experimental Phoebus Deployment • Place Phoebus depots at each HOPI node • Ingress/egress spans via ordinary Internet2/ Abilene IP infrastructure • Backbone span can use either/both of: • 10-Gb/s path through Abilene • dynamic 10-Gb/s lambda • Initial test user sites: • SDSC host with gigE connectivity • Columbia Univ host with gigE connectivity

  17. Initial Performance Results • In very early tests: • SDSC to losa: about 900 Mb/s • losa to nycm: about 5.1 Gb/s • nycm to Columbia: about 900 Mb/s • direct: 380 ± 88 Mb/s • Phoebus: 762 ± 36 Mb/s • In later tests with a variety of file sizes, SDSC to losa performance became worse

  18. Initial Performance Results

  19. Initial Test Results • What about the three components? • SDSC to losa depot: 429-491 Mb/s • losa depot to nycm depot: 5.13-5.15 Gb/s • nycm depot to Columbia: 908-930 Mb/s • Whatever caused that weakness in the SDSC-to-losa path did slow things down

  20. Plans for Summer 2006 • ‘Experimental production’ Phoebus, reaching out to interested users • Improve access control and instrumentation: • Maintain a log of achieved performance • Test use of dynamic HOPI lambdas • Evaluate Phoebus as a service within newnet • Test use of Phoebus internationally

  21. Comments on Backbone Span • Backbone could ensure flow performance between pairs of backbone depots • Backbone could provide a Phoebus Service in addition to its “IP” service • Relatively easy to use dynamic lambdas within the backbone portion of the Phoebus infrastructure • Alternatively, the backbone portion could use IP, but a non-TCP transport protocol!

  22. Comments on the Local (Ingress and Egress) Spans • Near ends, we have good, but not perfect, local/metro-area infrastructure • Relatively hard to deploy dynamic lambdas • Small RTTs allow high-speed TCP flows to be extended to many local sites in a scalable way

  23. Thus, Phoebus leverages both: • innovative wide-area infrastructure and • conventional local-area infrastructure • Phoebus can thus extend the value of multi-lambda wide-area infrastructure to many science users on high-quality conventional campus networks

  24. Ongoing Work • Phoebus deployment on HOPI • We’re seeking project participants! • Please email for information • ESP-NP • ESP = Extensible Session Protocol • Implementation on an IXP Network Processor from Intel • The IXP2800 can forward at 10 Gb/s

  25. Acknowledgements • UD Students: • Aaron Brown, Matt Rein • Internet2: • Eric Boyd, Rick Summerhill, Matt Zekauskas, ... • HOPI Testbed Support Center (TSC) Team • MCNC, Indiana Univ NOC, Univ Maryland • San Diego Supercomputer Center: • Patricia Kovatch, Tony Vu • Columbia University: • Alan Crosswell, Megan Pengelly, the Unix group • Dept of Energy Office of Science: • MICS Early Career Principal Investigator program

  26. End • Thank you for your attention • Questions?

More Related