1 / 22

Experiences with a Large-Scale Deployment of Stanford Peer-to-Peer Multicast

Experiences with a Large-Scale Deployment of Stanford Peer-to-Peer Multicast. Jeong h un Noh , Pierpaolo Baccichet * , Bernd Girod Image, Video and Multimedia Systems Group Information Systems Laboratory Stanford University * Now with Dyyno Inc. Outline.

sef
Download Presentation

Experiences with a Large-Scale Deployment of Stanford Peer-to-Peer Multicast

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Experiences with a Large-Scale Deploymentof Stanford Peer-to-Peer Multicast Jeonghun Noh, Pierpaolo Baccichet*, Bernd Girod Image, Video and Multimedia Systems Group Information Systems Laboratory Stanford University *Now with Dyyno Inc.

  2. Outline • Review of Stanford P2P Multicast (SPPM) • Experimental setup • System configuration • P2P multicast of ESWC • Analysis • Peer statistics • System performance

  3. Stanford P2P Multicast (SPPM) • P2P-based low-latency and video-aware live streaming • Multiple-multicast-tree overlay • Pushing video: reduces end-to-end delay and duplicate video packets • Path diversity: improves resilience • Fine-grained bandwidth aggregation • Video/network aware packet handling • Tested on PlanetLab [CoopNet 2002, SplitStream 2003] [Setton et al., P2PMMS 2005] [Baccichet et al., ICME 2007] [Setton et al., Proc. IEEE 2008]

  4. Multiple Complementary Multicast Trees Video stream … … Tree 1 Tree 2

  5. I B P P P B B Packet Scheduling 7 1 6 1 4 1 2 Child 1 Parent Child 2 [Setton et al., ICIP 2006]

  6. I B P P … … Local Retransmission Parent 1 Child P Parent 2 fails Video Buffer [Setton et al., ICME 2006]

  7. Outline • Overview of Stanford P2P Multicast • Experimental setup • System configuration • P2P multicast of ESWC • Analysis • Peer statistics • System performance

  8. System Configuration Video source Requests for missing packets Retransmission Server Super node(s) Tree 1 Tree 2 System used for data collection

  9. Technical Details • P2P streaming of ESWC 2008 • Video • CG 3D game contents • Encoded using H.264/AVC • Bitrate: 600 kbps(video: 560kbps) • Resolution: 640 by 480 pixels • End-to-end delay < 10 seconds • Data logging • Peer statistics • Packet loss ratio

  10. Outline • Overview of Stanford P2P Multicast • Experimental setup • System configuration • P2P multicast of ESWC • Analysis • Peer statistics • System performance

  11. Audience Size Day 2 Day 1 Day 3

  12. Peer Uplink Bandwidth • 20% of users supply 70% of aggregate uplink bandwidth • Multiple trees allow more peers to contribute

  13. Peer Lifetime

  14. Departure Rate 30 minutes later… Will it stay in system?

  15. Peer Accessibility STATIC IP (Firewall) RESTR. PORT SYMMETRIC FULL CONE RESTR. IP STATIC IP STATIC IP FULL CONE STATIC IP (Firewall) RESTR. IP Access Type Distribution RESTR. PORT SYMMETRIC

  16. Average Uplink Bandwidth Data from Day 3

  17. Packet Loss Ratio (PLR)

  18. PLR over Time Data from Day 3

  19. Conclusions • Analysis of rich peer statistics • 80/20 rule • 20% of users supply 70% of uplink bandwidth • 85% of users come from 20% of countries • Older peers are more reliable than younger peers • Server assistance • 2.3x aggregate bandwidth w/o assistance • Extra bandwidth from super nodes • SPPM achieves low-latency transmission delay w/ low packet loss

  20. Thank you! Email: jhnoh@stanford.edu

  21. Uplink Contribution – 80/20 Rule

  22. Active Overlay Management [Noh et al., ICME 2008] [Noh et al., Globecom 2008]

More Related