1 / 18

National Light Rail Footprint

National Light Rail Footprint. SEA. POR. SAC. BOS. NYC. CHI. OGD. DEN. SVL. CLE. WDC. PIT. FRE. KAN. RAL. NAS. STR. LAX. PHO. WAL. ATL. SDG. OLG. DAL. 15808 Terminal, Regen or OADM site Fiber route. NLR Buildout Started November 2002 Initially 4 10 Gb Wavelengths

neila
Download Presentation

National Light Rail Footprint

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. National Light Rail Footprint SEA POR SAC BOS NYC CHI OGD DEN SVL CLE WDC PIT FRE KAN RAL NAS STR LAX PHO WAL ATL SDG OLG DAL 15808 Terminal, Regen or OADM site Fiber route • NLR • Buildout Started November 2002 • Initially 4 10 Gb Wavelengths • To 40 10Gb Waves in Future NREN Backbones reached 2.5-10 Gbps in 2002 in Europe, Japan and US;US: Transition now to optical, dark fiber, multi-wavelength R&E network

  2. Progress: Max. Sustained TCP Thruput on Transatlantic and US Links • 8-9/01 105 Mbps 30 Streams: SLAC-IN2P3; 102 Mbps 1 Stream CIT-CERN • 11/5/01 125 Mbps in One Stream (modified kernel): CIT-CERN • 1/09/02 190 Mbps for One stream shared on 2 155 Mbps links • 3/11/02 120 Mbps Disk-to-Disk with One Stream on 155 Mbps link (Chicago-CERN) • 5/20/02 450-600 Mbps SLAC-Manchester on OC12 with ~100 Streams • 6/1/02 290 Mbps Chicago-CERN One Stream on OC12 (mod. Kernel) • 9/02 850, 1350, 1900 Mbps Chicago-CERN 1,2,3 GbE Streams, OC48 Link • 11-12/02 FAST: 940 Mbps in 1 Stream SNV-CERN; 9.4 Gbps in 10 Flows SNV-Chicago * Also see http://www-iepm.slac.stanford.edu/monitoring/bulk/; and the Internet2 E2E Initiative: http://www.internet2.edu/e2e

  3. Rf (s) TCP AQM p Rb’(s) Theory Internet: distributed feedback system Experiment Geneva 7000km Sunnyvale Baltimore 3000km 1000km Chicago FAST (Caltech): A Scalable, “Fair” Protocol for Next-Generation Networks: from 0.1 To 100 Gbps SC2002 11/02 Highlights of FAST TCP • Standard Packet Size • 940 Mbps single flow/GE card • 9.4 petabit-m/sec • 1.9 times LSR • 9.4 Gbps with 10 flows • 37.0 petabit-m/sec • 6.9 times LSR • 22 TB in 6 hours; in 10 flows • Implementation • Sender-side (only) mods • Delay (RTT) based • Stabilized Vegas Sunnyvale-Geneva Baltimore-Geneva Baltimore-Sunnyvale SC2002 10 flows SC2002 2 flows I2 LSR 29.3.00 multiple SC2002 1 flow 9.4.02 1 flow 22.8.02 IPv6 URL: netlab.caltech.edu/FAST Next: 10GbE; 1 GB/sec disk to disk C. Jin, D. Wei, S. Low FAST Team & Partners

  4. HENP Major Links: Bandwidth Roadmap (Scenario) in Gbps Continuing the Trend: ~1000 Times Bandwidth Growth Per Decade;We are Rapidly Learning to Use and Share Multi-Gbps Networks

  5. HENP Lambda Grids:Fibers for Physics • Problem: Extract “Small” Data Subsets of 1 to 100 Terabytes from 1 to 1000 Petabyte Data Stores • Survivability of the HENP Global Grid System, with hundreds of such transactions per day (circa 2007)requires that each transaction be completed in a relatively short time. • Example: Take 800 secs to complete the transaction. Then Transaction Size (TB)Net Throughput (Gbps) 1 10 10 100 100 1000 (Capacity of Fiber Today) • Summary: Providing Switching of 10 Gbps wavelengthswithin ~3-5 years; and Terabit Switching within 5-8 yearswould enable “Petascale Grids with Terabyte transactions”,as required to fully realize the discovery potential of major HENP programs, as well as other data-intensive fields.

  6. PingER Monitoring Sites IEPM: PingER Deployment • Measurements from • 34 monitors in 14 countries • Over 790 remote hosts; 3600 monitor-remote site pairs • Recently added 23 Sites in 17Countries, due to ICTP Collaboration • Reports on RTT, loss, reachability, jitter, reorders, duplicates … • Measurements go 6ack to Jan-95 • 79 Countries Monitored • Contain > 80% of world population • 99% of online users of the Internet • Mainly A&R sites Remote Sites

  7. History - Throughput Quality Improvements from US 80% annual improvement Factor ~100/8 yr Bandwidth of TCP < MSS/(RTT*Sqrt(Loss)) (1) Progress: but Digital Divide is Maintained (1) Macroscopic Behavior of the TCP Congestion Avoidance Algorithm, Matthis, Semke, Mahdavi, Ott, Computer Communication Review 27(3), July 1997

  8. NREN Core Network Size (Mbps-km):http://www.terena.nl/compendium/2002 100M Logarithmic Scale Leading Nl 10M Fi Cz Advanced Hu Es 1M Ch In Transition It Pl Gr 100k Ir Lagging 10k Ro 1k Ukr 100

  9. Work on the Digital Divide:Several Perspectives • Identify & Help Solve Technical Problems: Nat’l, Regional, Last 10/1/0.1 km • Inter-Regional Proposals (Example: Brazil) • US NSF Proposal (10/2002); possible EU LIS Proposal • Work on Policies and/or Pricing: pk, in, br, cn, SE Europe, … • E.g. RoEduNet (2-6 to 34 Mbps); Pricing not so differentfrom US-CERN price in 2002 for a few Gbps • Find Ways to work with vendors, NRENs, and/or Gov’ts • Use Model Cases: Installation of new advanced fiber infrastructures; Convince Neighboring Countries • Poland (to 5k km Fiber); Slovakia; Ireland • Exploit One-off Solutions: E.g. extend the SILK Project (DESY/FSU satellite links) to a SE European site • Work with other organizations: Terena, Internet2, AMPATH, IEEAF, UN, etc. to help with technical and/or political sol’ns

  10. Digital Divide Committee

  11. Gigabit Ethernet Backbone; 100 Mbps Link to GEANT

  12. GEANT 155Mbps 34Mbps 34Mbps 34Mbps 34Mbps 34Mbps Romania: 155 Mbps to GEANT and Bucharest;Inter-City Links of 2-6 Mbps; to 34 Mbps in 2003 Annual Cost > 1 MEuro

  13. Digital Divide WG Activities • Questionnaire Distributed to the HENP Lab Directors and the Major Collaboration Managements • Plan on Project to Build HENP World Network Map; Updated and Maintained on a Web Site, Backed by Database: • Systematize and Track Needs and Status • Information: Link Bandwidths, Utilization, Quality, Pricing, Local Infrastructure, Last Mile Problems, Vendors, etc. • Identify Urgent Cases; Focus on Opportunities to Help • First ICFA SCIC Workshop: Focus on the Digital Divide • Target Date February 2004 in Rio de Janeiro (LISHEP) • Organization Meeting July 2003 • Plan Statement at the WSIS, Geneva (December 2003) • Install and Leave Behind a Good Network • Then 1 (to 2) Workshops Per Year, at Sites that Need Help

  14. We Must Close the Digital Divide Goal: To Make Scientists from All World Regions Full Partners in the Process of Search and Discovery What ICFA and the HENP Community Can Do • Help identify and highlight specific needs (to Work On) • Policy problems; Last Mile problems; etc. • Spread the message: ICFA SCIC is there to help; Coordinatewith AMPATH, IEEAF, APAN, Terena, Internet2, etc. • Encourage Joint programs [such as in DESY’s Silk project; Japanese links to SE Asia and China; AMPATH to So. America] • NSF & LIS Proposals: US and EU to South America • Make direct contacts, arrange discussions with gov’t officials • ICFA SCIC is prepared to participate • Help Start, or Get Support for Workshops on Networks (& Grids) • Discuss & Create opportunities • Encourage, help form funded programs • Help form Regional support & training groups (requires funding)

  15. “Cultivate and promote practical solutions to delivering scalable,universally available and equitable access to suitable bandwidth and necessary network resources in support of research and education collaborations.” Groningen Carrier Hotel: March 2002 http://www.ieeaf.org

  16. CA-Tokyo by ~1/03 NY-AMS 9/02 (Research)

  17. GlobalMedicalResearchExchangeInitiative Bio-Medicine and Health Sciences St. Petersburg Kazakhstan Uzbekistan NL MD CA Barcelona Greece CN Chenai Navi Mumbai SG GHANA Buenos Aires/San Paolo Layer 1 – Spoke & Hub Sites PERTH Layer 2 – Spoke & Hub Sites Layer 3 – Spoke & Hub Sites Global Quilt Initiative – GMRE Initiative - 001 Propose Global Research and Education Network for Physics

  18. Networks, Grids and HENP • Current generation of 2.5-10 Gbps network backbones arrived in the last 15 Months in the US, Europe and Japan • Major transoceanic links also at 2.5 - 10 Gbps in 2003 • Capability Increased ~4 Times, i.e. 2-3 Times Moore’s • Reliable high End-to-end Performance of network applications(large file transfers; Grids) is required. Achieving this requires: • End-to-end monitoring; a coherent approach • Getting high performance (TCP) toolkits in users’ hands • Digital Divide: Network improvements are especially needed in SE Europe, So. America; SE Asia, and Africa: • Key Examples: India, Pakistan, China; Brazil; Romania • Removing Regional, Last Mile Bottlenecks and Compromises in Network Quality are nowOn the critical path, in all world regions • Work in Concert with APAN, Internet2, Terena, AMPATH; DataTAG, the Grid projects and the Global Grid Forum

More Related