1 / 18

High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research

High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research. Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC San Diego/UC Irvine). Abstract.

emory
Download Presentation

High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. High Performance Cyberinfrastructure Discovery Tools for Data Intensive Research Larry Smarr Prof. Computer Science and Engineering Director, Calit2 (UC San Diego/UC Irvine)

  2. Abstract High performance cyberinfrastructure (10Gbps dedicated optical channels end-to-end) enable new levels of discovery for data-intensive research projects. I will use several different examples of large data projects drawn from cosmological simulations, ocean observing, and microbial metagenomics. I will discuss why local campus high performance clouds are essential for this sort of work in academia--as a high bandwidth, high I/O fast storage, large RAM compute augmentation to the remote commercial cloud.

  3. Dedicated 10,000Mbps (10Gbps) SupernetworksEnable Remote Visual Analysis of Big Data Also NLR 80 x 10Gb Wavelengths

  4. NSF’s OptIPuter Project: Using Supernetworksto Meet the Needs of Data-Intensive Researchers OptIPortal– Termination Device for the OptIPuter 10Gbps Backplane

  5. Supercomputer Output 148 TB Movie Output (0.25 TB/file) 80 TB Diagnostic Dumps (8 TB/file) Connected at 10Gbps Oak Ridge to ANL to SDSC Intergalactic Medium on 2 Billion Light Year Scale Exploring Cosmology With Supercomputers, Supernetworks, and Supervisualization Science: Norman, Harkness, Paschos SDSC Visualization: Insley, ANL; Wagner SDSC ANL * Calit2 * LBNL * NICS * ORNL * SDSC

  6. Providing End-to-End 10Gbps Cyberinfrastructure for Petascale End Users log of gas temperature log of gas density Mike Norman, SDSC Analyzing Super Data

  7. Calit2 Microbial Metagenomics Cluster-Next Generation Optically Linked Science Data Server Source: Phil Papadopoulos, SDSC, Calit2 512 Processors ~5 Teraflops ~ 200 Terabytes Storage Nearly 4000 Users Over 75 Countries

  8. Using 10 Gbps Big Data Access and Analysis-Collaboration Between Calit2 and U Washington Ginger Armbrust’s Diatom Chromosomes Photo Credit: Alan Decker Feb. 29, 2008 iHDTV: 1500 Mbits/sec Calit2 to UW Research Channel Over NLR

  9. MIT’s Ed DeLong & Darwin Project Team Using OptIPortal to Analyze 10km Coupled Ocean Microbial Simulation

  10. The NSF-Funded Ocean Observatory Initiative– a Complex System of Systems Cyberinfrastructure Source: Matthew Arrott, Calit2 Program Manager for OOI CI

  11. 1 cm. Taking Sensornets to the Ocean Floor:Remote Interactive HD Imaging of Deep Sea Vent Source: John Delaney and Research Channel, U Washington

  12. NSF OOI is a $400M Program -OOI CI is $34M Part of OOI Source: Matthew Arrott, Calit2 Program Manager for OOI CI 30-40 Software Engineers Housed at Calit2@UCSD

  13. OOI CI is Built on National LambdaRail’sand Internet2’s DCN Optical Infrastructure Source: John Orcutt, Matthew Arrott, SIO/Calit2

  14. High Definition Video Connected OptIPortals:Virtual Working Spaces for Data Intensive Research Source: Falko Kuester, Kai Doerr Calit2; Michael Sims, NASA

  15. Analyzing Big Data in 3D Stereo:The NexCAVE OptIPortal Source: Tom DeFanti, Calit2@UCSD

  16. “Blueprint for the Digital University”--Report of the UCSD Research Cyberinfrastructure Design Teamresearch.ucsd.edu/documents/rcidt/RCIDTReportFinal2009.pdf CENIC, NLR, I2DCN N x 10Gbe April 24, 2009 DataOasis(Central) Storage Gordon – HPC System Cluster Condo Triton – Petadata Analysis Scientific Instruments Digital Data Collections Campus Lab Cluster OptIPortal Source: Philip Papadopoulos, SDSC, UCSD

  17. California and Washington Universities Are Testing a 10Gbps Connected Commercial Data Cloud • Amazon Experiment for Big Data • Only Available Through CENIC and Pacific NW GigaPOP • Private 10Gbps Peering Path • Includes Amazon Computing and Storage Services

  18. You Can Download This Presentation at lsmarr.calit2.net

More Related