1 / 52

Algorithms and the Grid

Explore the role of algorithms in the development of the Grid, an integrated information technology infrastructure for managing and understanding large amounts of data. Discover key styles of Grids and their applications in various fields such as e-business, e-science, and aerospace engineering.

gcarignan
Download Presentation

Algorithms and the Grid

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Algorithms and the Grid Geoffrey FoxComputer Science, Informatics, Physics Pervasive Technology Laboratories Indiana University Bloomington IN 47401 March 18 2005 gcf@indiana.edu http://www.infomall.org

  2. Trends in Simulation Research • 1990-2000 the HPCC High Performance Computing and Communication Initiative • Established Parallel Computing • Developed wonderful algorithms – especially in partial differential equation and particle dynamics areas • Almost no useful software except for MPI – messaging between parallel computer nodes • 1995-now Internet explosion and development of Web Service distributed system model • Replaces CORBA, Java RMI, HLA, COM etc. • 2000- now: almost no USA academic work in core simulation • Major projects like ASCI (DoE) and HPCMO (DoD)thrive • 2003-? Data Deluge apparent and Grid links Internet and HPCC with focus on data-simulation integration

  3. e-Business e-Science and the Grid • e-Business captures an emerging view of corporations as dynamic virtual organizations linking employees, customers and stakeholders across the world. • e-Science is the similar vision for scientific research with international participation in large accelerators, satellites or distributed gene analyses. • The Grid or CyberInfrastructure integrates the best of the Web, Agents, traditional enterprise software, high performance computing and Peer-to-peer systems to provide the information technology e-infrastructure for e-moreorlessanything. • A deluge of data of unprecedented and inevitable size must be managed and understood. • People, computers, data and instruments must be linked. • On demand assignment of experts, computers, networks and storage resources must be supported

  4. Some Important Styles of Grids • Computational Grids were origin of concepts and link computers across the globe – high latency stops this from being used as parallel machine • Knowledge and Information Grids link sensors and information repositories as in Virtual Observatories or BioInformatics • More detail on next slide • Collaborative Grids link multidisciplinary researchers across laboratories and universities • Community Grids focus on Grids involving large numbers of peers rather than focusing on linking major resources – links Grid and Peer-to-peer network concepts • Semantic Grid links Grid, and AI community with Semantic web (ontology/meta-data enriched resources) and Agent concepts • Grid Service Farms supply services-on-demand as in collaboration, GIS support, Image processing filter

  5. Information/Knowledge Grids • Distributed (10’s to 1000’s) of data sources (instruments, file systems, curated databases …) • Data Deluge: 1 (now) to 100’s petabytes/year (2012) • Moore’s law for Sensors • Possible filters assigned dynamically (on-demand) • Run image processing algorithm on telescope image • Run Gene sequencing algorithm on compiled data • Needs decision support front end with “what-if” simulations • Metadata (provenance) critical to annotate data • Integrate across experiments as in multi-wavelength astronomy Data Deluge comes from pixels/year available

  6. Virtual Observatory Astronomy GridIntegrate Experiments Radio Far-Infrared Visible Dust Map Visible + X-ray Galaxy Density Map

  7. e-Business and (Virtual) Organizations • Enterprise Grid supports information system for an organization; includes “university computer center”, “(digital) library”, sales, marketing, manufacturing … • Outsourcing Grid links different parts of an enterprise together Manufacturing plants with designers • Animators with electronic game or film designers and producers • Coaches with aspiring players (e-NCAA or e-NFL etc.) • Outsourcing will become easier …….. • Customer Grid links businesses and their customers as in many web sites such as amazon.com • e-Multimedia can use secure peer-to-peer Grids to link creators, distributors and consumers of digital music, games and films respecting rights • Distance education Grid links teacher at one place, students all over the place, mentors and graders; shared curriculum, homework, live classes …

  8. DAME In flight data ~5000 engines ~ Gigabyte per aircraft per Engine per transatlantic flight Global Network Such as SITA Ground Station Airline Engine Health (Data) Center Maintenance Centre Internet, e-mail, pager Rolls Royce and UK e-Science ProgramDistributed Aircraft Maintenance Environment

  9. It takes a distributed virtual organization to design, simulate and build a complex system like an aircraft NASA Aerospace Engineering Grid

  10. e-Defense and e-Crisis • Grids support Command and Control and provide Global Situational Awareness • Link commanders and frontline troops to themselves and to archival and real-time data; link to what-if simulations • Dynamic heterogeneous wired and wireless networks • Security and fault tolerance essential • System of Systems; Grid of Grids • The command and information infrastructure of each ship is a Grid; each fleet is linked together by a Grid; the President is informed by and informs the national defense Grid • Grids must be heterogeneous and federated • Crisis Management and Response enabled by a Grid linking sensors, disaster managers, and first responders with decision support • Define and Build DoD relevant Services – Collaboration, Sensors, GIS, Database etc.

  11. Analysis and Visualization Large Disks Old Style Metacomputing Grid Large Scale Parallel Computers Spread a single large Problem over multiple supercomputers

  12. Classes of Computing Grid Applications • Running “Pleasing Parallel Jobs” as in United Devices, Entropia (Desktop Grid) “cycle stealing systems” • Can be managed (“inside” the enterprise as in Condor) or more informal (as in SETI@Home) • Computing-on-demand in Industry where jobs spawned are perhaps very large (SAP, Oracle …) • Support distributed file systems as in Legion (Avaki), Globus with (web-enhanced) UNIX programming paradigm • Particle Physics will run some 30,000 simultaneous jobs this way • Pipelinedapplications linking data/instruments, compute, visualization • Seamless Access where Grid portals allow one to choose oneof multiple resources with a common interfaces

  13. What is Happening? • Grid ideas are being developed in (at least) two communities • Web Service – W3C, OASIS • Grid Forum (High Performance Computing, e-Science) • Open Middleware Infrastructure Institute OMII currently only in UK but maybe spreads to EU and USA • Service Standards are being debated • Grid Operational Infrastructure is being deployed • Grid Architecture and core software being developed • Particular System Services are being developed “centrally” – OGSA framework for this in • Lots of fields are setting domain specific standards and building domain specific services • Grids are viewed differently in different areas • Largely “computing-on-demand” in industry (IBM, Oracle, HP, Sun) • Largely distributed collaboratories in academia

  14. PortalService Security Catalog A typical Web Service • In principle, services can be in any language (Fortran .. Java .. Perl .. Python) and the interfaces can be method calls, Java RMI Messages, CGI Web invocations, totally compiled away (inlining) • The simplest implementations involve XML messages (SOAP) and programs written in net friendly languages like Java and Python PaymentCredit Card Web Services WSDL interfaces Warehouse Shipping control WSDL interfaces Web Services

  15. Services and Distributed Objects • A web service is a computer program running on either the local or remote machine with a set of well defined interfaces (ports) specified in XML (WSDL) • Web Services (WS) have many similarities with Distributed Object (DO) technology but there are some (important) technical and religious points (not easy to distinguish) • CORBA Java COM are typical DO technologies • Agents are typically SOA (Service Oriented Architecture) • Both involve distributed entities but Web Services are more loosely coupled • WS interact with messages; DO with RPC (Remote Procedure Call) • DO have “factories”; WS manage instances internally and interaction-specific state not exposed and hence need not be managed • DO have explicit state (statefull services); WS use context in the messages to link interactions (statefull interactions) • Claim: DO’s do NOT scale; WS build on experience (with CORBA) and do scale

  16. Grid impact on Algorithms I • Your favorite parallel algorithm will often run untouched on a Grid node linked to other simulations using traditional algorithms • Algorithms tolerant of high latency • Algorithms for new applications enabled by the Grid • Data assimilation for data-deluged science generalizing data mining • Where and how to process data • Incorporation of data in simulation • Complex Systems algorithms for non traditional simulations as in biology, social systems • Cellular automata

  17. Grid impact on Algorithms II • MPI software model not suited for Grid; use SOAP and publish/subscribe • Microseconds and milliseconds Latency • Grid workflow needs “integration algorithms” • Multidisciplinary algorithms for loose code coupling • Workflow scheduling algorithms (data oriented) • Data caching algorithms • Algorithms like distributed hash tables for distributed storage and look up of data • Algorithms for Grid security • Efficient support of group keys for multicast • Detection of Denial of Service attacks • Much better software available for building toolkits and Problem Solving Environments i.e. for using algorithms

  18. Data Deluged Science • In the past, we worried about data in the form of parallel I/O or MPI-IO, but we didn’t consider it as an enabler of new algorithms and new ways of computing • Data assimilation was not central to HPCC • DoE ASCI set up because didn’t want test data! • Now particle physics will get 100 petabytes from CERN • Nuclear physics (Jefferson Lab) in same situation • Use around 30,000 CPU’s simultaneously 24X7 • Weather, climate, solid earth (EarthScope) • Bioinformatics curated databases (Biocomplexity only 1000’s of data points at present) • Virtual Observatory and SkyServer in Astronomy • Environmental Sensor nets

  19. Weather Requirements

  20. Data Assimilation Information Simulation Model Datamining Ideas Reasoning ComputationalScience Data DelugedScienceComputing Paradigm Informatics

  21. USArray Seismic Sensors

  22. a Site-specific Irregular Scalar Measurements a Constellations for Plate Boundary-Scale Vector Measurements Ice Sheets a Volcanoes PBO Greenland Long Valley, CA Topography 1 km Stress Change Northridge, CA Earthquakes Hector Mine, CA

  23. Field Trip Data ? Discovery Services RepositoriesFederated Databases Streaming Data Sensors Database Database Research Education SERVOGrid Data FilterServices Customization Services From Researchto Education ResearchSimulations Analysis and VisualizationPortal EducationGrid Computer Farm Geoscience Research and Education Grids

  24. SERVOGrid Requirements • Seamless Access to Data repositories and large scale computers • Integration of multiple data sources including sensors, databases, file systems with analysis system • Including filtered OGSA-DAI (Grid database access) • Rich meta-data generation and access with SERVOGrid specific Schema extending openGIS (Geography as a Web service) standards and using Semantic Grid • Portalswith component model for user interfaces and web control of all capabilities • Collaboration to support world-wide work • Basic Grid tools: workflow and notification • NOT metacomputing

  25. Data Data Filter Filter Filter Data Filter Data OGSA-DAIGrid Services AnalysisControl Visualize Grid Data Filter Data Deluged ScienceComputing Architecture HPC Simulation Grid Data Assimilation Other Gridand Web Services Distributed Filters massage data For simulation

  26. Data Assimilation • Data assimilation implies one is solving some optimization problem which might have Kalman Filter like structure • Due to data deluge, one will become more and more dominated by the data (Nobs much larger than number of simulation points). • Natural approach is to form for each local (position, time) patch the “important” data combinations so that optimization doesn’t waste time on large error or insensitive data. • Data reduction done in natural distributed fashion NOT on HPC machine as distributed computing most cost effective if calculations essentially independent • Filter functions must be transmitted from HPC machine

  27. Distributed Filtering Nobslocal patch >> Nfilteredlocalpatch≈ Number_of_Unknownslocalpatch In simplest approach, filtered data gotten by linear transformations on original data based on Singular Value Decomposition of Least squares matrix Send needed Filter Receive filtered data Nobslocal patch 1 Filter Data Nfilteredlocal patch 1 Geographically DistributedSensor patches Nobslocal patch 2 Filter Data Nfilteredlocal patch 2 HPC Machine Factorize Matrixto product of local patches Distributed Machine

  28. Non Traditional Applications: Critical Infrastructure Simulations • These include electrical/gas/water grids and Internet, transportation, cell/wired phone dynamics. • One has some “classic SPICE style” network simulations in area like power grid (although load and infrastructure data incomplete) • 6000 to 17000 generators • 50000 to 140000 transmission lines • 40000 to 100000 substations • Need algorithms both forsimulating infrastructuresbut also to link them

  29. Non Traditional Applications: Critical Infrastructure Simulations • Activity data for people/institutions essential for detailed dynamics but again these are not “classic” data but need to be “fitted” in data assimilation style in terms of some assumed lower level model. • They tell you goals of people but not their low level movement • Disease and Internet virus spread and social network simulations can be built on dynamics coming from infrastructure simulations • Many results like “small world” internet connection structure are qualitative and unclear if they can be extended to detailed simulations • A lot of interest in (regulatory) networks in Biology

  30. (Non) Traditional Structure • 1) Traditional: Known equations plus boundary values • 2) Data assimilation: somewhat uncertain initial conditions and approximations corrected by data assimilation • 3) Data deluged Science: Phenomenological degrees of freedom swimming in a sea of data Known Data Known Equations on Agreed DoF Prediction PhenomenologicalDegrees of Freedom Swimming in a Sea of Data

  31. Some Questions for Non Traditional Applications • No systematic study of how best to represent data deluged sciences without known equations • Obviously data assimilation very relevant • Role of Cellular Automata (CA) and refinements like the New Kind of Science by Wolfram • Can CA or Potts model parameterize any system? • Relationship to back propagation and other neural network representations • Relationship to “just” interpolating data and then extrapolating a little • Role of Uncertainty Analysis – everything (equations, model, data) is uncertain! • Relationship of data mining and simulation • A new trade-off: How to split funds between sensors and simulation engines

  32. When is a High Performance Computer? • We might wish to consider three classes of multi-node computers • 1) Classic MPP with microsecond latency and scalable internode bandwidth (tcomm/tcalc ~ 10 or so) • 2) Classic Cluster which can vary from configurations like 1) to 3) but typically have millisecond latency and modest bandwidth • 3) Classic Grid or distributed systems of computers around the network • Latencies of inter-node communication – 100’s of milliseconds but can have good bandwidth • All have same peak CPU performance but synchronization costs increase as one goes from 1) to 3) • Cost of system (dollars per gigaflop) decreases by factors of 2 at each step from 1) to 2) to 3) • One should NOT use classic MPP if class 2) or 3) suffices unless some security or data issues dominates over cost-performance • One should not use a Grid as a true parallel computer – it can link parallel computers together for convenient access etc.

  33. Nugget3 Nugget4 Interaction Nugget1 Nugget2 Data Building PSE’s with theRule of the Millisecond I • Typical Web Services are used in situations with interaction delays (network transit) of 100’s of milliseconds • But basic message-based interaction architecture only incurs fraction of a millisecond delay • Thus use Web Services to build ALL PSE components • Use messages and NOT method/subroutine call or RPC

  34. Building PSE’s with theRule of the Millisecond II • Messaging has several advantages over scripting languages • Collaboration trivial by sharing messages • Software Engineering due to greater modularity • Web Services do/will have wonderful support • “Loose” Application coupling uses workflow technologies • Find characteristic interaction time (millisecond programs; microseconds MPI and particle) and use best supported architecture at this level • Two levels: Web Service (Grid) and C/C++/C#/Fortran/Java/Python • Major difficulty in frameworks is NOT building them but rather in supporting them • IMHO only hope is to always minimize life-cycle support risks • Simulation/science is too small a field to support much! • Expect to use DIFFERENT technologies at each level even though possible to do everything with one technology • Trade off support versus performance/customization

  35. Requirements for MPI Messaging tcalc tcomm tcalc • MPI and SOAP Messaging both send data from a source to a destination • MPI supports multicast (broadcast) communication; • MPI specifies destination and a context (in comm parameter) • MPI specifies data to send • MPI has a tag to allow flexibility in processing in source processor • MPI has calls to understand context (number of processors etc.) • MPI requires very low latency and high bandwidth so that tcomm/tcalc is at most 10 • BlueGene/L has bandwidth between 0.25 and 3 Gigabytes/sec/node and latency of about 5 microseconds • Latency determined so Message Size/Bandwidth > Latency

  36. Requirements for SOAP Messaging • Web Services has much of the same requirements as MPI with two differences where MPI more stringent than SOAP • Latencies are inevitably 1 (local) to 100 milliseconds which is 200 to 20,000 times that of BlueGene/L • 1) 0.000001 ms – CPU does a calculation • 2) 0.001 to 0.01 ms – MPI latency • 3) 1 to 10 ms – wake-up a thread or process • 4) 10 to 1000 ms – Internet delay • Bandwidths for many business applications are low as one just needs to send enough information for ATM and Bank to define transactions • SOAP has MUCH greater flexibility in areas like security, fault-tolerance, “virtualizing addressing” because one can run a lot of software in 100 milliseconds • Typically takes 1-3 milliseconds to gobble up a modest message in Java and “add value”

  37. Structure of SOAP • SOAP defines a very obvious message structure with a header and a bodyjust like email • The header contains information used by the “Internet operating system” • Destination, Source, Routing, Context, Sequence Number … • The message body is partly further information used by the operating system and partly information for application when it is not looked at by “operating system” except to encrypt, compress it etc. • Note WS-Security supports separate encryption for different parts of a document • Much discussion in field revolves around what is referenced in header • This structure makes it possible to define VERY Sophisticated messaging

  38. MPI and SOAP Integration • Note SOAP Specifies format and through WSDL interfaces • MPI only specifies interface and so interoperability between different MPIs requires additional work • IMPI http://impi.nist.gov/IMPI/ • Pervasive networks can support high bandwidth (Terabits/sec soon) but latency issue is not resolvable in general way • Can combine MPI interfaces with SOAP messaging but I don’t think this has been done • Just as walking, cars, planes, phones coexist with different properties; so SOAP and MPI are both good and should be used where appropriate

  39. NaradaBrokering • http://www.naradabrokering.org • We have built a messaging system that is designed to support traditional Web Services but has an architecture that allows it to support high performance data transport as required for Scientific applications • We suggest using this system whenever your application can tolerate 1-10 millisecond latency in linking components • Use MPI when you need much lower latency • Use SOAP approach when MPI interfaces required but latency high • As in linking two parallel applications at remote sites • Technically it forms an overlay network supporting in software features often done at IP Level

  40. Mean transit delay for message samples in NaradaBrokering: Different communication hops 9 hop-2 hop-3 8 hop-5 7 hop-7 6 5 Transit Delay (Milliseconds) 4 3 2 1 0 100 1000 Message Payload Size (Bytes) Pentium-3, 1GHz, 256 MB RAM 100 Mbps LAN JRE 1.3 Linux

  41. Pure SOAP SOAP over UDP Binary over UDP Fast Web Service Communication I • Internet Messaging systems allow one to optimize message streams at the cost of “startup time”, • Web Services can deliver the fastest possible interconnections with or without reliable messaging • Typical results from Grossman (UIC) comparing Slow SOAP over TCP with binary and UDP transport (latter gains a factor of 1000) 7020 5.60

  42. Fast Web Service Communication II • Mechanism only works for streams – sets of related messages • SOAP header in streams is constant except for sequence number (Message ID), time-stamp .. • One needs two types of new Web Service Specification • “WS-StreamNegotiation” to define how one can use WS-Policy to send messages at start of a stream to define the methodology for treating remaining messages in stream • “WS-FlexibleRepresentation” to define new encodings of messages

  43. Fast Web Service Communication III • Then use “WS-StreamNegotiation” to negotiate stream in Tortoise SOAP – ASCII XML over HTTP and TCP – • Deposit basic SOAP header through connection – it is part of context for stream (linking of 2 services) • Agree on firewall penetration, reliability mechanism, binary representation and fast transport protocol • Naturally transport UDP plus WS-RM • Use “WS-FlexibleRepresentation” to define encoding of a Fast transport (On a different port) with messages just having “FlexibleRepresentationContextToken”, Sequence Number, Time stamp if needed • RTP packets have essentially this structure • Could add stream termination status • Can monitor and control with original negotiation stream • Can generate different streams optimized for different end-points

  44. Service-1 Service-3 Role of Workflow • Programming SOAP and Web Services (the Grid): Workflow describes linkage between services • As distributed, linkage must be by messages • Linkage is two-way and has both control and data • Apply to multi-disciplinary, multi-scale linkage, multi-program linkage, link visualization to simulation, GIS to simulations and visualization filters to each other • Microsoft-IBM specification BPEL is current preferred Web Service XML specification of workflow Service-2

  45. Example workflow Here a sensor feeds a data-mining application (We are extending data-mining in DoD applications with Grossman from UIC) The data-mining application drives a visualization

  46. Example Flood Simulation workflow

  47. SERVOGrid Codes, Relationships Elastic Dislocation Inversion Viscoelastic FEM Viscoelastic Layered BEM Elastic Dislocation Pattern Recognizers Fault Model BEM This linkage called Workflow in Grid/Web Service parlance

  48. Two-level Programming I Service Data • The Web Service (Grid) paradigm implicitly assumes a two-level Programming Model • We make a Service (same as a “distributed object” or “computer program” running on a remote computer) using conventional technologies • C++ Java or Fortran Monte Carlo module • Data streaming from a sensor or Satellite • Specialized (JDBC) database access • Such services accept and produce data from users files and databases • The Grid is built by coordinating such services assuming we have solved problem of programming the service

  49. Service1 Service3 Service2 Service4 Two-level Programming II • The Grid is discussing the composition of distributed serviceswith the runtime interfaces to Grid as opposed to UNIX pipes/data streams • Familiar from use of UNIX Shell, PERL or Python scripts to produce real applications from core programs • Such interpretative environments are the single processor analog of Grid Programming • Some projects like GrADS from Rice University are looking at integration between service and composition levels but dominant effort looks at each level separately

More Related