hpc across heterogeneous resources n.
Skip this Video
Loading SlideShow in 5 Seconds..
HPC across Heterogeneous Resources PowerPoint Presentation
Download Presentation
HPC across Heterogeneous Resources

Loading in 2 Seconds...

play fullscreen
1 / 12

HPC across Heterogeneous Resources - PowerPoint PPT Presentation

  • Uploaded on

HPC across Heterogeneous Resources. Sathish Vadhiyar. Motivation. MPI assumes global communicator with the help of which all processes can communicate with each other. Hence all nodes on which MPI application is started must be accessible by all nodes. This is not always possible

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
Download Presentation

HPC across Heterogeneous Resources

An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript
  • MPI assumes global communicator with the help of which all processes can communicate with each other.
  • Hence all nodes on which MPI application is started must be accessible by all nodes.
  • This is not always possible
    • Due to the shortage of IP address space.
    • Security concerns – Large MPP sites and beowulf clusters have only 1 master node in public IP address space.
  • Grand challenging applications require the use of many MPPs.
pacx mpi
  • Parallel Computer eXtension
  • MPI on a cluster of MPPs
  • Initially developed to connect Cray-YMP with Intel Paragon
pacx mpi1
  • CFDs or crash simulation of automobiles – 1 MPP not enough
  • Initial application – flow solver application across Pittsburgh supercomputing centre, Sandia national lab and High Performance Computing Center, Stuttgart
  • PACX sits between applications and MPI;
pacx mpi2
  • On each MPP, 2 extra nodes with running daemons to take care of communication between MPPs, compression and decompression of data, communication with local nodes
  • Daemon nodes implemented as additional local MPI processes
  • Communication among processes internal to MPP through vendor MPI on a local MPP network.
  • Communication between MPPs through daemons via Internet or specialized network.
data conversion
Data Conversion
  • If the sender and receiver are in two separate heterogeneous MPPs, the sender converts its data to XDR (external data representation) format
  • Receiver converts data from XDR format to its own data representation
pacx mpi results
PACX-MPI: Results
  • Between T3Es at PSC and SDSC T3E
  • URANUS application
    • Navier-Strokes application – an iterative application based on convergence
    • Simulation of rentry vehicle
    • Between PSC and Univ. of Stuttgart
    • Closely coupled application due to frequent communication due to convergence.
    • Had to be modified for metacomputing setting – application made more asynchronous compromising on convergence
  • P3TDSMC – Monte Carlo for particle tracking
    • More amenable to metacomputing because of high computation-communication ratio
    • The latency effects are hidden as larger number of particles are considered.
other related projects
Other Related Projects
  • PLUS
  • MPI-Connect
  • Edgar Gabriel, Michael Resch, Thomas Beisel, Rainer Keller: 'Distributed computing in a heterogenous computing environment', (gzipped postscript) to appear at EuroPVMMPI'98 Liverpool/UK, 1998.
  • Thomas Beisel, Edgar Gabriel, Michael Resch: 'An Extension to MPI for Distributed Computing on MPPs' (gzipped postscript) in Marian Bubak, Jack Dongarra, Jerzy Wasniewski (Eds.) 'Recent Advances in Parallel Virtual Machine and Message Passing Interface', Lecture Notes in Computer Science, Springer, 1997, 75-83.
  • Message-passing environments for metacomputing , Pages 699-712 , Matthias A. Brune, Graham E. Fagg and Michael M. Resch , FGCS, Volume 15, 1999
  • PVMPI: An Integration of PVM and MPI systems. http://www.netlib.org/utk/papers/pvmpi/paper.html
  • A Grid-Enabled MPI: Message Passing in Heterogeneous Distributed Computing Systems. I. Foster, N. Karonis. Proc. 1998 SC Conference, November, 1998