1 / 48

CSCI-455/522

CSCI-455/522. Introduction to High Performance Computing Lecture 2. Types of Parallel Computers. Two principal types: Shared memory multiprocessor Distributed memory multicomputer. Shared vs. Distributed Memory. P. P. P. P. P. P.

prisca
Download Presentation

CSCI-455/522

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CSCI-455/522 Introduction to High Performance Computing Lecture 2

  2. Types of Parallel Computers Two principal types: • Shared memory multiprocessor • Distributed memory multicomputer

  3. Shared vs. Distributed Memory P P P P P P Shared memory - single address space. All processors have access to a pool of shared memory. (Ex: SGI Origin, Sun E10000) BUS Memory P P P P P P Distributed memory - each processor has it’s own local memory. Must do message passing to exchange data between processors. (Ex: CRAY T3E, IBM SP, clusters) M M M M M M Network

  4. Shared Memory Multiprocessor

  5. Conventional Computer Consists of a processor executing a program stored in a (main) memory: Each main memory location located by its address. Addresses start at 0 and extend to 2b - 1 when there are b bits (binary digits) in address. Main memory Instr uctions (to processor) Data (to or from processor) Processor

  6. Shared Memory Multiprocessor System Natural way to extend single processor model - have multiple processors connected to multiple memory modules, such that each processor can access any memory module : Memory module One address space Interconnection network Processors

  7. Simplistic View of a Small Shared Memory Multiprocessor Examples: • Dual Pentiums • Quad Pentiums Processors Shared memory Bus

  8. Quad Pentium Shared Memory Multiprocessor Processor Processor Processor Processor L1 cache L1 cache L1 cache L1 cache L2 Cache L2 Cache L2 Cache L2 Cache Bus interface Bus interface Bus interface Bus interface Processor/ memory b us I/O interf ace Memory controller I/O b us Memory Shared memory

  9. Programming Shared Memory Multiprocessors • Threads - programmer decomposes program into individual parallel sequences, (threads), each being able to access variables declared outside threads. Example Pthreads • Sequential programming language with preprocessor compiler directives to declare shared variables and specify parallelism. Example OpenMP - industry standard - needs OpenMP compiler

  10. Sequential programming language with added syntax to declare shared variables and specify parallelism. Example UPC (Unified Parallel C) - needs a UPC compiler. • Parallel programming language with syntax to express parallelism - compiler creates executable code for each processor (not now common) • Sequential programming language and ask parallelizing compiler to convert it into parallel executable code. - also not now common

  11. Shared Memory: UMA vs. NUMA Uniform memory access (UMA): Each processor has uniform access to memory. Also known as symmetric multiprocessors, or SMPs (Sun E10000) P P P P P P BUS Memory P P P P P P P P Non-uniform memory access (NUMA): Time for memory access depends on location of data. Local access is faster than non-local access. Easier to scale than SMPs (SGI Origin) BUS BUS Memory Memory Network

  12. Distributed Memory /Message-Passing Multicomputers

  13. Message-Passing Multicomputer Complete computers connected through an interconnection network: Interconnection network Messages Processor Local memory Computers

  14. Distributed Memory: MPPs vs. Clusters • Processor-memory nodes are connected by some type of interconnect network • Massively Parallel Processor (MPP): tightly integrated, single system image. • Cluster: individual computers connected by s/w Interconnect Network CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM

  15. Distributed Memory: MPPs vs. Clusters • Processor-memory nodes are connected by some type of interconnect network • Massively Parallel Processor (MPP): tightly integrated, single system image. • Cluster: individual computers connected by s/w Interconnect Network CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM CPU MEM

  16. Clusters • Similar to MPPs • Commodity processors and memory • Processor performance must be maximized • Memory hierarchy includes remote memory • No shared memory--message passing • Communication overhead must be minimized • Different from MPPs • All commodity, including interconnect and OS • Multiple independent systems: more robust • Separate I/O systems

  17. Interconnection Networks • Limited and exhaustive interconnections • 2- and 3-dimensional meshes • Hypercube (not now common) • Using Switches: • Crossbar • Trees • Multistage interconnection networks

  18. Communications Networks • Custom • Many vendors have custom interconnects that provide high performance for their system, specially MPP • CRAY T3E interconnect is the fastest for MPPs: lowest latency, highest bandwidth • Commodity • Used in some MPPs and all clusters • Myrinet, Gigabit Ethernet, Fast Ethernet, etc.

  19. Types of Interconnects • Fully connected • not feasible • Array and torus • Intel Paragon (2D array), CRAY T3E (3D torus) • Crossbar • IBM SP (8 nodes) • Hypercube • SGI Origin 2000 (hypercube), Meiko CS-2 (fat tree) • Combinations of some of the above • IBM SP (crossbar & fully connected for 80 nodes) • IBM SP (fat tree for > 80 nodes)

  20. Two-dimensional Array (Mesh) Computer/ Links processor Also three-dimensional - used in some large high performance systems.

  21. Three-dimensional Hypercube

  22. Four-dimensional Hypercube Hypercubes popular in 1980’s - not now

  23. Crossbar Switch Memor ies Switches Processors

  24. Tree Root Switch Links element Processors

  25. Multistage Interconnection NetworkExample: Omega network 2 ´ 2 switch elements (straight-through or crossover connections) 000 000 001 001 010 010 011 011 Inputs Outputs 100 100 101 101 110 110 111 111

  26. Distributed Shared Memory Making main memory of group of interconnected computers look as though a single memory with single address space. Then can use shared memory programming techniques. Interconnection netw or k Messages Processor Shared memory Computers

  27. Flynn’s Classifications Flynn (1966) created a classification for computers based upon instruction streams and data streams: • Single instruction stream-single data stream (SISD) computer Single processor computer - single stream of instructions generated from program. Instructions operate upon a single stream of data items.

  28. Single Instruction Stream-Multiple Data Stream (SIMD) Computer • A specially designed computer - a single instruction stream from a single program, but multiple data streams exist. Instructions from program broadcast to more than one processor. Each processor executes same instruction in synchronism, but using different data. • Developed because a number of important applications that mostly operate upon arrays of data.

  29. Multiple Instruction Stream-Multiple Data Stream (MIMD) Computer General-purpose multiprocessor system - each processor has a separate program and one instruction stream is generated from each program for each processor. Each instruction operates upon different data. Both the shared memory and the message-passing multiprocessors so far described are in the MIMD classification.

  30. The Banking Analogy • Tellers: Parallel Processors • Customers: tasks • Transactions: operations • Accounts: data

  31. Vector/Array • Each teller/processor gets a very fine-grained task • Use pipeline parallelism • Good for handling batches when operations can be broken down into fine-grained stages

  32. SIMD (Single-Instruction-Multiple-Data) • All processors do the same things or idle • Phase 1: data partitioning and distributed • Phase 2: data-parallel processing • Efficient for big, regular data-sets

  33. Systolic Array • Combination of SIMD and Pipeline parallelism • 2-d array of processors with memory at the boundary • Tighter coordination between processors • Achieve very high speeds by circulating data among processors before returning to memory

  34. MIMD(Multi-Instruction-Multiple-Data) • Each processor (teller) operates independently • Need synchronization mechanism • by message passing • or mutual exclusion (locks) • Best suited for large-grained problems • Less than data-flow parallelism

  35. Networked Computers as a Computing Platform • A network of computers became a very attractive alternative to expensive supercomputers and parallel computer systems for high-performance computing in early 1990’s. • Several early projects. Notable: • Berkeley NOW (network of workstations) project. • NASA Beowulf project.

  36. Key advantages: • Very high performance workstations and PCs readily available at low cost. • The latest processors can easily be incorporated into the system as they become available. • Existing software can be used or modified.

  37. Software Tools for Clusters • Based upon Message Passing Parallel Programming: • Parallel Virtual Machine (PVM) - developed in late 1980’s. Became very popular. • Message-Passing Interface (MPI) - standard defined in 1990s. • Both provide a set of user-level libraries for message passing. Use with regular programming languages (C, C++, ...).

  38. Beowulf Clusters* • A group of interconnected “commodity” computers achieving high performance with low cost. • Typically using commodity interconnects - high speed Ethernet, and Linux OS. * Beowulf comes from name given by NASA Goddard Space Flight Center cluster project.

  39. Cluster Interconnects • Originally fast Ethernet on low cost clusters • Gigabit Ethernet - easy upgrade path More Specialized/Higher Performance • Myrinet - 2.4 Gbits/sec - disadvantage: single vendor • cLan • SCI (Scalable Coherent Interface) • QNet • Infiniband - may be important as infininband interfaces may be integrated on next generation PCs

  40. Dedicated cluster with a master node Dedicated Cluster User Compute nodes Master node Up link Exter nal netw or k 2nd Ether net Switch interf ace

More Related