1 / 70

Distributed Systems CS 15-440

Distributed Systems CS 15-440. Programming Models Gregory Kesden Borrowed and adapted from our good friends at CMU-Doha , Qatar Majd F. Sakr , Mohammad Hammoud andVinay Kolar. Objectives. Discussion on Programming Models. MapReduce. Message Passing Interface (MPI).

gyula
Download Presentation

Distributed Systems CS 15-440

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Distributed SystemsCS 15-440 Programming Models Gregory Kesden Borrowed and adapted from our good friends at CMU-Doha, Qatar Majd F. Sakr, Mohammad HammoudandVinayKolar

  2. Objectives Discussion on Programming Models MapReduce Message Passing Interface (MPI) Examples of parallel processing Traditional models of parallel programming Parallel computer architectures Why parallelism? Why parallelism?

  3. Amdahl’s Law • We parallelize our programs in order to run them faster • How much faster will a parallel program run? • Suppose that the sequential execution of a program takes T1time units and the parallel execution on p processors takes Tp time units • Suppose that out of the entire execution of the program, s fraction of it is not parallelizable while 1-s fraction is parallelizable • Then the speedup (Amdahl’s formula):

  4. Amdahl’s Law: An Example • Suppose that 80% of you program can be parallelized and that you use 4 processors to run your parallel version of the program • The speedup you can get according to Amdahl is: • Although you use 4 processors you cannot get a speedup more than 2.5 times (or 40% of the serial running time)

  5. Real Vs. Actual Cases • Amdahl’s argument is too simplified to be applied to real cases • When we run a parallel program, there are a communication overhead and a workload imbalance among processes in general 20 80 20 80 Serial Serial Parallel 20 20 Parallel 20 20 Process 1 Process 1 Process 2 Process 2 Cannot be parallelized Process 3 Process 3 Cannot be parallelized Can be parallelized Communication overhead Process 4 Process 4 Can be parallelized Load Unbalance 2. Parallel Speed-up: An Actual Case 1. Parallel Speed-up: An Ideal Case

  6. Guidelines • In order to efficiently benefit from parallelization, we ought to follow these guidelines: • Maximize the fraction of our program that can be parallelized • Balance the workload of parallel processes • Minimize the time spent for communication

  7. Objectives Discussion on Programming Models MapReduce Message Passing Interface (MPI) Examples of parallel processing Traditional models of parallel programming Parallel computer architectures Parallel computer architectures Why parallelism?

  8. Parallel Computer Architectures • We can categorize the architecture of parallel computers in terms of two aspects: • Whether the memory is physically centralized or distributed • Whether or not the address space is shared

  9. Symmetric Multiprocessor • Symmetric Multiprocessor (SMP) architecture uses shared system resources that can be accessed equally from all processors • A single OS controls the SMP machine and it schedules processes and threads on processors so that the load is balanced Processor Processor Processor Processor Cache Cache Cache Cache Bus or Crossbar Switch Memory I/O

  10. Massively Parallel Processors • Massively Parallel Processors (MPP) architecture consists of nodes with each having its own processor, memory and I/O subsystem • An independent OS runs at each node Interconnection Network Processor Processor Processor Processor Cache Cache Cache Cache Bus Bus Bus Bus Memory I/O Memory I/O Memory I/O Memory I/O

  11. Non-Uniform Memory Access • Non-Uniform Memory Access (NUMA) architecture machines are built on a similar hardware model as MPP • NUMA typically provides a shared address space to applications using a hardware/software directory-based coherence protocol • The memory latency varies according to whether you access memory directly (local) or through the interconnect (remote). Thus the name non-uniform memory access • As in an SMP machine, a single OS controls the whole system

  12. Objectives Discussion on Programming Models MapReduce Message Passing Interface (MPI) Examples of parallel processing Traditional Models of parallel programming Traditional Models of parallel programming Parallel computer architectures Why parallelizing our programs?

  13. Models of Parallel Programming • What is a parallel programming model? • A programming model is an abstraction provided by the hardware to programmers • It determines how easily programmers can specify their algorithms into parallel unit of computations (i.e., tasks) that the hardware understands • It determines how efficiently parallel tasks can be executed on the hardware • Main Goal: utilize all the processors of the underlying architecture (e.g., SMP, MPP, NUMA) and minimize the elapsed time of your program

  14. Traditional Parallel Programming Models Parallel Programming Models Shared Memory Message Passing Message Passing

  15. Shared Memory Model • In the shared memory programming model, the abstraction is that parallel tasks can access any location of the memory • Parallel tasks can communicate through reading and writing common memory locations • This is similar to threads from a single process which share a single address space • Multi-threaded programs (e.g., OpenMP programs)are the best fit with shared memory programming model

  16. Shared Memory Model Si = Serial Pj = Parallel Single Thread Multi-Thread S1 S1 Time Time Spawn P1 P3 P1 P2 P3 P2 Join P3 S2 Shared Address Space P4 S2 Process Process

  17. Shared Memory Example begin parallel // spawn a child thread private int start_iter, end_iter, i; shared int local_iter=4, sum=0; shared double sum=0.0, a[], b[], c[]; shared lock_type mylock; start_iter = getid() * local_iter; end_iter = start_iter + local_iter; for (i=start_iter; i<end_iter; i++) a[i] = b[i] + c[i]; barrier; for (i=start_iter; i<end_iter; i++) if (a[i] > 0) { lock(mylock); sum = sum + a[i]; unlock(mylock); } barrier; // necessary end parallel // kill the child thread Print sum; for (i=0; i<8; i++) a[i] = b[i] + c[i]; sum = 0; for (i=0; i<8; i++) if (a[i] > 0) sum = sum + a[i]; Print sum; Sequential Parallel

  18. Traditional Parallel Programming Models Parallel Programming Models Shared Memory Shared Memory Message Passing

  19. Message Passing Model • In message passing, parallel tasks have their own local memories • One task cannot access another task’s memory • Hence, to communicate data they have to rely on explicit messages sent to each other • This is similar to the abstraction of processes which do not share an address space • MPI programs are the best fit with message passing programming model

  20. Message Passing Model Message Passing Single Thread S = Serial P = Parallel S1 S1 S1 S1 S1 Time Time P1 P1 P1 P1 P1 P2 S2 S2 S2 S2 P3 P4 Process 0 Process 1 Process 2 Process 3 S2 Node 1 Node 2 Node 3 Node 4 Data transmission over the Network Process

  21. Message Passing Example id = getpid(); local_iter = 4; start_iter = id * local_iter; end_iter = start_iter + local_iter; if (id == 0) send_msg (P1, b[4..7], c[4..7]); else recv_msg (P0, b[4..7], c[4..7]); for (i=start_iter; i<end_iter; i++) a[i] = b[i] + c[i]; local_sum = 0; for (i=start_iter; i<end_iter; i++) if (a[i] > 0) local_sum = local_sum + a[i]; if (id == 0) { recv_msg (P1, &local_sum1); sum = local_sum + local_sum1; Print sum; } else send_msg (P0, local_sum); for (i=0; i<8; i++) a[i] = b[i] + c[i]; sum = 0; for (i=0; i<8; i++) if (a[i] > 0) sum = sum + a[i]; Print sum; Sequential Parallel

  22. Shared Memory Vs. Message Passing • Comparison between shared memory and message passing programming models:

  23. Objectives Discussion on Programming Models MapReduce Message Passing Interface (MPI) Examples of parallel processing Examples of parallel processing Traditional Models of parallel programming Parallel computer architectures Why parallelizing our programs?

  24. SPMD and MPMD • When we run multiple processes with message-passing, there are further categorizations regarding how many different programs are cooperating in parallel execution • We distinguish between two models: • Single Program Multiple Data (SPMD) model • Multiple Programs Multiple Data (MPMP) model

  25. SPMD • In the SPMD model, there is only one program and each process uses the same executable working on different sets of data a.out Node 1 Node 3 Node 2

  26. MPMD • The MPMD model uses different programs for different processes, but the processes collaborate to solve the same problem • MPMD has two styles, the master/worker and the coupled analysis a.out a.out b.out b.out c.out a.out= Structural Analysis, b.out = fluid analysis and c.out = thermal analysis Example Node 1 Node 1 Node 2 Node 2 Node 3 Node 3 1. MPMD: Master/Slave 2. MPMD: Coupled Analysis

  27. 3 Key Points • To summarize, keep the following 3 points in mind: • The purpose of parallelization is to reduce the time spent for computation • Ideally, the parallel program is p times faster than the sequential program, where p is the number of processes involved in the parallel execution, but this is not always achievable • Message-passing is the tool to consolidate what parallelization has separated. It should not be regarded as the parallelization itself

  28. Objectives Discussion on Programming Models MapReduce Message Passing Interface (MPI) Message Passing Interface (MPI) Examples of parallel processing Traditional Models of parallel programming Parallel computer architectures Why parallelizing our programs?

  29. Message Passing Interface • In this part, the following concepts of MPI will be described: • Basics • Point-to-point communication • Collective communication

  30. What is MPI? • The Message Passing Interface (MPI) is a message passing library standard  for writing message passing programs • The goal of MPI is to establish a portable, efficient, and flexible standard for message passing • By itself, MPI is NOT a library - but rather the specification of what such a library should be • MPI is not an IEEE or ISO standard, but has in fact, become the industry standardfor writing message passing programs on HPC platforms

  31. Reasons for using MPI

  32. Programming Model • MPI is an example of a message passing programming model • MPI is now used on just about any common parallel architecture including MPP, SMP clusters, workstation clusters and heterogeneous networks • With MPIthe programmer is responsible for correctly identifying parallelism and implementing parallel algorithms using MPI constructs

  33. Communicators and Groups • MPI uses objects called communicators and groups to define which collection of processes may communicate with each other to solve a certain problem • Most MPI routines require you to specify a communicator as an argument • The communicator MPI_COMM_WORLD is often used in calling communication subroutines • MPI_COMM_WORLD is the predefined communicator that includes all of your MPI processes

  34. Ranks • Within a communicator, every process has its own unique, integer identifier referred to as rank, assigned by the system when the process initializes • A rank is sometimes called a task ID. Ranks are contiguous and begin at zero • Ranks are used by the programmer to specify the source and destination of messages • Ranks are often also used conditionally by the application to control program execution (e.g., if rank=0 do this / if rank=1 do that)

  35. Multiple Communicators • It is possible that a problem consists of several sub-problems where each can be solved concurrently • This type of application is typically found in the category of MPMD coupled analysis • We can create a new communicator for each sub-problem as a subset of an existing communicator • MPI allows you to achieve that by using MPI_COMM_SPLIT

  36. Example of Multiple Communicators • Consider a problem with a fluid dynamics part and a structural analysis part, where each part can be computed in parallel MPI_COMM_WORLD Comm_Fluid Comm_Struct Rank=0 Rank=0 Rank=1 Rank=1 Rank=0 Rank=4 Rank=1 Rank=5 Rank=2 Rank=2 Rank=3 Rank=3 Rank=2 Rank=6 Rank=3 Rank=7 • Ranks within MPI_COMM_WORLD are printed in red • Ranks within Comm_Fluid are printed with green • Ranks within Comm_Struct are printed with blue

  37. Next Class Discussion on Programming Models MapReduce Message Passing Interface (MPI) Message Passing Interface (MPI) Examples of parallel processing Traditional Models of parallel programming Parallel computer architectures Programming Models- Part II Why parallelizing our programs?

  38. Message Passing Interface • In this part, the following concepts of MPI will be described: • Basics • Point-to-point communication • Collective communication

  39. Point-to-Point Communication • MPI point-to-point operations typically involve message passing between two, and only two, different MPI tasks • One task performs a send operation and the other performs a matchingreceive operation • Ideally, every send operation would be perfectly synchronized with its matching receive • This is rarely the case. Somehow or other, the MPI implementation must be able to deal with storing data when the two tasks are out of sync Processor1 Processor2 Network sendA recvA

  40. Two Cases • Consider the following two cases: • A send operation occurs 5 seconds before the receive is ready - where is the message stored while the receive is pending? • Multiple sends arrive at the same receiving task which can only accept one send at a time - what happens to the messages that are "backing up"?

  41. Steps Involved in Point-to-Point Communication • The data is copied to the user buffer by the user • The user calls one of the MPI send routines • The system copies the data from the user buffer to the system buffer • The system sends the data from the system buffer to the destination process Process 0 Sender The user calls one of the MPI receive routines The system receives the data from the source process and copies it to the system buffer The system copies data from the system buffer to the user buffer The user uses data in the user buffer User Mode Kernel Mode sendbuf 1 sysbuf 2 Call a send routine Copying data from sendbuf to sysbuf 3 Send data from sysbuf to destination Now sendbuf can be reused 4 Data Process 1 Receiver User Mode Kernel Mode Receive data from source to sysbuf Call a recev routine 1 2 4 sysbuf Now recvbuf contains valid data Copying data from sysbuf to recvbuf recvbuf 3

  42. Blocking Send and Receive • When we use point-to-point communication routines, we usually distinguish between blocking and non-blocking communication • A blocking send routine will only returnafter it is safe to modify the application buffer for reuse • Safe means that modifications will not affect the data intended for the receive task • This does not imply that the data was actually received by the receiver- it may be sitting in the system buffer at the sender side Safe to modify sendbuf Rank 0 Rank 1 sendbuf recvbuf Network recvbuf sendbuf

  43. Blocking Send and Receive • A blocking send can be: • Synchronous: Means there is a handshaking occurring with the receive task to confirm a safe send • Asynchronous: Means the system buffer at the sender side is used to hold the data for eventual delivery to the receiver • A blocking receive only returns after the data has arrived (i.e., stored at the application recvbuf) and is ready for use by the program

  44. Non-Blocking Send and Receive (1) • Non-blocking send and non-blocking receiveroutines behave similarly • They return almost immediately • They do not wait for any communication events to complete such as: • Message copying from user buffer to system buffer • Or the actual arrival of a message

  45. Non-Blocking Send and Receive (2) • However, it is unsafe to modify the application buffer until you make sure that the requested non-blocking operation was actually performed by the library • If you use the application buffer before the copy completes: • Incorrect data may be copied to the system buffer (in case of non-blocking send) • Or your receive buffer does not contain what you want (in case of non-blocking receive) • You can make sure of the completion of the copy by using MPI_WAIT() after the send or receive operations

  46. Why Non-Blocking Communication? • Why do we use non-blocking communication despite its complexity? • Non-blocking communication is generally faster than its corresponding blocking communication • We can overlap computations while the system is copying data back and forth between application and system buffers

  47.  MPI Point-To-Point Communication Routines

  48. Message Order • MPI guarantees that messages will not overtake each other • If a sender sends two messages M1and M2in succession to the same destination, and both match the same receive, the receive operation will receive M1before M2 • If a receiver posts two receives R1 and R2, in succession, and both are looking for the same message, R1will receive the message before R2

  49. Fairness • MPI does not guarantee fairness – itis up to the programmer to prevent operation starvation • For instance, if task 0 and task 1 send competing messages (i.e., messages that match the same receive) to task 2, only one of the sends will complete Task 0 Task 1 Msg A Msg A Task 2 ?

  50. Unidirectional Communication • When you send a message from process 0 to process 1, there are four combinations of MPI subroutines to choose from • Blocking send and blocking receive • Non-blocking send and blocking receive • Blocking send and non-blocking receive • Non-blocking send and non-blocking receive Rank 0 Rank 1 sendbuf recvbuf recvbuf sendbuf

More Related