Intro to Parallel Programming with MPI CS 475

  • Slides: 50
Download presentation
Intro. to Parallel Programming with MPI CS 475 By Dr. Ziad A. Al-Sharif Based

Intro. to Parallel Programming with MPI CS 475 By Dr. Ziad A. Al-Sharif Based on the tutorial from the Argonne National Laboratory https: //www. mcs. anl. gov/~raffenet/permalinks/argonne 19_mpi. php

Outline § Part 1 – Introduction to MPI • Basic concepts • MPI-1, MPI-2,

Outline § Part 1 – Introduction to MPI • Basic concepts • MPI-1, MPI-2, MPI-3 – Installing and running MPI – Point-to-point Communication – Collective Communication § Part 3 – Hybrid Programming • Thread safety specification in MPI and how it enables hybrid programming • MPI + Open. MP • MPI + shared memory • MPI + accelerators – Derived Datatypes § Part 2 – MPI One-sided Communication (RMA) 22

Introduction to MPI 3

Introduction to MPI 3

Sample Parallel Programming Models § Shared Memory Programming – Processes share memory address space

Sample Parallel Programming Models § Shared Memory Programming – Processes share memory address space (threads model) – Application ensures no data corruption (Lock/Unlock) § Transparent Parallelization – Compiler works magic on sequential programs § Directive-based Parallelization – Compiler needs help (e. g. , Open. MP) § Message Passing – Explicit communication between processes • (like sending and receiving emails) 4

The Message-Passing Model § A process is (traditionally) a program counter and address space.

The Message-Passing Model § A process is (traditionally) a program counter and address space. § Processes may have multiple threads – Program counters and associated stacks are sharing a single address space. – But, MPI is for communication among processes, which have separate address spaces. § Inter-process communication consists of – synchronization – movement of data from one process’s address space to another’s. MPI Process MPI 5

MPI programming models § Originally, MPI was designed for distributed memory architectures. § MPI

MPI programming models § Originally, MPI was designed for distributed memory architectures. § MPI implementers adapted, later, their libraries to handle both types of underlying memory architectures seamlessly. – So it is originally for Distributed Memory But can work on shared memory or even hybrid memory architectures

MPI programming models § Today, MPI runs on virtually any hardware platform: – Distributed

MPI programming models § Today, MPI runs on virtually any hardware platform: – Distributed Memory – Shared Memory – Hybrid § The programming model remains a distributed memory model regardless of the underlying physical architecture of the machine. § All parallelism is explicit: – the programmer is responsible for correctly identifying parallelism and implementing parallel algorithms using MPI constructs.

The Message-Passing Model (an example) § Each process has to send/receive data to/from other

The Message-Passing Model (an example) § Each process has to send/receive data to/from other processes § Example: Sorting Integers Process 1 8 O(N log N) 23 19 67 45 35 1 24 13 30 3 5 Process 2 Process 1 8 19 23 35 45 67 1 O(N/2 log N/2) 1 3 5 8 3 5 13 24 30 O(N/2 log N/2) 13 19 23 24 30 35 45 67 O(N) Process 1 8

Standardizing Message-Passing Models with MPI § Before MPI, most parallelization systems were not portable

Standardizing Message-Passing Models with MPI § Before MPI, most parallelization systems were not portable (or very capable) § Early portable systems (PVM, p 4, TCGMSG, Chameleon) were mainly research efforts – Did not address the full spectrum of message-passing issues – Lacked vendor support (Intel, IBM, …etc ) – Were not implemented at the most efficient level § The MPI Forum was a collection of vendors, portability writers and users that wanted to standardize all these efforts 9

What is MPI? § MPI is a Message Passing Interface – The MPI Forum

What is MPI? § MPI is a Message Passing Interface – The MPI Forum organized in 1992 with broad participation by: • • Vendors: IBM, Intel, TMC, SGI, Convex, Meiko Portability library writers: PVM, p 4 Users: application scientists and library writers MPI-1 finished in 18 months – MPI Design: Incorporates the best ideas in a “standard” way • Each function takes fixed arguments • Each function has fixed semantics – Standardizes what the MPI implementation provides and what the application can and cannot expect – Each system can implement it differently as long as the semantics match § MPI is not… – a language or compiler specification – a specific implementation or product 10

MPI-1 § MPI-1 supports the classical message-passing programming model: – basic point-to-point communication, collectives,

MPI-1 § MPI-1 supports the classical message-passing programming model: – basic point-to-point communication, collectives, datatypes, etc § MPI-1 was defined (1994) by a broadly based group of parallel computer vendors, computer scientists, and applications developers. – 2 -year intensive process § Implementations appeared quickly and now MPI is taken for granted as vendor-supported software on any parallel machine. § Free, portable implementations exist for clusters and other environments (MPICH, Open MPI) 1111

Following MPI Standards § MPI-2 was released in 1997 – Several additional features including

Following MPI Standards § MPI-2 was released in 1997 – Several additional features including MPI + threads, MPI-I/O, remote memory access functionality and many others § MPI-2. 1 (2008) and MPI-2. 2 (2009) were released with some corrections to the standard and small features § MPI-3 (2012) added several new features to MPI (see next slide) § MPI-3. 1 (2015) is the latest version of the standard with minor corrections and features § MPI 4. 0 (2020) MPI Standard Draft / MPI 4. 0 Release Candidate § The Standard itself lives at http: //www. mpi-forum. org 12

MPI & Applications 13

MPI & Applications 13

Applications (Science and Engineering) § MPI is widely used in large scale parallel applications

Applications (Science and Engineering) § MPI is widely used in large scale parallel applications in science and engineering – Atmosphere, Earth, Environment – Physics - applied, nuclear, particle, condensed matter, high pressure, fusion, photonics – Bioscience, Biotechnology, Genetics – Chemistry, Molecular Sciences – Geology, Seismology – Mechanical Engineering - from prosthetics to spacecraft – Electrical Engineering, Circuit Design, Microelectronics – Computer Science, Mathematics 14

Turbo machinery (Gas turbine/compressor) Biology (heart murmur simulation) Transportation & traffic application Drilling application

Turbo machinery (Gas turbine/compressor) Biology (heart murmur simulation) Transportation & traffic application Drilling application Astrophysics application 15 15

Reasons for Using MPI § Standardization - MPI is the only message passing library

Reasons for Using MPI § Standardization - MPI is the only message passing library which can be considered a standard. It is supported on virtually all HPC platforms. Practically, it has replaced all previous message passing libraries § Portability - There is no need to modify your source code when you port your application to a different platform that supports (and is compliant with) the MPI standard § Performance Opportunities - Vendor implementations should be able to exploit native hardware features to optimize performance § Functionality – Rich set of features § Availability - A variety of implementations are available, both vendor and public domain – MPICH is a popular open-source and free implementation of MPI – Vendors and other collaborators take MPICH and add support for their systems • Intel MPI, IBM Blue Gene MPI, Cray MPI, Microsoft MPI, MVAPICH, MPICH-MX 16

Important considerations while using MPI § All parallelism is explicit: the programmer is responsible

Important considerations while using MPI § All parallelism is explicit: the programmer is responsible for correctly identifying parallelism and implementing parallel algorithms using MPI constructs § Approach in this Course – Example driven • A few running examples used throughout the course • Other smaller examples used to illustrate specific features – Example exercises available to download 17

How does MPI work? (1/2) 18

How does MPI work? (1/2) 18

How does MPI work? Multicore Systems. (2/2) 19

How does MPI work? Multicore Systems. (2/2) 19

General MPI Program Structure 20

General MPI Program Structure 20

General MPI Program Structure § Include the header files – #include "mpi. h" §

General MPI Program Structure § Include the header files – #include "mpi. h" § Function declarations, global variables, …etc § Main function begin – Do some serial work – Initialize MPI environment. • MPI_Init( &argc, &argv ) – This function must be called in every MPI program, must be called before any other MPI functions and must be called only once in an MPI program – Do work that will be executed in parallel and perform message passing. – Terminate MPI environment. • MPI_Finalize() – This function should be the last MPI routine called in every MPI program – No other MPI routines may be called after it – Do some serial work § Program ends.

General MPI Program Structure: Basic Outline

General MPI Program Structure: Basic Outline

Simple MPI Program Identifying Processes #include <mpi. h> #include <stdio. h> int main(int argc,

Simple MPI Program Identifying Processes #include <mpi. h> #include <stdio. h> int main(int argc, char ** argv) { int rank, size; Basic requirements for an MPI program MPI_Init(&argc, &argv); MPI_Comm_rank(MPI_COMM_WORLD, &rank); MPI_Comm_size(MPI_COMM_WORLD, &size); printf("I am %d of %dn", rank + 1, size); MPI_Finalize(); return 0; hello. c Basic program where each process prints its rank } For compilation: mpicc hello. c –o hello For running : mpiexec –n N. /hello 23

General MPI Program Structure

General MPI Program Structure

Compiling and Running MPI applications 25

Compiling and Running MPI applications 25

Compiling and Running MPI applications § MPI is a library – Applications can be

Compiling and Running MPI applications § MPI is a library – Applications can be written in C, C++, Fortran or even Python and appropriate calls to MPI can be added where required § Compilation: – Regular applications: % gcc test. c -o test – MPI applications % mpicc test. c -o test § Execution: – Regular applications %. /test – MPI applications (running with 16 processes) % mpiexec –n 16. /test 26

Compilation wrapper script to compile $ mpicc -O -o mpi_hello source file mpi_hello. c

Compilation wrapper script to compile $ mpicc -O -o mpi_hello source file mpi_hello. c Mix with openmp $ mpicc -O -o mpi_hello. c -fopenmp

Execution with mpirun or mpiexec mpirun -n <number of processes> $ mpirun -n 1

Execution with mpirun or mpiexec mpirun -n <number of processes> $ mpirun -n 1 Start an mpi job $ mpirun <executable> . /mpi_hello run with 1 process -n with this number of processes 4 . /mpi_hello And use/run this executable

Building and Installation of MPICH 29

Building and Installation of MPICH 29

What is MPICH? § MPICH is a high-performance and widely portable opensource implementation of

What is MPICH? § MPICH is a high-performance and widely portable opensource implementation of MPI § It provides all features of MPI that have been defined so far (including MPI-1, MPI-2. 0, MPI-2. 1, MPI-2. 2, MPI-3. 0 and MPI-3. 1) § Active development led by Argonne National Laboratory – Several close collaborators who contribute many features, bug fixes, testing for quality assurance, etc. • Intel, Cray, Mellanox, The Ohio State University, Microsoft, and many others 30

Getting Started with MPICH § From UNIX distributions – Most UNIX/Linux distributions package MPICH

Getting Started with MPICH § From UNIX distributions – Most UNIX/Linux distributions package MPICH for easy installation – apt-get (Ubuntu/Debian), yum (Fedora, Centos), brew/port (Mac OS) $ sudo apt-get update $ sudo apt-get install mpich § Open-source for source-based installation – – Download MPICH • Go to http: //www. mpich. org and follow the downloads link • The download will be a zipped tarball Build MPICH • Unzip/untar the tarball % tar xvfz mpich-<latest_version>. tar. gz % cd mpich-<latest_version> %. /configure –-prefix=/where/to/install/mpich |& tee c. log % make |& tee m. log % make install |& tee mi. log % Add /where/to/install/mpich/bin to your PATH 31

Compiling MPI programs with MPICH § Compilation Wrappers – For C programs: $ mpicc

Compiling MPI programs with MPICH § Compilation Wrappers – For C programs: $ mpicc test. c –o test – For C++ programs: $ mpicxx test. cpp –o test – Fortran programs: $ mpifort test. f 90 –o test § You may link other libraries if required too – To link to a math library: $ mpicc test. c –o test -lm § You can just assume that “mpicc” and friends have replaced your regular compilers (gcc, gfortran, etc. ) 32

Running MPI programs with MPICH § Launch 16 processes on the local node: %

Running MPI programs with MPICH § Launch 16 processes on the local node: % mpiexec –n 16. /test § Launch 16 processes on 4 nodes (each has 4 cores) % mpiexec –hosts h 1: 4, h 2: 4, h 3: 4, h 4: 4 –n 16. /test – Runs the first four processes on h 1, the next four on h 2, etc. % mpiexec –hosts h 1, h 2, h 3, h 4 –n 16. /test – Runs the first process on h 1, the second on h 2, etc. , and wraps around – So, h 1 will have the 1 st, 5 th, 9 th and 13 th processes § If there are many nodes, it might be easier to create a host file % cat hf h 1: 4 h 2: 2 % mpiexec –hostfile hf –n 16. /test 33

Trying some example programs § MPICH comes packaged with several example programs using almost

Trying some example programs § MPICH comes packaged with several example programs using almost ALL of MPICH’s functionality § A simple program to try out is the PI example written in C (cpi. c) – calculates the value of PI in parallel (available in the examples directory when you build MPICH) % mpiexec –n 16. /examples/cpi § The output will show many processes are running, and the error in calculating PI § Next, try it with multiple hosts % mpiexec –hosts h 1: 2, h 2: 4 –n 16. /examples/cpi 34

Programming With MPI An Example in C, C++, and Python 36

Programming With MPI An Example in C, C++, and Python 36

MPI is Simple § Many parallel programs can be written using just these six

MPI is Simple § Many parallel programs can be written using just these six functions, only two of which are non-trivial: – MPI_INIT – MPI_FINALIZE – MPI_COMM_SIZE – MPI_COMM_RANK – MPI_SEND – MPI_RECV § And in order to measure time, you may need: – MPI_Wtime()

Finding Out About the Environment § Two important questions raised early: – How many

Finding Out About the Environment § Two important questions raised early: – How many processes are participating in this computation? – Which one am I? § MPI provides functions to answer these questions: – MPI_Comm_size : • reports the number of processes. – MPI_Comm_rank : • reports the rank, a number between 0 and size-1, identifying the calling process • p processes are numbered 0, 1, 2, . . p-1

Finding Out About the Environment: Communicators number of processes in the communicator my rank

Finding Out About the Environment: Communicators number of processes in the communicator my rank (the process making this call)

mpi_hello. c (C) #include <mpi. h> #include <stdio. h> int main( int argc, char

mpi_hello. c (C) #include <mpi. h> #include <stdio. h> int main( int argc, char *argv[] ) { int rank, size; MPI_Init( &argc, &argv ); MPI_Comm_rank( MPI_COMM_WORLD, &rank ); MPI_Comm_size( MPI_COMM_WORLD, &size ); printf("I am %d of %d!n", rank, size ); MPI_Finalize(); return 0; }

Sample Execution $ mpirun -n 1. /mpi_hello I am 0 of 1 ! $

Sample Execution $ mpirun -n 1. /mpi_hello I am 0 of 1 ! $ mpirun -n 4. /mpi_hello I am 0 of 4 ! I am 1 of 4 ! I am 2 of 4 ! I am 3 of 4 !

Notes on Hello World § All MPI programs – begin with MPI_Init() – and

Notes on Hello World § All MPI programs – begin with MPI_Init() – and ends with MPI_Finalize(). § MPI_COMM_WORLD is defined by mpi. h (in C) and designates/labels all processes in the MPI “job” § Each statement executes independently in each process – including the printf/print statements

mpi_hello. cpp (C++) #include <mpi. h> #include <iostream> int main( int argc, char *argv[]

mpi_hello. cpp (C++) #include <mpi. h> #include <iostream> int main( int argc, char *argv[] ) { int rank, size; MPI: : Init(argc, argv); rank = MPI: : COMM_WORLD. Get_rank(); size = MPI: : COMM_WORLD. Get_size(); std: : cout <<"I am "<< rank <<" of "<<size<< "!n"; MPI: : Finalize(); $ mpi. CC hello. cpp –o hello return 0; OR } $ mpicxx hello. cpp –o hello OR $ mpic++ hello. cpp –o hello $ mpiexec –n N. /hello

mpi_hello. py (Python) § MPI in Python using the mpi 4 py, this package

mpi_hello. py (Python) § MPI in Python using the mpi 4 py, this package provides Python bindings for the Message Passing Interface (MPI) standard. It is implemented on top of the MPI-1/2/3 specification and exposes an API which grounds on the standard MPI-2 C++ bindings. § Make sure to install mpi 4 py – You may think of pip or conda – Then, you can run it $ mpirun -n 4 python mpi_hello. py from mpi 4 py import MPI size = MPI. COMM_WORLD. Get_size() rank = MPI. COMM_WORLD. Get_rank() name = MPI. Get_processor_name() print("I am process {0: 3} of {1: 3} on {2: 3}n". format(rank, size, name)) Source: https: //mpi 4 py. readthedocs. io/en/stable/tutorial. html

Extra References

Extra References

Web Pointers § MPI Standard : http: //www. mpi-forum. org/docs/ § MPI Forum :

Web Pointers § MPI Standard : http: //www. mpi-forum. org/docs/ § MPI Forum : http: //www. mpi-forum. org/ § Currently, there are different MPI implementations including: – MPICH : http: //www. mpich. org – MVAPICH : http: //mvapich. cse. ohio-state. edu/ – Intel MPI: http: //software. intel. com/en-us/intel-mpi-library/ – Microsoft MPI: www. microsoft. com/en-us/download/details. aspx? id=39961 – Open MPI : http: //www. open-mpi. org/ – IBM Spectrum MPI, Cray MPI, TH MPI, … § Several MPI tutorials can be found on the web 46

Latest MPI 3. 1 Standard in Book Form Available from amazon. com http: //www.

Latest MPI 3. 1 Standard in Book Form Available from amazon. com http: //www. amazon. com/dp/B 015 CJ 42 CU/ 47

Tutorial Books on MPI Basic MPI Advanced MPI, including MPI-3 48

Tutorial Books on MPI Basic MPI Advanced MPI, including MPI-3 48

Book on Parallel Programming Models Edited by Pavan Balaji • • • MPI: W.

Book on Parallel Programming Models Edited by Pavan Balaji • • • MPI: W. Gropp and R. Thakur • • • Chapel: B. Chamberlain • • Cn. C: K. Knobe, M. Burke, and F. Schlimbach • • Cilk Plus: A. Robison and C. Leiserson GASNet: P. Hargrove Open. SHMEM: J. Kuehn and S. Poole UPC: K. Yelick and Y. Zheng Global Arrays: S. Krishnamoorthy, J. Daily, A. Vishnu, and B. Palmer Charm++: L. Kale, N. Jain, and J. Lifflander ADLB: E. Lusk, R. Butler, and S. Pieper Scioto: J. Dinan SWIFT: T. Armstrong, J. M. Wozniak, M. Wilde, and I. Foster Open. MP: B. Chapman, D. Eachempati, and S. Chandrasekaran Intel TBB: A. Kukanov CUDA: W. Hwu and D. Kirk Open. CL: T. Mattson 49

Overview of New Features in MPI-3 § Major new features – Nonblocking collectives –

Overview of New Features in MPI-3 § Major new features – Nonblocking collectives – Neighborhood collectives – Improved one-sided communication interface – Tools interface – Fortran 2008 bindings § Other new features – Matching Probe and Recv for thread-safe probe and receive – Noncollective communicator creation function – “const” correct C bindings – Comm_split_type function – Nonblocking Comm_dup – Type_create_hindexed_block function § C++ bindings removed § Previously deprecated functions removed § MPI 3. 1 added nonblocking collective I/O functions 50

MVAPICH Open MPI Cray Tianhe HPE Fujitsu MS MPC NEC Sunway RIKEN AMPI IMPI

MVAPICH Open MPI Cray Tianhe HPE Fujitsu MS MPC NEC Sunway RIKEN AMPI IMPI MPICH-OFI BG/Q (legacy) 1 PE (legacy) 2 Spectrum NBC ✔ ✔ ✔ ✔ ✔ Nbr. Coll. ✔ ✔ ✔ ✘ ✔ ✔ ✔ RMA ✔ ✔ ✔ (*) ✔ ✔ Q 2 ‘ 18 Shr. mem ✔ ✔ ✔ ✔ ✔ Q 1 ‘ 18 MPI_T ✔ ✔ ✔ * ✔ ✔ Q 2 ‘ 18 Comm-create group ✔ ✔ ✔ * ✔ ✔ ✔ F 08 Bindings ✔ ✔ ✔ ✔ ✘ ✘ ✔ ✔ Q 2 ‘ 18 New Dtypes ✔ ✔ ✔ ✔ ✔ Large Counts ✔ ✔ ✔ ✔ ✔ MProbe ✔ ✔ ✔ ✔ ✔ Q 1 ‘ 18 NBC I/O ✔ ✔ ✘ ✘ * ✔ ✘ ✔ Q 3 ‘ 18 IBM Intel MPICH Status of MPI-3. 1 Implementations Release dates are estimates; subject to change at any time “✘” indicates no publicly announced plan to implement/support that feature Platform-specific restrictions might apply to the supported features 1 Open Source but unsupported 2 No MPI_T variables exposed * Under development (*) Partly done 51