- •Contents
- •List of Figures
- •List of Tables
- •Acknowledgments
- •Introduction to MPI
- •Overview and Goals
- •Background of MPI-1.0
- •Background of MPI-1.1, MPI-1.2, and MPI-2.0
- •Background of MPI-1.3 and MPI-2.1
- •Background of MPI-2.2
- •Who Should Use This Standard?
- •What Platforms Are Targets For Implementation?
- •What Is Included In The Standard?
- •What Is Not Included In The Standard?
- •Organization of this Document
- •MPI Terms and Conventions
- •Document Notation
- •Naming Conventions
- •Semantic Terms
- •Data Types
- •Opaque Objects
- •Array Arguments
- •State
- •Named Constants
- •Choice
- •Addresses
- •Language Binding
- •Deprecated Names and Functions
- •Fortran Binding Issues
- •C Binding Issues
- •C++ Binding Issues
- •Functions and Macros
- •Processes
- •Error Handling
- •Implementation Issues
- •Independence of Basic Runtime Routines
- •Interaction with Signals
- •Examples
- •Point-to-Point Communication
- •Introduction
- •Blocking Send and Receive Operations
- •Blocking Send
- •Message Data
- •Message Envelope
- •Blocking Receive
- •Return Status
- •Passing MPI_STATUS_IGNORE for Status
- •Data Type Matching and Data Conversion
- •Type Matching Rules
- •Type MPI_CHARACTER
- •Data Conversion
- •Communication Modes
- •Semantics of Point-to-Point Communication
- •Buffer Allocation and Usage
- •Nonblocking Communication
- •Communication Request Objects
- •Communication Initiation
- •Communication Completion
- •Semantics of Nonblocking Communications
- •Multiple Completions
- •Non-destructive Test of status
- •Probe and Cancel
- •Persistent Communication Requests
- •Send-Receive
- •Null Processes
- •Datatypes
- •Derived Datatypes
- •Type Constructors with Explicit Addresses
- •Datatype Constructors
- •Subarray Datatype Constructor
- •Distributed Array Datatype Constructor
- •Address and Size Functions
- •Lower-Bound and Upper-Bound Markers
- •Extent and Bounds of Datatypes
- •True Extent of Datatypes
- •Commit and Free
- •Duplicating a Datatype
- •Use of General Datatypes in Communication
- •Correct Use of Addresses
- •Decoding a Datatype
- •Examples
- •Pack and Unpack
- •Canonical MPI_PACK and MPI_UNPACK
- •Collective Communication
- •Introduction and Overview
- •Communicator Argument
- •Applying Collective Operations to Intercommunicators
- •Barrier Synchronization
- •Broadcast
- •Example using MPI_BCAST
- •Gather
- •Examples using MPI_GATHER, MPI_GATHERV
- •Scatter
- •Examples using MPI_SCATTER, MPI_SCATTERV
- •Example using MPI_ALLGATHER
- •All-to-All Scatter/Gather
- •Global Reduction Operations
- •Reduce
- •Signed Characters and Reductions
- •MINLOC and MAXLOC
- •All-Reduce
- •Process-local reduction
- •Reduce-Scatter
- •MPI_REDUCE_SCATTER_BLOCK
- •MPI_REDUCE_SCATTER
- •Scan
- •Inclusive Scan
- •Exclusive Scan
- •Example using MPI_SCAN
- •Correctness
- •Introduction
- •Features Needed to Support Libraries
- •MPI's Support for Libraries
- •Basic Concepts
- •Groups
- •Contexts
- •Intra-Communicators
- •Group Management
- •Group Accessors
- •Group Constructors
- •Group Destructors
- •Communicator Management
- •Communicator Accessors
- •Communicator Constructors
- •Communicator Destructors
- •Motivating Examples
- •Current Practice #1
- •Current Practice #2
- •(Approximate) Current Practice #3
- •Example #4
- •Library Example #1
- •Library Example #2
- •Inter-Communication
- •Inter-communicator Accessors
- •Inter-communicator Operations
- •Inter-Communication Examples
- •Caching
- •Functionality
- •Communicators
- •Windows
- •Datatypes
- •Error Class for Invalid Keyval
- •Attributes Example
- •Naming Objects
- •Formalizing the Loosely Synchronous Model
- •Basic Statements
- •Models of Execution
- •Static communicator allocation
- •Dynamic communicator allocation
- •The General case
- •Process Topologies
- •Introduction
- •Virtual Topologies
- •Embedding in MPI
- •Overview of the Functions
- •Topology Constructors
- •Cartesian Constructor
- •Cartesian Convenience Function: MPI_DIMS_CREATE
- •General (Graph) Constructor
- •Distributed (Graph) Constructor
- •Topology Inquiry Functions
- •Cartesian Shift Coordinates
- •Partitioning of Cartesian structures
- •Low-Level Topology Functions
- •An Application Example
- •MPI Environmental Management
- •Implementation Information
- •Version Inquiries
- •Environmental Inquiries
- •Tag Values
- •Host Rank
- •IO Rank
- •Clock Synchronization
- •Memory Allocation
- •Error Handling
- •Error Handlers for Communicators
- •Error Handlers for Windows
- •Error Handlers for Files
- •Freeing Errorhandlers and Retrieving Error Strings
- •Error Codes and Classes
- •Error Classes, Error Codes, and Error Handlers
- •Timers and Synchronization
- •Startup
- •Allowing User Functions at Process Termination
- •Determining Whether MPI Has Finished
- •Portable MPI Process Startup
- •The Info Object
- •Process Creation and Management
- •Introduction
- •The Dynamic Process Model
- •Starting Processes
- •The Runtime Environment
- •Process Manager Interface
- •Processes in MPI
- •Starting Processes and Establishing Communication
- •Reserved Keys
- •Spawn Example
- •Manager-worker Example, Using MPI_COMM_SPAWN.
- •Establishing Communication
- •Names, Addresses, Ports, and All That
- •Server Routines
- •Client Routines
- •Name Publishing
- •Reserved Key Values
- •Client/Server Examples
- •Ocean/Atmosphere - Relies on Name Publishing
- •Simple Client-Server Example.
- •Other Functionality
- •Universe Size
- •Singleton MPI_INIT
- •MPI_APPNUM
- •Releasing Connections
- •Another Way to Establish MPI Communication
- •One-Sided Communications
- •Introduction
- •Initialization
- •Window Creation
- •Window Attributes
- •Communication Calls
- •Examples
- •Accumulate Functions
- •Synchronization Calls
- •Fence
- •General Active Target Synchronization
- •Lock
- •Assertions
- •Examples
- •Error Handling
- •Error Handlers
- •Error Classes
- •Semantics and Correctness
- •Atomicity
- •Progress
- •Registers and Compiler Optimizations
- •External Interfaces
- •Introduction
- •Generalized Requests
- •Examples
- •Associating Information with Status
- •MPI and Threads
- •General
- •Initialization
- •Introduction
- •File Manipulation
- •Opening a File
- •Closing a File
- •Deleting a File
- •Resizing a File
- •Preallocating Space for a File
- •Querying the Size of a File
- •Querying File Parameters
- •File Info
- •Reserved File Hints
- •File Views
- •Data Access
- •Data Access Routines
- •Positioning
- •Synchronism
- •Coordination
- •Data Access Conventions
- •Data Access with Individual File Pointers
- •Data Access with Shared File Pointers
- •Noncollective Operations
- •Collective Operations
- •Seek
- •Split Collective Data Access Routines
- •File Interoperability
- •Datatypes for File Interoperability
- •Extent Callback
- •Datarep Conversion Functions
- •Matching Data Representations
- •Consistency and Semantics
- •File Consistency
- •Random Access vs. Sequential Files
- •Progress
- •Collective File Operations
- •Type Matching
- •Logical vs. Physical File Layout
- •File Size
- •Examples
- •Asynchronous I/O
- •I/O Error Handling
- •I/O Error Classes
- •Examples
- •Subarray Filetype Constructor
- •Requirements
- •Discussion
- •Logic of the Design
- •Examples
- •MPI Library Implementation
- •Systems with Weak Symbols
- •Systems Without Weak Symbols
- •Complications
- •Multiple Counting
- •Linker Oddities
- •Multiple Levels of Interception
- •Deprecated Functions
- •Deprecated since MPI-2.0
- •Deprecated since MPI-2.2
- •Language Bindings
- •Overview
- •Design
- •C++ Classes for MPI
- •Class Member Functions for MPI
- •Semantics
- •C++ Datatypes
- •Communicators
- •Exceptions
- •Mixed-Language Operability
- •Problems With Fortran Bindings for MPI
- •Problems Due to Strong Typing
- •Problems Due to Data Copying and Sequence Association
- •Special Constants
- •Fortran 90 Derived Types
- •A Problem with Register Optimization
- •Basic Fortran Support
- •Extended Fortran Support
- •The mpi Module
- •No Type Mismatch Problems for Subroutines with Choice Arguments
- •Additional Support for Fortran Numeric Intrinsic Types
- •Language Interoperability
- •Introduction
- •Assumptions
- •Initialization
- •Transfer of Handles
- •Status
- •MPI Opaque Objects
- •Datatypes
- •Callback Functions
- •Error Handlers
- •Reduce Operations
- •Addresses
- •Attributes
- •Extra State
- •Constants
- •Interlanguage Communication
- •Language Bindings Summary
- •Groups, Contexts, Communicators, and Caching Fortran Bindings
- •External Interfaces C++ Bindings
- •Change-Log
- •Bibliography
- •Examples Index
- •MPI Declarations Index
- •MPI Function Index
4.2. PACK AND UNPACK |
121 |
... other combiner values ...
default:
printf( "Unrecognized combiner type\n" );
}
return 1;
}
4.2 Pack and Unpack
Some existing communication libraries provide pack/unpack functions for sending noncontiguous data. In these, the user explicitly packs data into a contiguous bu er before sending it, and unpacks it from a contiguous bu er after receiving it. Derived datatypes, which are described in Section 4.1, allow one, in most cases, to avoid explicit packing and unpacking. The user speci es the layout of the data to be sent or received, and the communication library directly accesses a noncontiguous bu er. The pack/unpack routines are provided for compatibility with previous libraries. Also, they provide some functionality that is not otherwise available in MPI. For instance, a message can be received in several parts, where the receive operation done on a later part may depend on the content of a former part. Another use is that outgoing messages may be explicitly bu ered in user supplied space, thus overriding the system bu ering policy. Finally, the availability of pack and unpack operations facilitates the development of additional communication libraries layered on top of MPI.
MPI_PACK(inbuf, incount, datatype, outbuf, outsize, position, comm)
IN |
inbuf |
input bu er start (choice) |
IN |
incount |
number of input data items (non-negative integer) |
IN |
datatype |
datatype of each input data item (handle) |
OUT |
outbuf |
output bu er start (choice) |
IN |
outsize |
output bu er size, in bytes (non-negative integer) |
INOUT |
position |
current position in bu er, in bytes (integer) |
IN |
comm |
communicator for packed message (handle) |
int MPI_Pack(void* inbuf, int incount, MPI_Datatype datatype, void *outbuf, int outsize, int *position, MPI_Comm comm)
MPI_PACK(INBUF, INCOUNT, DATATYPE, OUTBUF, OUTSIZE, POSITION, COMM, IERROR) <type> INBUF(*), OUTBUF(*)
INTEGER INCOUNT, DATATYPE, OUTSIZE, POSITION, COMM, IERROR
fvoid MPI::Datatype::Pack(const void* inbuf, int incount, void *outbuf, int outsize, int& position, const MPI::Comm &comm) const
(binding deprecated, see Section 15.2) g
Packs the message in the send bu er speci ed by inbuf, incount, datatype into the bu er space speci ed by outbuf and outsize. The input bu er can be any communication bu er allowed in MPI_SEND. The output bu er is a contiguous storage area containing outsize
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
122 |
CHAPTER 4. DATATYPES |
1bytes, starting at the address outbuf (length is counted in bytes, not elements, as if it were
2a communication bu er for a message of type MPI_PACKED).
3The input value of position is the rst location in the output bu er to be used for
4packing. position is incremented by the size of the packed message, and the output value
5of position is the rst location in the output bu er following the locations occupied by the
6packed message. The comm argument is the communicator that will be subsequently used
7
8
9
for sending the packed message.
10 |
MPI_UNPACK(inbuf, insize, position, outbuf, outcount, datatype, comm) |
|||
11 |
IN |
inbuf |
input bu er start (choice) |
|
|
||||
12 |
IN |
insize |
size of input bu er, in bytes (non-negative integer) |
|
13 |
||||
|
|
|
||
14 |
INOUT |
position |
current position in bytes (integer) |
|
15 |
OUT |
outbuf |
output bu er start (choice) |
|
|
||||
16 |
IN |
outcount |
number of items to be unpacked (integer) |
|
17 |
||||
|
|
|
||
18 |
IN |
datatype |
datatype of each output data item (handle) |
|
19 |
IN |
comm |
communicator for packed message (handle) |
|
|
||||
20 |
|
|
|
|
21 |
int MPI_Unpack(void* inbuf, int insize, int *position, void *outbuf, |
|||
|
||||
22 |
|
int outcount, MPI_Datatype datatype, MPI_Comm comm) |
||
|
|
|||
23 |
|
|
|
|
24 |
MPI_UNPACK(INBUF, INSIZE, POSITION, OUTBUF, OUTCOUNT, DATATYPE, COMM, |
|||
25 |
|
IERROR) |
|
26<type> INBUF(*), OUTBUF(*)
27INTEGER INSIZE, POSITION, OUTCOUNT, DATATYPE, COMM, IERROR
28
fvoid MPI::Datatype::Unpack(const void* inbuf, int insize, void *outbuf,
29
int outcount, int& position, const MPI::Comm& comm) const
30
(binding deprecated, see Section 15.2) g
31
32Unpacks a message into the receive bu er speci ed by outbuf, outcount, datatype from
33the bu er space speci ed by inbuf and insize. The output bu er can be any communication
34bu er allowed in MPI_RECV. The input bu er is a contiguous storage area containing insize
35bytes, starting at address inbuf. The input value of position is the rst location in the input
36bu er occupied by the packed message. position is incremented by the size of the packed
37message, so that the output value of position is the rst location in the input bu er after
38the locations occupied by the message that was unpacked. comm is the communicator used
39to receive the packed message.
40
41Advice to users. Note the di erence between MPI_RECV and MPI_UNPACK: in
42MPI_RECV, the count argument speci es the maximum number of items that can
43be received. The actual number of items received is determined by the length of
44the incoming message. In MPI_UNPACK, the count argument speci es the actual
45number of items that are unpacked; the \size" of the corresponding message is the
46increment in position. The reason for this change is that the \incoming message size"
47is not predetermined since the user decides how much to unpack; nor is it easy to
48determine the \message size" from the number of items to be unpacked. In fact, in a
4.2. PACK AND UNPACK |
123 |
heterogeneous system, this number may not be determined a priori. (End of advice to users.)
To understand the behavior of pack and unpack, it is convenient to think of the data part of a message as being the sequence obtained by concatenating the successive values sent in that message. The pack operation stores this sequence in the bu er space, as if sending the message to that bu er. The unpack operation retrieves this sequence from bu er space, as if receiving a message from that bu er. (It is helpful to think of internal Fortran les or sscanf in C, for a similar function.)
Several messages can be successively packed into one packing unit. This is e ected by several successive related calls to MPI_PACK, where the rst call provides position = 0, and each successive call inputs the value of position that was output by the previous call, and the same values for outbuf, outcount and comm. This packing unit now contains the equivalent information that would have been stored in a message by one send call with a send bu er that is the \concatenation" of the individual send bu ers.
A packing unit can be sent using type MPI_PACKED. Any point to point or collective communication function can be used to move the sequence of bytes that forms the packing unit from one process to another. This packing unit can now be received using any receive operation, with any datatype: the type matching rules are relaxed for messages sent with type MPI_PACKED.
A message sent with any type (including MPI_PACKED) can be received using the type MPI_PACKED. Such a message can then be unpacked by calls to MPI_UNPACK.
A packing unit (or a message created by a regular, \typed" send) can be unpacked into several successive messages. This is e ected by several successive related calls to MPI_UNPACK, where the rst call provides position = 0, and each successive call inputs the value of position that was output by the previous call, and the same values for inbuf, insize and comm.
The concatenation of two packing units is not necessarily a packing unit; nor is a substring of a packing unit necessarily a packing unit. Thus, one cannot concatenate two packing units and then unpack the result as one packing unit; nor can one unpack a substring of a packing unit as a separate packing unit. Each packing unit, that was created by a related sequence of pack calls, or by a regular send, must be unpacked as a unit, by a sequence of related unpack calls.
Rationale. The restriction on \atomic" packing and unpacking of packing units allows the implementation to add at the head of packing units additional information, such as a description of the sender architecture (to be used for type conversion, in a heterogeneous environment) (End of rationale.)
The following call allows the user to nd out how much space is needed to pack a message and, thus, manage space allocation for bu ers.
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
1
2
3
4
5
6
7
8
9
124 |
|
CHAPTER 4. DATATYPES |
MPI_PACK_SIZE(incount, datatype, comm, size) |
||
IN |
incount |
count argument to packing call (non-negative integer) |
IN |
datatype |
datatype argument to packing call (handle) |
IN |
comm |
communicator argument to packing call (handle) |
OUT |
size |
upper bound on size of packed message, in bytes (non- |
|
|
negative integer) |
10 |
int MPI_Pack_size(int incount, MPI_Datatype datatype, MPI_Comm comm, |
|
|
11 |
int *size) |
|
|
12 |
MPI_PACK_SIZE(INCOUNT, DATATYPE, COMM, SIZE, IERROR) |
|
|
13 |
INTEGER INCOUNT, DATATYPE, COMM, SIZE, IERROR |
|
|
14 |
fint MPI::Datatype::Pack_size(int incount, const MPI::Comm& comm) const |
15 |
|
16 |
(binding deprecated, see Section 15.2) g |
17A call to MPI_PACK_SIZE(incount, datatype, comm, size) returns in size an upper bound
18on the increment in position that is e ected by a call to MPI_PACK(inbuf, incount, datatype,
19outbuf, outcount, position, comm).
20
21Rationale. The call returns an upper bound, rather than an exact bound, since the
22exact amount of space needed to pack the message may depend on the context (e.g.,
23rst message packed in a packing unit may take more space). (End of rationale.)
24 |
|
|
25 |
Example 4.21 An example using MPI_PACK. |
|
26 |
|
|
27 |
int |
position, i, j, a[2]; |
|
|
|
28 |
char |
buff[1000]; |
|
|
|
29 |
|
|
MPI_Comm_rank(MPI_COMM_WORLD, &myrank);
30
if (myrank == 0)
31
{
32
33
34
/* SENDER CODE */
35
36
37
38
position = 0; |
|
|
|
|
|
MPI_Pack(&i, |
1, |
MPI_INT, buff, |
1000, |
&position, |
MPI_COMM_WORLD); |
MPI_Pack(&j, |
1, |
MPI_INT, buff, |
1000, |
&position, |
MPI_COMM_WORLD); |
MPI_Send( buff, |
position, MPI_PACKED, 1, 0, MPI_COMM_WORLD); |
}
39
else /* RECEIVER CODE */
40
MPI_Recv( a, 2, MPI_INT, 0, 0, MPI_COMM_WORLD);
41
42
Example 4.22 An elaborate example.
43
44 int position, i;
45float a[1000];
46char buff[1000];
47 |
|
48 |
MPI_Comm_rank(MPI_Comm_world, &myrank); |
|
4.2. PACK AND UNPACK |
125 |
if (myrank == 0)
{
/* SENDER CODE */
int len[2]; MPI_Aint disp[2];
MPI_Datatype type[2], newtype;
/* build datatype for i followed by a[0]...a[i-1] */
len[0] = 1; len[1] = i;
MPI_Address( &i, disp); MPI_Address( a, disp+1); type[0] = MPI_INT; type[1] = MPI_FLOAT;
MPI_Type_struct( 2, len, disp, type, &newtype);
MPI_Type_commit( &newtype);
/* Pack i followed by a[0]...a[i-1]*/
position = 0;
MPI_Pack( MPI_BOTTOM, 1, newtype, buff, 1000, &position, MPI_COMM_WORLD);
/* Send */
MPI_Send( buff, position, MPI_PACKED, 1, 0,
MPI_COMM_WORLD);
/* *****
One can replace the last three lines with
MPI_Send( MPI_BOTTOM, 1, newtype, 1, 0, MPI_COMM_WORLD);
***** */
}
else if (myrank == 1)
{
/* RECEIVER CODE */
MPI_Status status;
/* Receive */
MPI_Recv( buff, 1000, MPI_PACKED, 0, 0, MPI_COMM_WORLD, &status);
/* Unpack i */
position = 0;
MPI_Unpack(buff, 1000, &position, &i, 1, MPI_INT, MPI_COMM_WORLD);
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
126 |
CHAPTER 4. DATATYPES |
1
2/* Unpack a[0]...a[i-1] */
3MPI_Unpack(buff, 1000, &position, a, i, MPI_FLOAT, MPI_COMM_WORLD);
4}
5
6Example 4.23 Each process sends a count, followed by count characters to the root; the
7root concatenates all characters into one string.
8 |
int count, gsize, counts[64], totalcount, k1, k2, k, |
|
9displs[64], position, concat_pos;
10 |
char chr[100], *lbuf, *rbuf, *cbuf; |
|
|
11 |
|
12MPI_Comm_size(comm, &gsize);
13MPI_Comm_rank(comm, &myrank);
14 |
|
15 |
/* allocate local pack buffer */ |
|
|
16 |
MPI_Pack_size(1, MPI_INT, comm, &k1); |
|
|
17 |
MPI_Pack_size(count, MPI_CHAR, comm, &k2); |
|
|
18 |
k = k1+k2; |
|
|
19 |
lbuf = (char *)malloc(k); |
|
|
20 |
|
21/* pack count, followed by count characters */
22position = 0;
23MPI_Pack(&count, 1, MPI_INT, lbuf, k, &position, comm);
24MPI_Pack(chr, count, MPI_CHAR, lbuf, k, &position, comm);
25 |
|
26 |
if (myrank != root) { |
|
|
27 |
/* gather at root sizes of all packed messages */ |
|
|
28 |
MPI_Gather( &position, 1, MPI_INT, NULL, 0, |
|
|
29 |
MPI_DATATYPE_NULL, root, comm); |
|
|
30 |
|
31 |
/* gather at root packed messages */ |
|
|
32 |
MPI_Gatherv( lbuf, position, MPI_PACKED, NULL, |
|
|
33 |
NULL, NULL, NULL, root, comm); |
|
|
34 |
|
35 |
} else { /* root code */ |
|
|
36 |
/* gather sizes of all packed messages */ |
|
|
37 |
MPI_Gather( &position, 1, MPI_INT, counts, 1, |
|
|
38 |
MPI_INT, root, comm); |
|
39
40
41
42
43
44
45
46
47
48
/* gather all packed messages */ displs[0] = 0;
for (i=1; i < gsize; i++)
displs[i] = displs[i-1] + counts[i-1]; totalcount = displs[gsize-1] + counts[gsize-1]; rbuf = (char *)malloc(totalcount);
cbuf = (char *)malloc(totalcount); MPI_Gatherv( lbuf, position, MPI_PACKED, rbuf,
counts, displs, MPI_PACKED, root, comm);