The displs
argument from MPI_Scatterv()
function is said to be an "integer array (of length group size). Entry i specifies the displacement (relative to sendbuf from which to take the outgoing data to process i".
Let's say then that I have sendcounts
argument
int sendcounts[7] = {3, 3, 3, 3, 4, 4, 4};
The way I'm reasoning this out is that the displs
array should always start with value of 0 since the first entry's displacement is 0 relative to sendbuf
, so in my example above, displs
should look like:
int displs[7] = {0, 3, 6, 9, 13, 17, 21};
Is that correct? I know this is a trivial question, but for some reason the web does not help at all. There are no good examples out there, hence my question.
Yes, displacements gives the root information the information as to which items to send to a particular task - the offset of the starting item. So in most simple cases (e.g., you'd use MPI_Scatter
but the counts don't evenly divide) this can be immediately calculated from counts information:
displs[0] = 0; // offsets into the global array
for (size_t i=1; i<comsize; i++)
displs[i] = displs[i-1] + counts[i-1];
But it doesn't need to be that way; the only restriction is that the data you're sending can't overlap. You could count from the back just as well:
displs[0] = globalsize - counts[0];
for (size_t i=1; i<comsize; i++)
displs[i] = displs[i-1] - counts[i];
or any arbitrary order would work as well.
And in general the calculations can be more complicated because the types of the send buffer and receive buffers have to be consistent but not necessarily the same - you often get this if you're sending multidimensional array slices, for instance.
As an example of the simple cases, the below does the forward and backward cases:
#include <iostream>
#include <vector>
#include "mpi.h"
int main(int argc, char **argv) {
const int root = 0; // the processor with the initial global data
size_t globalsize;
std::vector<char> global; // only root has this
const size_t localsize = 2; // most ranks will have 2 items; one will have localsize+1
char local[localsize+2]; // everyone has this
int mynum; // how many items
MPI_Init(&argc, &argv);
int comrank, comsize;
MPI_Comm_rank(MPI_COMM_WORLD, &comrank);
MPI_Comm_size(MPI_COMM_WORLD, &comsize);
// initialize global vector
if (comrank == root) {
globalsize = comsize*localsize + 1;
for (size_t i=0; i<globalsize; i++)
global.push_back('a'+i);
}
// initialize local
for (size_t i=0; i<localsize+1; i++)
local[i] = '-';
local[localsize+1] = '\0';
int counts[comsize]; // how many pieces of data everyone has
for (size_t i=0; i<comsize; i++)
counts[i] = localsize;
counts[comsize-1]++;
mynum = counts[comrank];
int displs[comsize];
if (comrank == 0)
std::cout << "In forward order" << std::endl;
displs[0] = 0; // offsets into the global array
for (size_t i=1; i<comsize; i++)
displs[i] = displs[i-1] + counts[i-1];
MPI_Scatterv(global.data(), counts, displs, MPI_CHAR, // For root: proc i gets counts[i] MPI_CHARAs from displs[i]
local, mynum, MPI_CHAR, // I'm receiving mynum MPI_CHARs into local */
root, MPI_COMM_WORLD); // Task (root, MPI_COMM_WORLD) is the root
local[mynum] = '\0';
std::cout << comrank << " " << local << std::endl;
std::cout.flush();
if (comrank == 0)
std::cout << "In reverse order" << std::endl;
displs[0] = globalsize - counts[0];
for (size_t i=1; i<comsize; i++)
displs[i] = displs[i-1] - counts[i];
MPI_Scatterv(global.data(), counts, displs, MPI_CHAR, // For root: proc i gets counts[i] MPI_CHARAs from displs[i]
local, mynum, MPI_CHAR, // I'm receiving mynum MPI_CHARs into local */
root, MPI_COMM_WORLD); // Task (root, MPI_COMM_WORLD) is the root
local[mynum] = '\0';
std::cout << comrank << " " << local << std::endl;
MPI_Finalize();
}
Running gives:
In forward order
0 ab
1 cd
2 ef
3 ghi
In reverse order
0 hi
1 fg
2 de
3 abc