A MPI datatype is defined by its size and content, as well as its extent. When multiple elements of the same size are used in a contiguous manner (such in your scatter operation with a count > 1), the extent is used to compute where the next element will start. As, your extent is as big as the size of the derived datatype, the first elements of the second structure will start after the last element of the first structure. Which obviously will not do what you expect.
/* Create a type that span over the whole column */ MPI_Type_vector( N, 1, N, MPI_DOUBLE, &temp_type );/* Restrict the extent of this type, in such a way that when used with a count
* the extent of the whole column is just one element. */ MPI_Type_create_resized( temp_type, 0, sizeof(double), column_type ); MPI_Scatter( sbuf, 1, column_type, rbuf, 1, column_type, root, comm ); george. On Jun 16, 2008, at 2:36 AM, Siegmar Gross wrote:
Hi,I have a problem with derived data types and MPI_Scatter/MPI_Gather in C(Solaris 10 sparc, openmpi-1.2.4).I want to distribute the columns of a matrix with MPI_Scatter/ MPI_Gather. Unfortunately my program didn't work with my derived data type, so I used a 2x2 matrix to figure out what's wrong and added a printf statement after MPI_Scatter to print the contents of the column vector of each process. The process with rank 1 didn't get the values "2" and "4" (see below), but more or less 0. Now I used a 4x2 matrix and still a 2-element column vector (so I should see the upper 2x2 "matrix" in my columns) to get an idea which values are used for process 1. As you can see below it got "5" and "7", i.e. the values of the first column of the lower 2x2 matrix and not the values of the block which starts after the first element of the first block (a[2][0]instead of a[0][1]).Since I wasn't sure if I could use MPI_Type_struct I rewrote the programwith MPI_Type_vector. This time the result was better but still notsatisfying. Process 1 got values from the second column but one value toolate (starting with a[1][1] instead of a[1][0]).I run into the same problem with MPICH, MPICH2, and LAM-MPI, so that I have probably misunderstood something or have a programming error. Since I dont't know how to proceed, I would be very grateful if someone could tell me if I can't use a strided vector with MPI_Scatter or if it is eventually a bug in the implementations of the MPI libraries (how unlikely it is). Mainly I've implemented this part in the following way (the full program code is appended).int blockcounts[2] = {1, 1}; MPI_Datatype types[2] = {MPI_DOUBLE, MPI_UB}; MPI_Aint offsets[2] = {0, Q * sizeof (double)}; MPI_Datatype coltype; MPI_Type_struct (2, blockcounts, offsets, types, &coltype); MPI_Type_commit (&coltype); MPI_Scatter (matrix, P, coltype, column, P, MPI_DOUBLE, 0, MPI_COMM_WORLD); MPI_Type_struct =============== tyr e5 158 mpiexec -mca btl ^udapl -np 2 e5_1a original matrix: 1 2 3 4 rank: 0 c0: 1 c1: 3 rank: 1 c0: 5.51719e-313 c1: 4.24399e-314 tyr e5 160 mpiexec -mca btl ^udapl -np 2 e5_1a original matrix: 1 2 3 4 5 6 7 8 rank: 0 c0: 1 c1: 3 rank: 1 c0: 5 c1: 7 MPI_Type_vector =============== tyr e5 119 mpiexec -mca btl ^udapl -np 2 e5_1b original matrix: 1 2 3 4 5 6 7 8 rank: 0 c0: 1 c1: 3 rank: 1 c0: 4 c1: 6 Thank you very much for any help or suggestions in advance. Kind regards Siegmar #include <stdio.h> #include <stdlib.h> #include "mpi.h" #define P 4 /* # of rows */ #define Q 2 /* # of columns */ #define FAKTOR 2 /* multiplicator for col. elem. */ static void print_matrix (int p, int q, double **mat); int main (int argc, char *argv[]) { int ntasks, /* number of parallel tasks */ mytid, /* my task id */ namelen, /* length of processor name */ i, j, /* loop variables */ tmp; /* temporary value */ double matrix[P][Q], column[2]; /* column[P];*/ int blockcounts[2] = {1, 1}; MPI_Datatype types[2] = {MPI_DOUBLE, MPI_UB}; MPI_Aint offsets[2] = {0, Q * sizeof (double)}; MPI_Datatype coltype; MPI_Init (&argc, &argv); MPI_Comm_rank (MPI_COMM_WORLD, &mytid); MPI_Comm_size (MPI_COMM_WORLD, &ntasks);/* check that we have the correct number of processes in our universe */if (mytid == 0) { if (ntasks != Q) { printf ("\n\nWe need exactly %d processes.\n" "Usage: mpirun -w -np %d N %s\n\n\n", Q, Q, argv[0]); } } if (ntasks != Q) { MPI_Finalize (); exit (EXIT_FAILURE); } /* build the new type for a strided vector */ MPI_Type_struct (2, blockcounts, offsets, types, &coltype); MPI_Type_commit (&coltype); if (mytid == 0) { tmp = 1; for (i = 0; i < P; ++i) /* initialize matrix */ { for (j = 0; j < Q; ++j) { matrix[i][j] = tmp++; } } printf ("\n\noriginal matrix:\n\n"); print_matrix (P, Q, (double **) matrix); } /* distribute columns */ MPI_Scatter (matrix, 2, coltype, column, 2, MPI_DOUBLE, 0, /* MPI_Scatter (matrix, P, coltype, column, P, MPI_DOUBLE, 0,*/ MPI_COMM_WORLD); printf ("rank: %d c0: %g c1: %g\n", mytid, column[0], column[1]); for (i = 0; i < 2; ++i) /* for (i = 0; i < P; ++i)*/ { if ((mytid % 2) == 0) { column[i] = column[i] * column[i]; } else { column[i] = column[i] * FAKTOR; } } /* wait for result vectors */ MPI_Gather (column, 2, MPI_DOUBLE, matrix, 2, coltype, 0, /* MPI_Gather (column, P, MPI_DOUBLE, matrix, P, coltype, 0,*/ MPI_COMM_WORLD); if (mytid == 0) { printf ("\n\nresult matrix:\n" "(odd columns: elements squared; even columns: elements " "multiplied with %d)\n\n", FAKTOR); print_matrix (P, Q, (double **) matrix); } MPI_Type_free (&coltype); MPI_Finalize (); return EXIT_SUCCESS; } /* print the values of a matrix * * input parameters: p number of rows * q number of columns * mat 2D-matrix of "double" values * output parameters: none * return value: none * side effects: none * */ void print_matrix (int p, int q, double **mat) { int i, j; /* loop variables */ for (i = 0; i < p; ++i) { for (j = 0; j < q; ++j) { printf ("%6g", *((double *) mat + i * q + j)); } printf ("\n"); } printf ("\n"); } #include <stdio.h> #include <stdlib.h> #include "mpi.h" #define P 4 /* # of rows */ #define Q 2 /* # of columns */ #define FAKTOR 2 /* multiplicator for col. elem. */ static void print_matrix (int p, int q, double **mat); int main (int argc, char *argv[]) { int ntasks, /* number of parallel tasks */ mytid, /* my task id */ namelen, /* length of processor name */ i, j, /* loop variables */ tmp; /* temporary value */ double matrix[P][Q], column[2]; MPI_Datatype coltype; MPI_Init (&argc, &argv); MPI_Comm_rank (MPI_COMM_WORLD, &mytid); MPI_Comm_size (MPI_COMM_WORLD, &ntasks);/* check that we have the correct number of processes in our universe */if (mytid == 0) { if (ntasks != Q) { printf ("\n\nWe need exactly %d processes.\n" "Usage: mpirun -w -np %d N %s\n\n\n", Q, Q, argv[0]); } } if (ntasks != Q) { MPI_Finalize (); exit (EXIT_FAILURE); } /* build the new type for a strided vector */ MPI_Type_vector (2, 1, Q, MPI_DOUBLE, &coltype); MPI_Type_commit (&coltype); if (mytid == 0) { tmp = 1; for (i = 0; i < P; ++i) /* initialize matrix */ { for (j = 0; j < Q; ++j) { matrix[i][j] = tmp++; } } printf ("\n\noriginal matrix:\n\n"); print_matrix (P, Q, (double **) matrix); } /* distribute columns */ MPI_Scatter (matrix, 1, coltype, column, 2, MPI_DOUBLE, 0, MPI_COMM_WORLD); printf ("rank: %d c0: %g c1: %g\n", mytid, column[0], column[1]); for (i = 0; i < 2; ++i) { if ((mytid % 2) == 0) { column[i] = column[i] * column[i]; } else { column[i] = column[i] * FAKTOR; } } /* wait for result vectors */ MPI_Gather (column, 2, MPI_DOUBLE, matrix, 1, coltype, 0, MPI_COMM_WORLD); if (mytid == 0) { printf ("\n\nresult matrix:\n" "(odd columns: elements squared; even columns: elements " "multiplied with %d)\n\n", FAKTOR); print_matrix (P, Q, (double **) matrix); } MPI_Type_free (&coltype); MPI_Finalize (); return EXIT_SUCCESS; } /* print the values of a matrix * * input parameters: p number of rows * q number of columns * mat 2D-matrix of "double" values * output parameters: none * return value: none * side effects: none * */ void print_matrix (int p, int q, double **mat) { int i, j; /* loop variables */ for (i = 0; i < p; ++i) { for (j = 0; j < q; ++j) { printf ("%6g", *((double *) mat + i * q + j)); } printf ("\n"); } printf ("\n"); } _______________________________________________ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
smime.p7s
Description: S/MIME cryptographic signature