Hi Gilles and Howard,

The use of MPI_Dist_graph_create_adjacent solves the issue :)

Thanks for your help!

Best reagrds,
Ghislain

2014-11-21 7:23 GMT+01:00 Gilles Gouaillardet <gilles.gouaillar...@iferc.org
>:

>  Hi Ghislain,
>
> that sound like a but in MPI_Dist_graph_create :-(
>
> you can use MPI_Dist_graph_create_adjacent instead :
>
> MPI_Dist_graph_create_adjacent(MPI_COMM_WORLD, degrees, &targets[0],
> &weights[0],
>                         degrees, &targets[0], &weights[0], info,
> rankReordering, &commGraph);
>
> it does not crash and as far as i understand, it produces correct results,
>
> according the the mpi standard (example 7.3) that should do the same
> thing, that's why
> i think there is a bug in MPI_Dist_graph_create
>
> Cheers,
>
> Gilles
>
>
>
>
> On 2014/11/21 2:21, Howard Pritchard wrote:
>
> Hi Ghislain,
>
> I tried to run your test with mvapich 1.9 and get a "message truncated"
> failure at three ranks.
>
> Howard
>
>
> 2014-11-20 8:51 GMT-07:00 Ghislain Viguier <ghislain.vigu...@gmail.com> 
> <ghislain.vigu...@gmail.com>:
>
>
>  Dear support,
>
> I'm encountering an issue with the MPI_Neighbor_alltoallw request of
> mpi-1.8.3.
> I have enclosed a test case with information of my workstation.
>
> In this test, I define a weighted topology for 5 processes, where the
> weight represent the number of buffers to send/receive :
>     rank
>       0 : | x |
>       1 : | 2 | x |
>       2 : | 1 | 1 | x |
>       3 : | 3 | 2 | 3 | x |
>       4 : | 5 | 2 | 2 | 2 | x |
>
> In this topology, the rank 1 will send/receive :
>    2 buffers to/from the rank 0,
>    1 buffer to/from the rank 2,
>    2 buffers to/from the rank 3,
>    2 buffers to/from the rank 4,
>
> The send buffer are defined with the MPI_Type_create_hindexed_block. This
> allows to use a same buffer for several communications without duplicating
> it (read only).
> Here the rank 1 will have 2 send buffers (the max of 2, 1, 2, 2).
> The receiver buffer is a contiguous buffer defined with
> MPI_Type_contiguous request.
> Here, the receiver buffer of the rank 1 is of size : 7 (2+1+2+2)
>
> This test case succesful for 2 or 3 processes. For 4 processes, the test
> fails 1 times for 3 successes. For 5 processes, the test fails all the time.
>
> The error code is : *** MPI_ERR_IN_STATUS: error code in status
>
> I don't understand what I am doing wrong.
>
> Could you please have a look on it?
>
> Thank you very much.
>
> Best regards,
> Ghislain Viguier
>
> --
> Ghislain Viguier
> Tél. 06 31 95 03 17
>
> _______________________________________________
> users mailing listus...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this 
> post:http://www.open-mpi.org/community/lists/users/2014/11/25850.php
>
>
>
> _______________________________________________
> users mailing listus...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
>
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2014/11/25852.php
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
> Link to this post:
> http://www.open-mpi.org/community/lists/users/2014/11/25853.php
>



-- 
Ghislain Viguier
Tél. 06 31 95 03 17

Reply via email to