Deal all,
probably I have found the error.
Let's me check. Probably I have not properly set-up colors.

Thanks a lot,
I hope that you have not lost too much time for me,
I will let you know If that was the problem.

Thanks again

Diego


On 3 August 2018 at 19:57, Diego Avesani <diego.aves...@gmail.com> wrote:

> Dear R, Dear all,
>
> I do not know.
> I have isolated the issues. It seem that I have some problem with:
>   CALL MPI_COMM_SPLIT(MPI_COMM_WORLD,colorl,MPIworld%rank,MPI_
> LOCAL_COMM,MPIworld%iErr)
>   CALL MPI_COMM_RANK(MPI_LOCAL_COMM, MPIlocal%rank,MPIlocal%iErr)
>   CALL MPI_COMM_SIZE(MPI_LOCAL_COMM, MPIlocal%nCPU,MPIlocal%iErr)
>
> openMPI seems not able to create properly MPIlocal%rank.
>
> what should be? a bug?
>
> thanks again
>
> Diego
>
>
> On 3 August 2018 at 19:47, Ralph H Castain <r...@open-mpi.org> wrote:
>
>> Those two command lines look exactly the same to me - what am I missing?
>>
>>
>> On Aug 3, 2018, at 10:23 AM, Diego Avesani <diego.aves...@gmail.com>
>> wrote:
>>
>> Dear all,
>>
>> I am experiencing a strange error.
>>
>> In my code I use three group communications:
>> MPI_COMM_WORLD
>> MPI_MASTERS_COMM
>> LOCAL_COMM
>>
>> which have in common some CPUs.
>>
>> when I run my code as
>>  mpirun -np 4 --oversubscribe ./MPIHyperStrem
>>
>> I have no problem, while when I run it as
>>
>>  mpirun -np 4 --oversubscribe ./MPIHyperStrem
>>
>> sometimes it crushes and sometimes not.
>>
>> It seems that all is linked to
>> CALL MPI_REDUCE(QTS(tstep,:), QTS(tstep,:), nNode, MPI_DOUBLE_PRECISION,
>> MPI_SUM, 0, MPI_LOCAL_COMM, iErr)
>>
>> which works with in local.
>>
>> What do you think? Can you please suggestion some debug test?
>> Is a problem related to local communications?
>>
>> Thanks
>>
>>
>>
>> Diego
>>
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
>>
>>
>>
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
>>
>
>
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to