The error message is self explanatory, the application calls MPI_Recv with
an invalid TAG. The MPI standard defines a valid tag as a positive integer
between 0 and the value of the MPI_UB_TAG attribute on MPI_COMM_WORLD. At
this point it seems plausible this is an application issue.

Check that the application is correctly restricting the tags used to the
valid range. If this is the case, then the issue might be on OMPI side in
which case a reproducer would be appreciated.

George.



On Sun, Sep 19, 2021 at 11:42 Feng Wade via users <users@lists.open-mpi.org>
wrote:

> Hi,
>
> Good morning.
>
> I am using openmpi/4.0.3 on Compute Canada to do 3D flow simulation. My
> grid size is Lx*Ly*Lz=700*169*500. It worked quite well for lower
> resolution. However, after increasing my resolution from Nx*Ny*Nz=64*109*62
> to 256*131*192, openmpi reported errors as shown below:
>
> [gra541:21749] *** An error occurred in MPI_Recv
> [gra541:21749] *** reported by process [2068774913,140]
> [gra541:21749] *** on communicator MPI COMMUNICATOR 13 DUP FROM 0
> [gra541:21749] *** MPI_ERR_TAG: invalid tag
> [gra541:21749] *** MPI_ERRORS_ARE_FATAL (processes in this communicator
> will now abort,
> [gra541:21749] ***    and potentially your MPI job)
> [gra529:07588] 210 more processes have sent help message
> help-mpi-errors.txt / mpi_errors_are_fatal
> [gra529:07588] Set MCA parameter "orte_base_help_aggregate" to 0 to see
> all help / error messages
>
> This is my computation parameters and command to run openmpi:
> #!/bin/bash
> #SBATCH --time=0-10:00:00
> #SBATCH --job-name=3D_EIT_Wi64
> #SBATCH --output=log-%j
> #SBATCH --ntasks=128
> #SBATCH --nodes=4
> #SBATCH --mem-per-cpu=4000M
> mpirun ./vepoiseuilleFD_5.x
>
> I guess The value of the PATH and LD_LIBRARY_PATH environment variables
> are all set correct because my simulation worked for lower resolution ones.
>
> Thank you for your time.
>
> Sincerely
>
> Wade
>

Reply via email to