Joseph,

Would you mind trying again with
export OMPI_MCA_osc=^pt2pt
export OMPI_MCA_osc_base_verbose=10

If it still does not work, then please post the output

Cheers,

Gilles

Joseph Schuchart <schuch...@hlrs.de> wrote:
>Hi Howard, 
>
>Thanks for your quick reply and your suggestions. I exported both variables as 
>you suggested but neither has any impact. The error message stays the same 
>with both env variables set. Is there any other way to get more information 
>from OpenMPI? 
>
>Sorry for not mentioning my OS. I'm running on a Linux Mint 18.1 with stock 
>kernel 4.8.0-36.
>
>Joseph
>
>On 02/18/2017 05:18 PM, Howard Pritchard wrote:
>
>Hi Joseph
>
>
>What OS are you using when running the test?
>
>
>Could you try running with
>
>
>export OMPI_mca_osc=^pt2pt
>
>and
>
>export OMPI_mca_osc_base_verbose=10
>
>
>This error message was put in to this OMPI release because this part of the 
>code has known problems when used multi threaded.
>
>
>
>
>Joseph Schuchart <schuch...@hlrs.de> schrieb am Sa. 18. Feb. 2017 um 04:02:
>
>All,
>
>I am seeing a fatal error with OpenMPI 2.0.2 if requesting support for
>MPI_THREAD_MULTIPLE and afterwards creating a window using
>MPI_Win_create. I am attaching a small reproducer. The output I get is
>the following:
>
>```
>MPI_THREAD_MULTIPLE supported: yes
>MPI_THREAD_MULTIPLE supported: yes
>MPI_THREAD_MULTIPLE supported: yes
>MPI_THREAD_MULTIPLE supported: yes
>--------------------------------------------------------------------------
>The OSC pt2pt component does not support MPI_THREAD_MULTIPLE in this
>release.
>Workarounds are to run on a single node, or to use a system with an RDMA
>capable network such as Infiniband.
>--------------------------------------------------------------------------
>[beryl:10705] *** An error occurred in MPI_Win_create
>[beryl:10705] *** reported by process [2149974017,2]
>[beryl:10705] *** on communicator MPI_COMM_WORLD
>[beryl:10705] *** MPI_ERR_WIN: invalid window
>[beryl:10705] *** MPI_ERRORS_ARE_FATAL (processes in this communicator
>will now abort,
>[beryl:10705] ***    and potentially your MPI job)
>[beryl:10698] 3 more processes have sent help message help-osc-pt2pt.txt
>/ mpi-thread-multiple-not-supported
>[beryl:10698] Set MCA parameter "orte_base_help_aggregate" to 0 to see
>all help / error messages
>[beryl:10698] 3 more processes have sent help message
>help-mpi-errors.txt / mpi_errors_are_fatal
>```
>
>I am running on a single node (my laptop). Both OpenMPI and the
>application were compiled using GCC 5.3.0. Naturally, there is no
>support for Infiniband available. Should I signal OpenMPI that I am
>indeed running on a single node? If so, how can I do that? Can't this be
>detected by OpenMPI automatically? The test succeeds if I only request
>MPI_THREAD_SINGLE.
>
>OpenMPI 2.0.2 has been configured using only
>--enable-mpi-thread-multiple and --prefix configure parameters. I am
>attaching the output of ompi_info.
>
>Please let me know if you need any additional information.
>
>Cheers,
>Joseph
>
>--
>Dipl.-Inf. Joseph Schuchart
>High Performance Computing Center Stuttgart (HLRS)
>Nobelstr. 19
>D-70569 Stuttgart
>
>Tel.: +49(0)711-68565890
>Fax: +49(0)711-6856832
>E-Mail: schuch...@hlrs.de
>
>_______________________________________________
>users mailing list
>users@lists.open-mpi.org
>https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>
>
>
>_______________________________________________ users mailing list 
>users@lists.open-mpi.org 
>https://rfd.newmexicoconsortium.org/mailman/listinfo/users 
>
>
>-- Dipl.-Inf. Joseph Schuchart High Performance Computing Center Stuttgart 
>(HLRS) Nobelstr. 19 D-70569 Stuttgart Tel.: +49(0)711-68565890 Fax: 
>+49(0)711-6856832 E-Mail: schuch...@hlrs.de 
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to