Joseph, Would you mind trying again with export OMPI_MCA_osc=^pt2pt export OMPI_MCA_osc_base_verbose=10
If it still does not work, then please post the output Cheers, Gilles Joseph Schuchart <schuch...@hlrs.de> wrote: >Hi Howard, > >Thanks for your quick reply and your suggestions. I exported both variables as >you suggested but neither has any impact. The error message stays the same >with both env variables set. Is there any other way to get more information >from OpenMPI? > >Sorry for not mentioning my OS. I'm running on a Linux Mint 18.1 with stock >kernel 4.8.0-36. > >Joseph > >On 02/18/2017 05:18 PM, Howard Pritchard wrote: > >Hi Joseph > > >What OS are you using when running the test? > > >Could you try running with > > >export OMPI_mca_osc=^pt2pt > >and > >export OMPI_mca_osc_base_verbose=10 > > >This error message was put in to this OMPI release because this part of the >code has known problems when used multi threaded. > > > > >Joseph Schuchart <schuch...@hlrs.de> schrieb am Sa. 18. Feb. 2017 um 04:02: > >All, > >I am seeing a fatal error with OpenMPI 2.0.2 if requesting support for >MPI_THREAD_MULTIPLE and afterwards creating a window using >MPI_Win_create. I am attaching a small reproducer. The output I get is >the following: > >``` >MPI_THREAD_MULTIPLE supported: yes >MPI_THREAD_MULTIPLE supported: yes >MPI_THREAD_MULTIPLE supported: yes >MPI_THREAD_MULTIPLE supported: yes >-------------------------------------------------------------------------- >The OSC pt2pt component does not support MPI_THREAD_MULTIPLE in this >release. >Workarounds are to run on a single node, or to use a system with an RDMA >capable network such as Infiniband. >-------------------------------------------------------------------------- >[beryl:10705] *** An error occurred in MPI_Win_create >[beryl:10705] *** reported by process [2149974017,2] >[beryl:10705] *** on communicator MPI_COMM_WORLD >[beryl:10705] *** MPI_ERR_WIN: invalid window >[beryl:10705] *** MPI_ERRORS_ARE_FATAL (processes in this communicator >will now abort, >[beryl:10705] *** and potentially your MPI job) >[beryl:10698] 3 more processes have sent help message help-osc-pt2pt.txt >/ mpi-thread-multiple-not-supported >[beryl:10698] Set MCA parameter "orte_base_help_aggregate" to 0 to see >all help / error messages >[beryl:10698] 3 more processes have sent help message >help-mpi-errors.txt / mpi_errors_are_fatal >``` > >I am running on a single node (my laptop). Both OpenMPI and the >application were compiled using GCC 5.3.0. Naturally, there is no >support for Infiniband available. Should I signal OpenMPI that I am >indeed running on a single node? If so, how can I do that? Can't this be >detected by OpenMPI automatically? The test succeeds if I only request >MPI_THREAD_SINGLE. > >OpenMPI 2.0.2 has been configured using only >--enable-mpi-thread-multiple and --prefix configure parameters. I am >attaching the output of ompi_info. > >Please let me know if you need any additional information. > >Cheers, >Joseph > >-- >Dipl.-Inf. Joseph Schuchart >High Performance Computing Center Stuttgart (HLRS) >Nobelstr. 19 >D-70569 Stuttgart > >Tel.: +49(0)711-68565890 >Fax: +49(0)711-6856832 >E-Mail: schuch...@hlrs.de > >_______________________________________________ >users mailing list >users@lists.open-mpi.org >https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > > >_______________________________________________ users mailing list >users@lists.open-mpi.org >https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > >-- Dipl.-Inf. Joseph Schuchart High Performance Computing Center Stuttgart >(HLRS) Nobelstr. 19 D-70569 Stuttgart Tel.: +49(0)711-68565890 Fax: >+49(0)711-6856832 E-Mail: schuch...@hlrs.de
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users