Re: [OMPI users] mpirun on Kubuntu 20.4.1 hangs
Hello, In spite of the delay, I was not able to solve my problem. Thanks to Joseph and Prentice for their interesting suggestions. I uninstalled AppAmor (SElinux is not installed ) as suggested by Prentice but there were no changes, mpirun sttill hangs. The result of gdb stack trace is the following: $ sudo gdb -batch -ex "thread apply all bt" -p $(ps -C mpirun -o pid= | head -n 1) [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". 0x7f9289544307 in __libc_connect (fd=9, addr=..., len=16) at ../sysdeps/unix/sysv/linux/connect.c:26 26 ../sysdeps/unix/sysv/linux/connect.c: Aucun fichier ou dossier de ce type. Thread 1 (Thread 0x7f92891f4e80 (LWP 4948)): #0 0x7f9289544307 in __libc_connect (fd=9, addr=..., len=16) at ../sysdeps/unix/sysv/linux/connect.c:26 #1 0x7f9288fff59d in ?? () from /lib/x86_64-linux-gnu/libxcb.so.1 #2 0x7f9288fffc49 in xcb_connect_to_display_with_auth_info () from /lib/x86_64-linux-gnu/libxcb.so.1 #3 0x7f928906cb7a in _XConnectXCB () from /lib/x86_64-linux-gnu/libX11.so.6 #4 0x7f928905d319 in XOpenDisplay () from /lib/x86_64-linux-gnu/libX11.so.6 #5 0x7f92897de4fb in ?? () from /usr/lib/x86_64-linux-gnu/hwloc/hwloc_gl.so #6 0x7f92893b901e in ?? () from /lib/x86_64-linux-gnu/libhwloc.so.15 #7 0x7f92893c13a0 in hwloc_topology_load () from /lib/x86_64-linux-gnu/libhwloc.so.15 #8 0x7f92896df564 in opal_hwloc_base_get_topology () from /lib/x86_64-linux-gnu/libopen-pal.so.40 #9 0x7f92891da6be in ?? () from /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_ess_hnp.so #10 0x7f92897a22fc in orte_init () from /lib/x86_64-linux-gnu/libopen-rte.so.40 #11 0x7f92897a6c86 in orte_submit_init () from /lib/x86_64-linux-gnu/libopen-rte.so.40 #12 0x55819fd0b3a3 in ?? () #13 0x7f92894480b3 in __libc_start_main (main=0x55819fd0b1c0, argc=1, argv=0x7fff5334fe48, init=, fini=, rtld_fini=, stack_end=0x7fff5334fe38) at ../csu/libc-start.c:308 #14 0x55819fd0b1fe in ?? () [Inferior 1 (process 4948) detached] So it seems to be a problem in the connection via libxcb (socket ?) but this is out of my system computer skills.. Is there any authorization needed? As is libX11 at the origin of the call I tried to execute in a bare terminal (ctrl-alt-f2 and via ssh) but the message is the same. I tried to recompile/install the hole package and have the same result. Thank you for your help. Jorge Le 22/10/2020 à 12:16, Joseph Schuchart via users a écrit : Hi Jorge, Can you try to get a stack trace of mpirun using the following command in a separate terminal? sudo gdb -batch -ex "thread apply all bt" -p $(ps -C mpirun -o pid= | head -n 1) Maybe that will give some insight where mpirun is hanging. Cheers, Joseph On 10/21/20 9:58 PM, Jorge SILVA via users wrote: Hello Jeff, The program is not executed, seems waits for something to connect with (why twice ctrl-C ?) jorge@gcp26:~/MPIRUN$ mpirun -np 1 touch /tmp/foo ^C^C jorge@gcp26:~/MPIRUN$ ls -l /tmp/foo ls: impossible d'accéder à '/tmp/foo': Aucun fichier ou dossier de ce type no file is created.. In fact, my question was if are there differences in mpirun usage between these versions.. The mpirun -help gives a different output as expected, but I tried a lot of options without any success. Le 21/10/2020 à 21:16, Jeff Squyres (jsquyres) a écrit : There's huge differences between Open MPI v2.1.1 and v4.0.3 (i.e., years of development effort); it would be very hard to categorize them all; sorry! What happens if you mpirun -np 1 touch /tmp/foo (Yes, you can run non-MPI apps through mpirun) Is /tmp/foo created? (i.e., did the job run, and mpirun is somehow not terminating) On Oct 21, 2020, at 12:22 PM, Jorge SILVA via users mailto:users@lists.open-mpi.org>> wrote: Hello Gus, Thank you for your answer.. Unfortunately my problem is much more basic. I didn't try to run the program in both computers , but just to run something in one computer. I just installed the new OS an openmpi in two different computers, in the standard way, with the same result. For example: In kubuntu20.4.1 LTS with openmpi 4.0.3-0ubuntu jorge@gcp26:~/MPIRUN$ cat hello.f90 print*,"Hello World!" end jorge@gcp26:~/MPIRUN$ mpif90 hello.f90 -o hello jorge@gcp26:~/MPIRUN$ ./hello Hello World! jorge@gcp26:~/MPIRUN$ mpirun -np 1 hello <---here the program hangs with no output ^C^Cjorge@gcp26:~/MPIRUN$ The mpirun task sleeps with no output, and only twice ctrl-C ends the execution : jorge 5540 0.1 0.0 44768 8472 pts/8 S+ 17:54 0:00 mpirun -np 1 hello In kubuntu 18.04.5 LTS with openmpi 2.1.1, of course, the same program gives jorge@gcp30:~/MPIRUN$ cat hello.f90 print*, "Hello World!" END jorge@gcp30:~/MPIRUN$ mpif90 hello.f90 -o hello jorge@gcp30:~/MPIRUN$ ./hello Hello World! jorge@gcp30:~/MPIRUN$ mpirun -np 1 hello Hello Worl
Re: [OMPI users] mpirun on Kubuntu 20.4.1 hangs
Sorry, if I execute mpirun in a *really *bare terminal, without X Server running it works! but with an error message : Invalid MIT-MAGIC-COOKIE-1 key So the problem is related to X, but I have still no solution Jorge Le 14/11/2020 à 12:33, Jorge Silva via users a écrit : Hello, In spite of the delay, I was not able to solve my problem. Thanks to Joseph and Prentice for their interesting suggestions. I uninstalled AppAmor (SElinux is not installed ) as suggested by Prentice but there were no changes, mpirun sttill hangs. The result of gdb stack trace is the following: $ sudo gdb -batch -ex "thread apply all bt" -p $(ps -C mpirun -o pid= | head -n 1) [Thread debugging using libthread_db enabled] Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1". 0x7f9289544307 in __libc_connect (fd=9, addr=..., len=16) at ../sysdeps/unix/sysv/linux/connect.c:26 26 ../sysdeps/unix/sysv/linux/connect.c: Aucun fichier ou dossier de ce type. Thread 1 (Thread 0x7f92891f4e80 (LWP 4948)): #0 0x7f9289544307 in __libc_connect (fd=9, addr=..., len=16) at ../sysdeps/unix/sysv/linux/connect.c:26 #1 0x7f9288fff59d in ?? () from /lib/x86_64-linux-gnu/libxcb.so.1 #2 0x7f9288fffc49 in xcb_connect_to_display_with_auth_info () from /lib/x86_64-linux-gnu/libxcb.so.1 #3 0x7f928906cb7a in _XConnectXCB () from /lib/x86_64-linux-gnu/libX11.so.6 #4 0x7f928905d319 in XOpenDisplay () from /lib/x86_64-linux-gnu/libX11.so.6 #5 0x7f92897de4fb in ?? () from /usr/lib/x86_64-linux-gnu/hwloc/hwloc_gl.so #6 0x7f92893b901e in ?? () from /lib/x86_64-linux-gnu/libhwloc.so.15 #7 0x7f92893c13a0 in hwloc_topology_load () from /lib/x86_64-linux-gnu/libhwloc.so.15 #8 0x7f92896df564 in opal_hwloc_base_get_topology () from /lib/x86_64-linux-gnu/libopen-pal.so.40 #9 0x7f92891da6be in ?? () from /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_ess_hnp.so #10 0x7f92897a22fc in orte_init () from /lib/x86_64-linux-gnu/libopen-rte.so.40 #11 0x7f92897a6c86 in orte_submit_init () from /lib/x86_64-linux-gnu/libopen-rte.so.40 #12 0x55819fd0b3a3 in ?? () #13 0x7f92894480b3 in __libc_start_main (main=0x55819fd0b1c0, argc=1, argv=0x7fff5334fe48, init=, fini=out>, rtld_fini=, stack_end=0x7fff5334fe38) at ../csu/libc-start.c:308 #14 0x55819fd0b1fe in ?? () [Inferior 1 (process 4948) detached] So it seems to be a problem in the connection via libxcb (socket ?) but this is out of my system computer skills.. Is there any authorization needed? As is libX11 at the origin of the call I tried to execute in a bare terminal (ctrl-alt-f2 and via ssh) but the message is the same. I tried to recompile/install the hole package and have the same result. Thank you for your help. Jorge Le 22/10/2020 à 12:16, Joseph Schuchart via users a écrit : Hi Jorge, Can you try to get a stack trace of mpirun using the following command in a separate terminal? sudo gdb -batch -ex "thread apply all bt" -p $(ps -C mpirun -o pid= | head -n 1) Maybe that will give some insight where mpirun is hanging. Cheers, Joseph On 10/21/20 9:58 PM, Jorge SILVA via users wrote: Hello Jeff, The program is not executed, seems waits for something to connect with (why twice ctrl-C ?) jorge@gcp26:~/MPIRUN$ mpirun -np 1 touch /tmp/foo ^C^C jorge@gcp26:~/MPIRUN$ ls -l /tmp/foo ls: impossible d'accéder à '/tmp/foo': Aucun fichier ou dossier de ce type no file is created.. In fact, my question was if are there differences in mpirun usage between these versions.. The mpirun -help gives a different output as expected, but I tried a lot of options without any success. Le 21/10/2020 à 21:16, Jeff Squyres (jsquyres) a écrit : There's huge differences between Open MPI v2.1.1 and v4.0.3 (i.e., years of development effort); it would be very hard to categorize them all; sorry! What happens if you mpirun -np 1 touch /tmp/foo (Yes, you can run non-MPI apps through mpirun) Is /tmp/foo created? (i.e., did the job run, and mpirun is somehow not terminating) On Oct 21, 2020, at 12:22 PM, Jorge SILVA via users mailto:users@lists.open-mpi.org>> wrote: Hello Gus, Thank you for your answer.. Unfortunately my problem is much more basic. I didn't try to run the program in both computers , but just to run something in one computer. I just installed the new OS an openmpi in two different computers, in the standard way, with the same result. For example: In kubuntu20.4.1 LTS with openmpi 4.0.3-0ubuntu jorge@gcp26:~/MPIRUN$ cat hello.f90 print*,"Hello World!" end jorge@gcp26:~/MPIRUN$ mpif90 hello.f90 -o hello jorge@gcp26:~/MPIRUN$ ./hello Hello World! jorge@gcp26:~/MPIRUN$ mpirun -np 1 hello <---here the program hangs with no output ^C^Cjorge@gcp26:~/MPIRUN$ The mpirun task sleeps with no output, and only twice ctrl-C ends the execution : jorge 5540 0.1 0.0 44768 8472 pts/8 S+ 17:54 0:00 mpirun -np 1 hello In k
[OMPI users] PRRTE DVM: how to tell prun to not share nodes among prun jobs?
Hi, in context of the PRRTE Distributed Virtual Machine, is there a way to tell the task mapper inside prun to not share a node across separate prun jobs? For example, inside a resource allocation from Cobalt/ALPS: 2 nodes with 64 cores each: prte --daemonize prun ... & ... prun ... & pterm Scenario A: $ prun --map-by ppr:64:node -n 64 ./mpitest & $ prun --map-by ppr:64:node -n 64 ./mpitest & MPI World size = 64 processes Hello World from rank 0 running on nid03834 (hostname nid03834)! ... Hello World from rank 63 running on nid03834 (hostname nid03834)! MPI World size = 64 processes Hello World from rank 0 running on nid03835 (hostname nid03835)! ... Hello World from rank 63 running on nid03835 (hostname nid03835)! Scenario B: $ prun --map-by ppr:64:node -n 1 ./mpitest & $ prun --map-by ppr:64:node -n 1 ./mpitest & MPI World size = 1 processes Hello World from rank 0 running on nid03834 (hostname nid03834)! MPI World size = 1 processes Hello World from rank 0 running on nid03834 (hostname nid03834)! The question is: in Scneario B, how to tell prun that node nid03834 should not be used for the second prun job, because this node is already (partially) occupied by a different prun instance job? Scenario A implies that the DVM already tracks occupancy, so the question is just how to tell the mapper to treat a free core on a free node differently from a free core on a partially occupied node. The --map-by :NOOVERSUBSCRIBE does not look like the answer since there's no oversubscription of cores, right? Would need something like --map-by :exclusive:node? If not supported, how hard would it be for me to patch? Potential workarounds I can think of is to fill the unoccupied cores on partially occupied nodes with dummy jobs with --host pointing to the partially occupied nodes and a -n count matching the number of unoccupied cores, but is this even doable? also requires dumping the mapping from each prun which I am unable to achive with --map-by :DISPLAY (works with mpirun but not with prun). Or, run a Flux instance [1] instead of the PRRTE DVM on the resource allocation, which seems similar but features a scheduler with a queue (a feature proposed for the PRRTE DVM on the list earlier [1]). I am guessing that Flux has the flexibility to this exclusive node mapping, but not sure. The DVM is proving to be very useful to deal with restrictions on minimum nodecount per job on some HPC clusters, by batching many small jobs into one job. A queue would be even more useful, but even without a queue it is still useful for batching sets of jobs which are known to fit on an allocation in parallel (i.e. without having to wait at all). [1] https://flux-framework.readthedocs.io/en/latest/quickstart.html [2] https://www.mail-archive.com/users@lists.open-mpi.org/msg30692.html OpenMPI: commit 7a922c8774b184ecb3aa1cd06720390bd9200b50 Fri Nov 6 08:48:29 2020 -0800 PRRTE: commit 37dd45c4d9fe973df1000f1a1421c2718fd80050 Fri Nov 6 12:45:38 2020 -0600 Thank you.
Re: [OMPI users] PRRTE DVM: how to tell prun to not share nodes among prun jobs?
IIRC, the correct syntax is: prun -host +e ... This tells PRRTE that you want empty nodes for this application. You can even specify how many empty nodes you want: prun -host +e:2 ... I haven't tested that in a bit, so please let us know if it works or not so we can fix it if necessary. As for the queue - we do plan to add a queue to PRRTE in 1st quarter next year. Wasn't really thinking of a true scheduler - just a FIFO queue for now. > On Nov 14, 2020, at 11:52 AM, Alexei Colin via users > wrote: > > Hi, in context of the PRRTE Distributed Virtual Machine, is there a way > to tell the task mapper inside prun to not share a node across separate > prun jobs? > > For example, inside a resource allocation from Cobalt/ALPS: 2 nodes with > 64 cores each: > > prte --daemonize > prun ... & > ... > prun ... & > pterm > > Scenario A: > > $ prun --map-by ppr:64:node -n 64 ./mpitest & > $ prun --map-by ppr:64:node -n 64 ./mpitest & > > MPI World size = 64 processes > Hello World from rank 0 running on nid03834 (hostname nid03834)! > ... > Hello World from rank 63 running on nid03834 (hostname nid03834)! > > MPI World size = 64 processes > Hello World from rank 0 running on nid03835 (hostname nid03835)! > ... > Hello World from rank 63 running on nid03835 (hostname nid03835)! > > Scenario B: > > $ prun --map-by ppr:64:node -n 1 ./mpitest & > $ prun --map-by ppr:64:node -n 1 ./mpitest & > > MPI World size = 1 processes > Hello World from rank 0 running on nid03834 (hostname nid03834)! > > MPI World size = 1 processes > Hello World from rank 0 running on nid03834 (hostname nid03834)! > > The question is: in Scneario B, how to tell prun that node nid03834 > should not be used for the second prun job, because this node is already > (partially) occupied by a different prun instance job? > > Scenario A implies that the DVM already tracks occupancy, so the > question is just how to tell the mapper to treat a free core on a free > node differently from a free core on a partially occupied node. The > --map-by :NOOVERSUBSCRIBE does not look like the answer since there's > no oversubscription of cores, right? Would need something like --map-by > :exclusive:node? If not supported, how hard would it be for me to patch? > > Potential workarounds I can think of is to fill the unoccupied cores on > partially occupied nodes with dummy jobs with --host pointing to the > partially occupied nodes and a -n count matching the number of > unoccupied cores, but is this even doable? also requires dumping the > mapping from each prun which I am unable to achive with --map-by > :DISPLAY (works with mpirun but not with prun). > > Or, run a Flux instance [1] instead of the PRRTE DVM on the resource > allocation, which seems similar but features a scheduler with a queue (a > feature proposed for the PRRTE DVM on the list earlier [1]). I am > guessing that Flux has the flexibility to this exclusive node mapping, > but not sure. > > The DVM is proving to be very useful to deal with restrictions on > minimum nodecount per job on some HPC clusters, by batching many small > jobs into one job. A queue would be even more useful, but even without a > queue it is still useful for batching sets of jobs which are known to > fit on an allocation in parallel (i.e. without having to wait at all). > > [1] https://flux-framework.readthedocs.io/en/latest/quickstart.html > [2] https://www.mail-archive.com/users@lists.open-mpi.org/msg30692.html > > OpenMPI: commit 7a922c8774b184ecb3aa1cd06720390bd9200b50 > Fri Nov 6 08:48:29 2020 -0800 > PRRTE: commit 37dd45c4d9fe973df1000f1a1421c2718fd80050 > Fri Nov 6 12:45:38 2020 -0600 > > Thank you.
Re: [OMPI users] PRRTE DVM: how to tell prun to not share nodes among prun jobs?
On Sat, Nov 14, 2020 at 08:07:47PM +, Ralph Castain via users wrote: > IIRC, the correct syntax is: > > prun -host +e ... > > This tells PRRTE that you want empty nodes for this application. You can even > specify how many empty nodes you want: > > prun -host +e:2 ... > > I haven't tested that in a bit, so please let us know if it works or not so > we can fix it if necessary. Works! Thank you. $ prun --map-by ppr:64:node --host +e -n 1 ./mpitest & $ prun --map-by ppr:64:node --host +e -n 1 ./mpitest & MPI World size = 1 processes Hello World from rank 0 running on nid03835 (hostname nid03835)! MPI World size = 1 processes Hello World from rank 0 running on nid03834 (hostname nid03834)! Should I PR a patch to prun manpage to change this: -H, -host, --host List of hosts on which to invoke processes. to something like this?: -H, -host, --host List of hosts on which to invoke processes. Pass +e to allocate only onto empty nodes (i.e. none of whose cores have been allocated to other prun jobs) or +e:N to allocate to nodes at least N of which are empty.
Re: [OMPI users] PRRTE DVM: how to tell prun to not share nodes among prun jobs?
That would be very kind of you and most welcome! > On Nov 14, 2020, at 12:38 PM, Alexei Colin wrote: > > On Sat, Nov 14, 2020 at 08:07:47PM +, Ralph Castain via users wrote: >> IIRC, the correct syntax is: >> >> prun -host +e ... >> >> This tells PRRTE that you want empty nodes for this application. You can >> even specify how many empty nodes you want: >> >> prun -host +e:2 ... >> >> I haven't tested that in a bit, so please let us know if it works or not so >> we can fix it if necessary. > > Works! Thank you. > > $ prun --map-by ppr:64:node --host +e -n 1 ./mpitest & > $ prun --map-by ppr:64:node --host +e -n 1 ./mpitest & > > MPI World size = 1 processes > Hello World from rank 0 running on nid03835 (hostname nid03835)! > > MPI World size = 1 processes > Hello World from rank 0 running on nid03834 (hostname nid03834)! > > Should I PR a patch to prun manpage to change this: > > -H, -host, --host > List of hosts on which to invoke processes. > > to something like this?: > > -H, -host, --host > List of hosts on which to invoke processes. Pass >+e to allocate only onto empty nodes (i.e. none of >whose cores have been allocated to other prun jobs) or >+e:N to allocate to nodes at least N of which are empty.
Re: [OMPI users] mpirun on Kubuntu 20.4.1 hangs
Hello The hwloc/X11 stuff is caused by OpenMPI using a hwloc that was built with the GL backend enabled (in your case, it's because package libhwloc-plugins is installed). That backend is used for querying the locality of X11 displays running on NVIDIA GPUs (using libxnvctrl). Does running "lstopo" fail/hang too? (it will basically run hwloc without OpenMPI). One workaround should be to set HWLOC_COMPONENTS=-gl in your environment so that this backend is ignored. Recent hwloc releases have a way to avoid some plugins at runtime through the C interface, we should likely blacklist all plugins that are already blacklisted at compile time when OMPI builds its own hwloc. Brice Le 14/11/2020 à 12:33, Jorge Silva via users a écrit : > Hello, > > In spite of the delay, I was not able to solve my problem. Thanks to > Joseph and Prentice for their interesting suggestions. > > I uninstalled AppAmor (SElinux is not installed ) as suggested by > Prentice but there were no changes, mpirun sttill hangs. > > The result of gdb stack trace is the following: > > > $ sudo gdb -batch -ex "thread apply all bt" -p $(ps -C mpirun -o pid= > | head -n 1) > [Thread debugging using libthread_db enabled] > Using host libthread_db library > "/lib/x86_64-linux-gnu/libthread_db.so.1". > 0x7f9289544307 in __libc_connect (fd=9, addr=..., len=16) at > ../sysdeps/unix/sysv/linux/connect.c:26 > 26 ../sysdeps/unix/sysv/linux/connect.c: Aucun fichier ou dossier > de ce type. > > Thread 1 (Thread 0x7f92891f4e80 (LWP 4948)): > #0 0x7f9289544307 in __libc_connect (fd=9, addr=..., len=16) at > ../sysdeps/unix/sysv/linux/connect.c:26 > #1 0x7f9288fff59d in ?? () from /lib/x86_64-linux-gnu/libxcb.so.1 > #2 0x7f9288fffc49 in xcb_connect_to_display_with_auth_info () > from /lib/x86_64-linux-gnu/libxcb.so.1 > #3 0x7f928906cb7a in _XConnectXCB () from > /lib/x86_64-linux-gnu/libX11.so.6 > #4 0x7f928905d319 in XOpenDisplay () from > /lib/x86_64-linux-gnu/libX11.so.6 > #5 0x7f92897de4fb in ?? () from > /usr/lib/x86_64-linux-gnu/hwloc/hwloc_gl.so > #6 0x7f92893b901e in ?? () from /lib/x86_64-linux-gnu/libhwloc.so.15 > #7 0x7f92893c13a0 in hwloc_topology_load () from > /lib/x86_64-linux-gnu/libhwloc.so.15 > #8 0x7f92896df564 in opal_hwloc_base_get_topology () from > /lib/x86_64-linux-gnu/libopen-pal.so.40 > #9 0x7f92891da6be in ?? () from > /usr/lib/x86_64-linux-gnu/openmpi/lib/openmpi3/mca_ess_hnp.so > #10 0x7f92897a22fc in orte_init () from > /lib/x86_64-linux-gnu/libopen-rte.so.40 > #11 0x7f92897a6c86 in orte_submit_init () from > /lib/x86_64-linux-gnu/libopen-rte.so.40 > #12 0x55819fd0b3a3 in ?? () > #13 0x7f92894480b3 in __libc_start_main (main=0x55819fd0b1c0, > argc=1, argv=0x7fff5334fe48, init=, fini= out>, rtld_fini=, stack_end=0x7fff5334fe38) at > ../csu/libc-start.c:308 > #14 0x55819fd0b1fe in ?? () > [Inferior 1 (process 4948) detached] > > > So it seems to be a problem in the connection via libxcb (socket ?) > but this is out of my system computer skills.. Is there any > authorization needed? > > As is libX11 at the origin of the call I tried to execute in a bare > terminal (ctrl-alt-f2 and via ssh) but the message is the same. I > tried to recompile/install the hole package and have the same result. > > Thank you for your help. > > Jorge > > Le 22/10/2020 à 12:16, Joseph Schuchart via users a écrit : >> Hi Jorge, >> >> Can you try to get a stack trace of mpirun using the following >> command in a separate terminal? >> >> sudo gdb -batch -ex "thread apply all bt" -p $(ps -C mpirun -o pid= | >> head -n 1) >> >> Maybe that will give some insight where mpirun is hanging. >> >> Cheers, >> Joseph >> >> On 10/21/20 9:58 PM, Jorge SILVA via users wrote: >>> Hello Jeff, >>> >>> The program is not executed, seems waits for something to connect >>> with (why twice ctrl-C ?) >>> >>> jorge@gcp26:~/MPIRUN$ mpirun -np 1 touch /tmp/foo >>> ^C^C >>> >>> jorge@gcp26:~/MPIRUN$ ls -l /tmp/foo >>> ls: impossible d'accéder à '/tmp/foo': Aucun fichier ou dossier de >>> ce type >>> >>> no file is created.. >>> >>> In fact, my question was if are there differences in mpirun usage >>> between these versions.. The >>> >>> mpirun -help >>> >>> gives a different output as expected, but I tried a lot of options >>> without any success. >>> >>> >>> Le 21/10/2020 à 21:16, Jeff Squyres (jsquyres) a écrit : There's huge differences between Open MPI v2.1.1 and v4.0.3 (i.e., years of development effort); it would be very hard to categorize them all; sorry! What happens if you mpirun -np 1 touch /tmp/foo (Yes, you can run non-MPI apps through mpirun) Is /tmp/foo created? (i.e., did the job run, and mpirun is somehow not terminating) > On Oct 21, 2020, at 12:22 PM, Jorge SILVA via users > mailto:users@lists.open-mpi.org>> wrote: > > Hello Gus, > > Thank you for your answer.. Unfortunatel