FWIW, to avoid using the --prefix option, you can set your PATH / LD_LIBRARY_PATH to point to the Open MPI installation on all nodes.
Many organizations opt to have NFS-shared home directories, so that when you modify your "main" shell startup file (e.g., .bashrc) to point PATH and LD_LIBRARY_PATH to your Open MPI installation, it effectively modifies it for all nodes in the cluster. On Apr 29, 2013, at 8:56 AM, E.O. <ooyama.eii...@gmail.com> wrote: > It works!!! > By putting two dash'es and no equal sign, it worked fine!! > > [root@host1 tmp]# mpirun --prefix /myname --host host2 /tmp/hello.out > Hello World from processor host2, rank 0 out of 1 processors > [root@host1 tmp]# > > It looks like one dash "-prefix" also works if I don't put an equal sign.. > > Thank you very much!! > > Eiichi > > > > On Mon, Apr 29, 2013 at 8:29 AM, Ralph Castain <r...@open-mpi.org> wrote: > Hmm....okay. No, let's not bother to install a bunch of stuff you don't > otherwise need. > > I probably mis-typed the "prefix" option - it has two dashes in front of it > and no equal sign: > > mpirun --prefix ./myname ... > > I suspect you only put one dash, and the equal sign was a definite problem, > which is why it gave you an error. > > > On Apr 29, 2013, at 5:12 AM, E.O. <ooyama.eii...@gmail.com> wrote: > >> I tried configuring/building an OMPI on the remote host but I was not able >> to... >> The remote host (host2) doesn't have any development tools, such as gcc, >> make, etc... >> >> Since I am able to run an MPI hello_c binary on the remote host, I believe >> the host has all the necessary libraries needed for MPI. I am also able to >> run an MPI hello_c binary on host1 from host2. >> >> [root@host2 tmp]# mpirun -host localhost /tmp/hello.out >> Hello World from processor host2, rank 0 out of 1 processors >> [root@host2 tmp]# mpirun -host host2 /tmp/hello.out >> Hello World from processor host2, rank 0 out of 1 processors >> [root@host2 tmp]# mpirun -host host1 /tmp/hello.out >> Hello World from processor host1, rank 0 out of 1 processors >> [root@host2 tmp]# >> >> However I still can't run hello_c binary on host2 from host1 >> >> [root@host1 tmp]# mpirun -host host2 /tmp/hello.out >> -------------------------------------------------------------------------- >> Sorry! You were supposed to get help about: >> opal_init:startup:internal-failure >> But I couldn't open the help file: >> //share/openmpi/help-opal-runtime.txt: No such file or directory. Sorry! >> -------------------------------------------------------------------------- >> [host2:02499] [[INVALID],INVALID] ORTE_ERROR_LOG: Error in file >> runtime/orte_init.c at line 79 >> [host2:02499] [[INVALID],INVALID] ORTE_ERROR_LOG: Error in file >> orted/orted_main.c at line 358 >> -------------------------------------------------------------------------- >> A daemon (pid 17710) died unexpectedly with status 255 while attempting >> to launch so we are aborting. >> >> There may be more information reported by the environment (see above). >> >> This may be because the daemon was unable to find all the needed shared >> libraries on the remote node. You may set your LD_LIBRARY_PATH to have the >> location of the shared libraries on the remote nodes and this will >> automatically be forwarded to the remote nodes. >> -------------------------------------------------------------------------- >> -------------------------------------------------------------------------- >> mpirun noticed that the job aborted, but has no info as to the process >> that caused that situation. >> -------------------------------------------------------------------------- >> [root@host1 tmp]# >> >> >> If I set -prefix=/myname, it returns a different output >> >> [root@host1 tmp]# mpirun -prefix=/myname -host host2 /tmp/hello.out >> -------------------------------------------------------------------------- >> mpirun was unable to launch the specified application as it could not access >> or execute an executable: >> >> Executable: -prefix=/myname >> Node: host1 >> >> while attempting to start process rank 0. >> -------------------------------------------------------------------------- >> [root@host1 tmp]# >> >> Do you still want me to try building OMPI on the remote host? >> >> eiichi >> >> >> >> On Sun, Apr 28, 2013 at 12:24 PM, Ralph Castain <r...@open-mpi.org> wrote: >> If you configure/build OMPI on the remote node using the same configure >> options you used on host1, does the problem go away? >> >> >> On Apr 28, 2013, at 8:58 AM, E.O. <ooyama.eii...@gmail.com> wrote: >> >>> Thank you Ralph! >>> I ran it with "-prefix" option but I got this... >>> >>> [root@host1 tmp]# mpirun -prefix /myname -np 4 -host host2 ./hello.out >>> -------------------------------------------------------------------------- >>> mpirun was unable to launch the specified application as it could not access >>> or execute an executable: >>> >>> Executable: -prefix=/myname >>> Node: host1 >>> >>> while attempting to start process rank 0. >>> -------------------------------------------------------------------------- >>> [root@host1 tmp]# >>> >>> I also updated PATH in the remote host (host2) to include /myname. >>> But it didn't seem change anything... >>> >>> eiichi >>> >>> >>> >>> >>> On Sun, Apr 28, 2013 at 11:48 AM, Ralph Castain <r...@open-mpi.org> wrote: >>> The problem is likely that your path variables aren't being set properly on >>> the remote machine when mpirun launches the remote daemon. You might check >>> to see that your default shell rc file is also setting those values >>> correctly. Alternatively, modify your mpirun cmd line a bit by adding >>> >>> mpirun -prefix /myname ... >>> >>> so it will set the remove prefix and see if that helps. If it does, you can >>> add --enable-orterun-prefix-by-default to your configure line so mpirun >>> always adds it. >>> >>> >>> On Apr 28, 2013, at 7:56 AM, "E.O." <ooyama.eii...@gmail.com> wrote: >>> >>> > Hello >>> > >>> > I have five linux machines (one is redhat and the other are busybox) >>> > I downloaded openmpi-1.6.4.tar.gz into my main redhat machine and >>> > configure'ed/compiled it successfully. >>> > ./configure --prefix=/myname >>> > I installed it to /myname directory successfully. I am able to run a >>> > simple hallo.c on my redhat machine. >>> > >>> > [root@host1 /tmp] # mpirun -np 4 ./hello.out >>> > I am parent >>> > I am a child >>> > I am a child >>> > I am a child >>> > [root@host1 /tmp] # >>> > >>> > Then, I sent entire /myname directory to the another machine (host2). >>> > [root@host1 /] # tar zcf - myname | ssh host2 "(cd /; tar zxf -)" >>> > >>> > and ran mpirun for the host (host2). >>> > >>> > [root@host1 tmp]# mpirun -np 4 -host host2 ./hello.out >>> > -------------------------------------------------------------------------- >>> > Sorry! You were supposed to get help about: >>> > opal_init:startup:internal-failure >>> > But I couldn't open the help file: >>> > //share/openmpi/help-opal-runtime.txt: No such file or directory. >>> > Sorry! >>> > -------------------------------------------------------------------------- >>> > [host2:26294] [[INVALID],INVALID] ORTE_ERROR_LOG: Error in file >>> > runtime/orte_init.c at line 79 >>> > [host2:26294] [[INVALID],INVALID] ORTE_ERROR_LOG: Error in file >>> > orted/orted_main.c at line 358 >>> > -------------------------------------------------------------------------- >>> > A daemon (pid 23691) died unexpectedly with status 255 while attempting >>> > to launch so we are aborting. >>> > >>> > There may be more information reported by the environment (see above). >>> > >>> > This may be because the daemon was unable to find all the needed shared >>> > libraries on the remote node. You may set your LD_LIBRARY_PATH to have the >>> > location of the shared libraries on the remote nodes and this will >>> > automatically be forwarded to the remote nodes. >>> > -------------------------------------------------------------------------- >>> > -------------------------------------------------------------------------- >>> > mpirun noticed that the job aborted, but has no info as to the process >>> > that caused that situation. >>> > -------------------------------------------------------------------------- >>> > [root@host1 tmp]# >>> > >>> > I set those environment variables >>> > >>> > [root@host1 tmp]# echo $LD_LIBRARY_PATH >>> > /myname/lib/ >>> > [root@host1 tmp]# echo $OPAL_PREFIX >>> > /myname/ >>> > [root@host1 tmp]# >>> > >>> > [root@host2 /] # ls -la /myname/lib/libmpi.so.1 >>> > lrwxrwxrwx 1 root root 15 Apr 28 10:21 >>> > /myname/lib/libmpi.so.1 -> libmpi.so.1.0.7 >>> > [root@host2 /] # >>> > >>> > If I ran the ./hello.out binary inside host2, it works fine >>> > >>> > [root@host1 tmp]# ssh host2 >>> > [root@host2 /] # /tmp/hello.out >>> > I am parent >>> > [root@host2 /] # >>> > >>> > Can someone help me figure out why I cannot run hello.out in host2 from >>> > host1 ? >>> > Am I missing any env variables ? >>> > >>> > Thank you, >>> > >>> > Eiichi >>> > >>> > >>> > _______________________________________________ >>> > users mailing list >>> > us...@open-mpi.org >>> > http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/