A few notes: 1. I'm guessing that your LD_LIBRARY_PATH is not set properly on the remote nodes, which is why it can't find libmpi.so on the remote nodes. Ensure that it's set properly on the other side (you'll likely need to modify your shell startup files), or use the --prefix functionality in mpirun (which will ensure to set your PATH and LD_LIBRARY_PATH properly on remote nodes), like this:
mpirun --prefix /opt/openmpi/1.1 -np 3 --hostfile /tmp/hosts ./first Or simply supply the full pathname to mpirun (exactly equivalent to --prefix): /opt/openmpi/1.1/bin/mpirun -np 3 --hostfile /tmp/hosts ./first Or if you're lazy (like me): `which mpirun` -np 3 --hostfile /tmp/hosts ./first 2. Note that your "ls" command was actually shell expanded on the node where you ran mpirun, and *then* it was executed on the remote nodes. This was not a problem because the files are actually the same on all nodes, but I thought you might want to know that for future reference. Hope that helps! On 7/28/06 4:55 PM, "Dan Lipsitt" <danlips...@gmail.com> wrote: > get the following error when I attempt to run an mpi program (called > "first", in this case) across several nodes (it works on a single > node): > > $ mpirun -np 3 --hostfile /tmp/nodes ./first > ./first: error while loading shared libraries: libmpi.so.0: cannot > open shared object file: No such file or directory > > My library path looks okay and I am able to run other programs, > including listing the supposedly missing library: > > $ echo $LD_LIBRARY_PATH > /opt/openmpi/1.1/lib/ > $ mpirun -np 3 --hostfile /tmp/nodes uptime > 16:42:51 up 22 days, 3:14, 10 users, load average: 0.01, 0.02, 0.04 > 19:49:32 up 1:36, 0 users, load average: 0.00, 0.00, 0.00 > 19:40:01 up 1:37, 0 users, load average: 0.00, 0.00, 0.00 > $ mpirun -np 3 --hostfile /tmp/nodes ls -l /opt/openmpi/1.1/lib/libmpi.so* > lrwxrwxrwx 1 root root 15 Jul 13 15:44 > /opt/openmpi/1.1/lib/libmpi.so -> libmpi.so.0.0.0 > lrwxrwxrwx 1 root root 15 Jul 13 15:44 > /opt/openmpi/1.1/lib/libmpi.so.0 -> libmpi.so.0.0.0 > -rwxr-xr-x 1 root root 6157698 Jul 12 18:08 > /opt/openmpi/1.1/lib/libmpi.so.0.0.0 > lrwxrwxrwx 1 root root 15 Jul 26 16:17 > /opt/openmpi/1.1/lib/libmpi.so -> libmpi.so.0.0.0 > lrwxrwxrwx 1 root root 15 Jul 26 16:17 > /opt/openmpi/1.1/lib/libmpi.so.0 -> libmpi.so.0.0.0 > -rwxr-xr-x 1 root root 6157698 Jul 12 18:08 > /opt/openmpi/1.1/lib/libmpi.so.0.0.0 > lrwxrwxrwx 1 root root 15 Jul 26 13:50 > /opt/openmpi/1.1/lib/libmpi.so -> libmpi.so.0.0.0 > lrwxrwxrwx 1 root root 15 Jul 26 13:50 > /opt/openmpi/1.1/lib/libmpi.so.0 -> libmpi.so.0.0.0 > -rwxr-xr-x 1 root root 6157698 Jul 12 18:08 > /opt/openmpi/1.1/lib/libmpi.so.0.0.0 > > Any suggestions? > > Thanks, > Dan > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres Server Virtualization Business Unit Cisco Systems