Have a look at the bash man page, and these two FAQ items: http://www.open-mpi.org/faq/?category=running#adding-ompi-to-path http://www.open-mpi.org/faq/?category=running#mpirun-prefix
On Feb 25, 2011, at 10:31 AM, Xianglong Kong wrote: > .bashrc is not executed when I ssh the node. How can I let it be executed? > > Kong > > On Fri, Feb 25, 2011 at 10:04 AM, Jeff Squyres <jsquy...@cisco.com> wrote: >> Ensure to check that a) your .bashrc is actually executing when you "ssh >> othernode env", and b) if .bashrc is executing, make sure that it isn't >> prematurely exiting for non-interactive jobs. >> >> >> On Feb 25, 2011, at 9:58 AM, Xianglong Kong wrote: >> >>> I'm using openmpi 1.4.3. The cluster consist of two desktop with Intel >>> core 2 duo running on Ubuntu 10.04. >>> >>> A weird thing that i found is that when I issued the command "env | >>> grep LD_LIBRARY_PATH" on the slave node, it showed the mpi lib path. >>> But when >>> I issude the command "ssh slave-node env | grep LD_LIBRARY_PATH" on >>> the master side to check the LD_LIBRARY_PATH of the slave node, it >>> showed nothing. Also, issuing the command "ssh master-node env | grep >>> LD_LIBRARY_PATH" on the slave side would return the correct mpi lib >>> path. >>> >>> I tried to modify the .bashrc and the /etc/ld.so.conf.d/*.conf file to >>> configure the LD_LIBRARY_PATH on the slave node, but it seems to work >>> only locally. How can I set the LD_LIBRARY_PATH on the slave node >>> side, so that I can get the correct path when I use "ssh slave-node >>> env | grep LD_LIBRARY_PATH" on the master side? >>> >>> Kong >>> >>> On Wed, Feb 23, 2011 at 5:22 PM, Bill Rankin <bill.ran...@sas.com> wrote: >>>> Jeff: >>>> >>>>> FWIW: I have rarely seen this to be the issue. >>>> >>>> Been bitten by similar situations before. But it may not have been >>>> OpenMPI. In any case it was a while back. >>>> >>>>> In short, programs are erroneous that do not guarantee that all their >>>>> outstanding requests have completed before calling finalize. >>>> >>>> Agreed 100%. The barrier won't prevent the case of unmatched >>>> sends/receives or outstanding request handles, but if the logic is correct >>>> it does make sure that everyone completes before anyone leaves. >>>> >>>> In any case, I also tried code #2 and it completed w/o issue on our >>>> cluster. I guess the next thing to ask Kong is regarding what version he >>>> is running and what is the platform. >>>> >>>> -b >>>> >>>> >>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>> >>> >>> >>> -- >>> Xianglong Kong >>> Department of Mechanical Engineering >>> University of Rochester >>> Phone: (585)520-4412 >>> MSN: dinosaur8...@hotmail.com >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> -- >> Jeff Squyres >> jsquy...@cisco.com >> For corporate legal information go to: >> http://www.cisco.com/web/about/doing_business/legal/cri/ >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > > > > -- > Xianglong Kong > Department of Mechanical Engineering > University of Rochester > Phone: (585)520-4412 > MSN: dinosaur8...@hotmail.com > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/