Am 02.03.2014 um 23:39 schrieb Victor:

> Thanks for your reply. There are some updates, but it was too late last night 
> to post it.
> 
> I now have the AMD/Intel heterogeneous cluster up and running. The initial 
> problem was that when I installed OpenMPI on the AMD nodes, the library paths 
> were set to a different location than on the Intel nodes. I am not sure why.

/lib vs. /lib64? It depends on the version of `libtool` on the machine IIRC.


> In any case I then followed the suggestion from the FAQ and instead shared 
> the same OpenMPI install directory with all the nodes via NFS. Now the job is 
> running so I can confirm that it is indeed possible to run the same job on a 
> heterogeneous cluster comprised of AMD and Intel nodes.
> 
> I am using OpenMPI 1.7.4 now.
> 
> There is a related problem though. I am sharing /opt/openmpi-1.7.4 via NFS 
> but there does not seem to be a way to tell the nodes where OpenMPI is 
> located when using non-interactive SSH (using secure key login). SSH does not 
> seem to parse .bash_profile so I do not know how to tell the jobs on the 
> nodes where to find OpenMPI except by starting the job with 
> /opt/openmpi-1.7.4/bin/mpirun.

For a non-interactive login ~/.bashrc is used.

-- Reuti


> Regarding open-mx, yes I will look into that next to see if the job is indeed 
> using it. My msa flag is --mca mx self
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to