My guess is that you aren't doing the allocation correctly - since you are 
using qsub, can I assume you have Moab as your scheduler?

aprun should be forwarding the envars - do you see them if you just run "aprun 
-n 1 printenv"?

On Nov 23, 2013, at 2:13 PM, Teranishi, Keita <knte...@sandia.gov> wrote:

> Hi,
> 
> I installed OpenMPI on our small XE6 using the configure options under 
> /contrib directory.  It appears it is working fine, but it ignores MCA 
> parameters (set in env var).  So I switched to mpirun (in OpenMPI) and it can 
> handle MCA parameters somehow.  However,  mpirun fails to allocate process by 
> cores.  For example, I allocated 32 cores (on 2 nodes) by "qsub –lmppwidth=32 
> –lmppnppn=16", mpirun recognizes it as 2 slots.    Is it possible to mpirun 
> to handle mluticore nodes of XE6 properly or is there any options to handle 
> MCA parameters for aprun?
> 
> Regards,
> -----------------------------------------------------------------------------
> Keita Teranishi
> Principal Member of Technical Staff
> Scalable Modeling and Analysis Systems
> Sandia National Laboratories
> Livermore, CA 94551
> +1 (925) 294-3738
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to