Hi,

I have some performance issue on a parallel machine composed of nodes of 16 
procs each. The application is launched on multiple of 16 procs for given 
numbers of nodes.
I was told by people using MX MPI with this machine to attach a script to 
mpiexec, which 'numactl' things, in order to make the execution performance 
stable.

Looking on the faq (the oldest one is for OpenMPI v1.3?), I saw that maybe the 
solution would be for me to use the --mca mpi_paffinity_alone 1
Is that correct? -- BTW, I have both memory and processor affinity:
>ompi_info | grep affinity
           MCA paffinity: linux (MCA v2.0, API v2.0, Component v1.4.2)
           MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.4.2)
           MCA maffinity: libnuma (MCA v2.0, API v2.0, Component v1.4.2)
Does it handle memory too, or do I have to use another option like 
--mca mpi_maffinity 1?

Still, I would like to test the numactl solution. Does OpenMPI provide an 
equivalent to $MXMPI_ID which gives at least gives the NODE on which a 
process is launched by OpenMPI, so that I can adapt the script that was given 
to me?

Tkx.

.Yves.

Reply via email to