Re: [OMPI users] problem with rankfile in openmpi-1.6.4rc2

2013-01-24 Thread Ralph Castain
I built the current 1.6 branch (which hasn't seen any changes that would impact this function) and was able to execute it just fine on a single socket machine. I then gave it your slot-list, which of course failed as I don't have two active sockets (one is empty), but it appeared to parse the li

Re: [OMPI users] Error when attempting to run LAMMPS on Centos 6.2 with OpenMPI

2013-01-24 Thread Ralph Castain
How was OMPI configured? What type of system are you running on (i.e., what is the launcher - ssh, lsf, slurm, ...)? On Jan 24, 2013, at 6:35 PM, #YEO JINGJIE# wrote: > Dear users, > > Maybe something went wrong as I was compiling OpenMPI, I am very new to > linux. When I try to run LAMMPS

[OMPI users] Error when attempting to run LAMMPS on Centos 6.2 with OpenMPI

2013-01-24 Thread #YEO JINGJIE#
Dear users, Maybe something went wrong as I was compiling OpenMPI, I am very new to linux. When I try to run LAMMPS using the following command: /usr/lib64/openmpi/bin/mpirun -n 16 /opt/lammps-21Jan13/lmp_linux < zigzag.in I get the following errors: [NTU-2:28895] [[INVALID],INVALID] O

Re: [OMPI users] mpivars.sh - Intel Fortran 13.1 conflict with OpenMPI 1.6.3

2013-01-24 Thread Tim Prince
On 01/24/2013 12:40 PM, Michael Kluskens wrote: This is for reference and suggestions as this took me several hours to track down and the previous discussion on "mpivars.sh" failed to cover this point (nothing in the FAQ): I successfully build and installed OpenMPI 1.6.3 using the following on

[OMPI users] mpivars.sh - Intel Fortran 13.1 conflict with OpenMPI 1.6.3

2013-01-24 Thread Michael Kluskens
This is for reference and suggestions as this took me several hours to track down and the previous discussion on "mpivars.sh" failed to cover this point (nothing in the FAQ): I successfully build and installed OpenMPI 1.6.3 using the following on Debian Linux: ./configure --prefix=/opt/openmpi

Re: [OMPI users] openmpi 1.6.3, job submitted through torque/PBS + Moab (scheduler) only land on one node even though multiple nodes/processors are specified

2013-01-24 Thread Ralph Castain
Sure - just add --with-openib=no --with-psm=no to your config line and we'll ignore it On Jan 24, 2013, at 7:09 AM, Sabuj Pattanayek wrote: > ahha, with --display-allocation I'm getting : > > mca: base: component_find: unable to open > /sb/apps/openmpi/1.6.3/x86_64/lib/openmpi/mca_mtl_psm: >

Re: [OMPI users] MPI_Alltoallv performance regression 1.6.0 to 1.6.1

2013-01-24 Thread Number Cruncher
I've looked in more detail at the current two MPI_Alltoallv algorithms and wanted to raise a couple of ideas. Firstly, the new default "pairwise" algorithm. * There is no optimisation for sparse/empty messages, compare to the old basic "linear" algorithm. * The attached "pairwise-nop" patch add

Re: [OMPI users] openmpi 1.6.3, job submitted through torque/PBS + Moab (scheduler) only land on one node even though multiple nodes/processors are specified

2013-01-24 Thread Brock Palen
On Jan 24, 2013, at 10:10 AM, Sabuj Pattanayek wrote: > or do i just need to compile two versions, one with IB and one without? You should not need to, we have OMPI compiled for openib/psm and run that same install on psm/tcp and verbs(openib) based gear. All the nodes assigned to your job have

Re: [OMPI users] openmpi 1.6.3, job submitted through torque/PBS + Moab (scheduler) only land on one node even though multiple nodes/processors are specified

2013-01-24 Thread Sabuj Pattanayek
or do i just need to compile two versions, one with IB and one without? On Thu, Jan 24, 2013 at 9:09 AM, Sabuj Pattanayek wrote: > ahha, with --display-allocation I'm getting : > > mca: base: component_find: unable to open > /sb/apps/openmpi/1.6.3/x86_64/lib/openmpi/mca_mtl_psm: > libpsm_infinipa

Re: [OMPI users] openmpi 1.6.3, job submitted through torque/PBS + Moab (scheduler) only land on one node even though multiple nodes/processors are specified

2013-01-24 Thread Sabuj Pattanayek
ahha, with --display-allocation I'm getting : mca: base: component_find: unable to open /sb/apps/openmpi/1.6.3/x86_64/lib/openmpi/mca_mtl_psm: libpsm_infinipath.so.1: cannot open shared object file: No such file or directory (ignored) I think the system I compiled it on has different ib libs than

Re: [OMPI users] openmpi 1.6.3, job submitted through torque/PBS + Moab (scheduler) only land on one node even though multiple nodes/processors are specified

2013-01-24 Thread Ralph Castain
How did you configure OMPI? If you add --display-allocation to your cmd line, does it show all the nodes? On Jan 24, 2013, at 6:34 AM, Sabuj Pattanayek wrote: > Hi, > > I'm submitting a job through torque/PBS, the head node also runs the > Moab scheduler, the .pbs file has this in the resource

[OMPI users] openmpi 1.6.3, job submitted through torque/PBS + Moab (scheduler) only land on one node even though multiple nodes/processors are specified

2013-01-24 Thread Sabuj Pattanayek
Hi, I'm submitting a job through torque/PBS, the head node also runs the Moab scheduler, the .pbs file has this in the resources line : #PBS -l nodes=2:ppn=4 I've also tried something like : #PBS -l procs=56 and at the end of script I'm running : mpirun -np 8 cat /dev/urandom > /dev/null or