On 16/04/2011 12:13 AM, Larcombe, Lee wrote:
Hi gmx-users
We have an HPC setup running HP_MPI and LSF/SLURM. Gromacs 4.5.3 has been
compiled with mpi support
The compute nodes on the system contain 2 x dual core Xeons which the system
sees as 4 processors
An LSF script called gromacs_run.lsf is as shown below
#BSUB -N
#BSUB -J "gromacsTest5"
#BSUB -u l.larco...@cranfield.ac.uk
#BSUB -n 4
#BSUB -q short
#BSUB -o %J.log
mpirun -srun mdrun_mpi -v -s xxx.tpr -o xxx.trr
Queued with:
Bsub< gromacs_run.lsf
This is intended to run 1 mdrun on a single node using all four cores of the
two xeons. The result is that although the job is only submitted to one compute
node, 4 mdruns are launched on each of the 4 cores = 16 jobs. These are all the
same as if mdrun has not been compiled with mpi support.
mdrun_mpi will run one process on each core that the MPI configuration
declares is available. Spawning four separate runs of four processes
indicates that MPI is not configured to reflect the hardware (since each
run thinks it can have four processes), or that the submission script is
inappropriate (since four runs get spawned), or both. We can't help
there. Trouble-shoot with a trivial MPI test program.
If I tell srun to start just one task with "mpirun -srun -n1 mdrun_mpi -v -s xxx.tpr
–o xxx.trr" it starts one job on each core instead of 4:
NNODES=1, MYRANK=0, HOSTNAME=comp195
NNODES=1, MYRANK=0, HOSTNAME=comp195
NNODES=1, MYRANK=0, HOSTNAME=comp195
NNODES=1, MYRANK=0, HOSTNAME=comp195
Logs show 4 mdrun_mpi starts, 4 file read ins and I get 4 of all run files in
CWD. I am sure
that mdrun_mpi is indeed compiled with mpi support - although our sysadmin did that, not me. For
example, if I try and execute "mdrun_mpi –h" I get a message from HP–MPI and have to
execute "mpirun mdrun_mpi –h" to see the help text.
Does anyone have any experience of running with this setup - any ideas?
Thanks
Lee
--
gmx-users mailing list gmx-users@gromacs.org
http://lists.gromacs.org/mailman/listinfo/gmx-users
Please search the archive at
http://www.gromacs.org/Support/Mailing_Lists/Search before posting!
Please don't post (un)subscribe requests to the list. Use the
www interface or send it to gmx-users-requ...@gromacs.org.
Can't post? Read http://www.gromacs.org/Support/Mailing_Lists