I suspect something else is going on there - I can't imagine how the LAMA
mapper could be interacting with the Torque launcher. The check for adequate
resources (per the error message) is done long before we get to the launcher.
I'll have to let the LAMA supporters chase it down.
Thanks
Ralph
Thanks, Ralph.
This is an additional information.
Just execute directly on the node without Torque:
mpirun -np 8 -report-bindings -mca rmaps lama -mca rmaps_lama_bind 1c
Myprog
Then it also works, which means the combination of LAMA and Torque would
case
the problem.
Tetsuya Mishima
> Okay,
Okay, so the problem is a bug in LAMA itself. I'll file a ticket and let the
LAMA folks look into it.
On Nov 7, 2013, at 4:18 PM, tmish...@jcity.maeda.co.jp wrote:
>
>
> Hi Ralph,
>
> I quickly tried 2 runs:
>
> mpirun -report-bindings -bind-to core Myprog
> mpirun -machinefile pbs_hosts -np
Hi Ralph,
I quickly tried 2 runs:
mpirun -report-bindings -bind-to core Myprog
mpirun -machinefile pbs_hosts -np ${NPROCS} -report-bindings -bind-to core
Myprog
It works fine in both cases on node03 and node08.
Regards,
Tetsuya Mishima
> What happens if you drop the LAMA request and instead
What happens if you drop the LAMA request and instead run
mpirun -report-bindings -bind-to core Myprog
This would do the same thing - does it work? If so, then we know it is a
problem in the LAMA mapper. If not, then it is likely a problem in a different
section of the code.
On Nov 7, 2013,
Dear openmpi developers,
I tried the new function LAMA of openmpi-1.7.3 and
unfortunately it is not stable under my environment, which
is built with torque.
(1) I used 4 scripts as shown below to clarify the problem:
(COMMON PART)
#!/bin/sh
#PBS -l nodes=node03:ppn=8 / nodes=node08:ppn=8
expor
Jeff,
Good to know. Thanks!
Seems really like MPI_ABORT should only be used within error traps after MPI
functions have been started.
Code-wise, the sample I got was not the best. Usage should be checked before
MPI_Initialize, I think :)
It seems the expectation is that MPI_ABORT is only call
Hi Steven, Dmytry
Not sure if this web page is still valid or totally out of date,
but there it goes anyway, in the hopes that it may help:
http://www.mcs.anl.gov/research/projects/mpi/mpich1-old/docs/install/node38.htm
On the other hand, one expert seems to dismiss NFS
for paralllel IO:
http:
That's a relatively old version of OMPI. Maybe try the latest
release? That's always the safe bet since the issue might have been
fixed already.
I recall that OMPI uses ROMIO so you might try to reproduce with MPICH
so you can report it to the people that wrote the MPI-IO code. Of
course, this mi
Not sure if this is related, but:
I've seen a case of performance degradation on NFS and Lustre when
writing NetCDF files. The reason was that the file was filled with a
loop writing one 4-byte record at once. Performance became close to
local hard drive, when I simply introduced buffering of reco
The simple C program attached below hangs on MPI_File_write when I am using an
NFS-mounted filesystem. Is MPI-IO supported in OpenMPI for NFS filesystems?
I'm using OpenMPI 1.4.5 on Debian stable (wheezy), 64-bit Opteron CPU, Linux
3.2.51. I was surprised by this because the problems only st
11 matches
Mail list logo