quot;--with-tm") is now responding to the Torque mode allocation correctly and
is no longer simply running all the jobs on the first node.
That is$PBS_NODEFILE ,pbsdsh hostname and mpirun hostname
are all in agreement.
Thank you all for your help, and putt
.
Take you for your help, though I am still open to suggestions for a
replacement.
Anthony Thyssen ( System Programmer )
--
Encryption... is a powerful defensive weapon for free people.
It offers a technical guarantee o
eptember/018858.html
>
> and no conclusion was reached.
>
>
> Cheers,
>
>
> Gilles
>
>
> On 10/3/2017 2:02 PM, Anthony Thyssen wrote:
>
>> The stdin and stdout are saved to separate channels.
>>
>> It is interesting that the output from pbsds
The stdin and stdout are saved to separate channels.
It is interesting that the output from pbsdsh is node21.emperor 5 times,
even though $PBS_NODES is the 5 individual nodes.
Attached are the two compressed files, as well as the pbs_hello batch used.
Anthony Thyssen ( System Programmer
=UP
node22.emperor: slots=1 max_slots=0 slots_inuse=0 state=UP
=
node21.emperor
node21.emperor
node21.emperor
node21.emperor
node21.emperor
===8<----CUT HERE--
ss as to how to solve this problem..
ANY and all suggestions, or even ways I can get other information as to
what is causing this will be most welcome.
Anthony Thyssen ( System Programmer )
--
Using encryption on th
"pbsnodes")
*ALL 5 processes was run on the first node. Vastly over-subscribing that
node.*
Anyone have any ideas as to what went wrong?
*Why did OpenMPI not follow the node mapping it says it should be
following!*
Additional... OpenMPI on its own (without torque) does appear to work
gnoring it, and just running everything (over-subscribing) on
the first node given. The previous problem did not over subscribe the
nodes. It just did not spread out the processes as requested.
I am starting a new thread about this problem to try and get some help.
Anthony Thyssen
d, then we use what they give us
>
> Sent from my iPad
>
> On Sep 26, 2017, at 8:11 PM, Anthony Thyssen
> wrote:
>
>
> I have been having problems with OpenMPI on a new cluster of machines,
> using
> stock RHEL7 packages.
>
> ASIDE: This will be used with
I have been having problems with OpenMPI on a new cluster of machines, using
stock RHEL7 packages.
ASIDE: This will be used with Torque-PBS (from EPEL archives), though
OpenMPI
(currently) does not have the "tm" resource manager configured to use PBS,
as you
will be able to see in the debug output
10 matches
Mail list logo