Hi there,
Slurm version 18.08
I am trying to find out if there is a way to add a specific gres, in this
case a GPU to a reservation? I think I can reserve a portion of a node
that has a specific gres quantity attached but I cannot figure out how to
reserve the gres, so I cannot guarantee that i
The uid and gid are the same for the slurm and munge users on each node. The
two new nodes, one of which can’t connect with the controller, have the same
users and were created with the same sequence of steps. The only exception is
that the node that won’t connect has the software stack to com
There is a third user account on all machines in the cluster that is the
user account for using the cluster. That account has uid 1000 on all four
worker nodes, but on the controller it is 1001. So that is probably why the
question marks.
I doubt this is the issue when 3 of the 4 nodes that work
I added two new nodes to my cluster (5 nodes total including controller).
One of the new nodes works, but the other one can't connect to the
controller. Both new nodes were created the same way except that the one
that can't connect to the controller has some extra packages installed to
build slur
On 4/22/20 12:56 PM, dean.w.schu...@gmail.com wrote:
There is a third user account on all machines in the cluster that is the
user account for using the cluster. That account has uid 1000 on all four
worker nodes, but on the controller it is 1001. So that is probably why the
question marks.
Hi all:
[ BTW this is the same situation that the submitter of
https://bugs.schedmd.com/show_bug.cgi?id=2692 presented. ]
We have a non-Slurm cluster in production and are developing our next one,
which will run Slurm 20.02.X.
We have a partition "batch" which is open to all users. Half of th
Even for users other than slurm and munge? It seems strange that 3 of 4 worker
nodes work with the same UIDs/GIDs as the non-working nodes.
-Original Message-
From: slurm-users On Behalf Of
Christopher Samuel
Sent: Wednesday, April 22, 2020 2:27 PM
To: slurm-users@lists.schedmd.com
Sub
Hi Michael,
Thanks for your insightful explanation of the Slurm RPM build process!
This clarified the topic a lot for me. I have updated my Slurm
installation Wiki page based upon your information:
https://wiki.fysik.dtu.dk/niflheim/Slurm_installation#build-slurm-rpms
/Ole
On 21-04-2020 22
Hi Dean,
On Wed, Apr 22, 2020 at 07:28:15PM -0600, dean.w.schu...@gmail.com wrote:
> Even for users other than slurm and munge? It seems strange that 3 of
> 4 worker nodes work with the same UIDs/GIDs as the non-working nodes.
As in:
https://slurm.schedmd.com/quickstart_admin.html
Super Quick