Hi Felix,
On 11/23/23 18:14, Felix wrote:
Will slurm-20.02 which is installed on a management node comunicate with
slurm-22.05 installed on a work nodes?
They have the same configuration file slurm.conf
Or do the version have to be the same. Slurm 20.02 was installed manually
and slurm 22.05
Hello
I have a curiosity and question in the same time,
Will slurm-20.02 which is installed on a management node comunicate with
slurm-22.05 installed on a work nodes?
They have the same configuration file slurm.conf
Or do the version have to be the same. Slurm 20.02 was installed
manually
ego...@posteo.me writes:
> ok, I understand synching of users to slurm database is a task which
> it not built-in, but could be added outside of slurm :-)
>
> With regards to the QoS or Partition QoS setting I've tried several
> settings and configurations however it was not possible at all to
> c
Thanks for confirming, Brian. That was my understanding as well. Do you
have it working that way on a machine you have access to? If so, I'd be
interested to see the config file, because that's not the behavior I am
experiencing in my tests.
In fact, in my tests Slurm will not bring down those "X
"Schneider, Gerald" writes:
> Is there any way to release the allocation manually?
I've only seen this once on our clusters, and that time it helped just
restarting slurmctld.
If this is a recurring problem, perhaps it will help to upgrade Slurm.
You are running quite an old version.
--
Regar
On 11/23/23 11:50, Markus Kötter wrote:
On 23.11.23 10:56, Schneider, Gerald wrote:
I have a recurring problem with allocated TRES, which are not
released after all jobs on that node are finished. The TRES are still
marked as allocated and no new jobs can't be scheduled on that node
using those
Hi,
On 23.11.23 10:56, Schneider, Gerald wrote:
I have a recurring problem with allocated TRES, which are not
released after all jobs on that node are finished. The TRES are still
marked as allocated and no new jobs can't be scheduled on that node
using those TRES.
Remove the node from slurm.c
Hi there,
I have a recurring problem with allocated TRES, which are not released after
all jobs on that node are finished. The TRES are still marked as allocated and
no new jobs can't be scheduled on that node using those TRES.
$ scontrol show node node2
NodeName=node2 Arch=x86_64 CoresPerSocke