Adrian and Diego,
Are you using AMD Epyc processors when viewing this issue? I've been
having the same issue but only on dual AMD Epyc systems. I haven't tried
changing the core file location from an NFS mount though so perhaps
there's an issue writing it out in time.
How did you disable core files?
Regards,
Willy Markuske
HPC Systems Engineer
Research Data Services
P: (619) 519-4435
On 8/6/21 6:16 AM, Adrian Sevcenco wrote:
On 8/6/21 3:19 PM, Diego Zuccato wrote:
IIRC we increased SlurmdTimeout to 7200 .
Thanks a lot!
Adrian
Il 06/08/2021 13:33, Adrian Sevcenco ha scritto:
On 8/6/21 1:56 PM, Diego Zuccato wrote:
We had a similar problem some time ago (slow creation of big core
files) and solved it by increasing the Slurm timeouts
oh, i see.. well, in principle i should not have core files, and i
do not find any...
to the point that even the slowest core wouldn't trigger it. Then,
once the need for core files was over, I disabled core files and
restored the timeouts.
and how much did you increased them? i have
SlurmctldTimeout=300
SlurmdTimeout=300
Thank you!
Adrian
Il 06/08/2021 12:46, Adrian Sevcenco ha scritto:
On 8/6/21 1:27 PM, Diego Zuccato wrote:
Hi.
Hi!
Might it be due to a timeout (maybe the killed job is creating a
core file, or caused heavy swap usage)?
i will have to search for culprit ..
the problem is why would the node be put in drain for the reason
of failed killing? and how can i control/disable
this?
Thank you!
Adrian
BYtE,
Diego
Il 06/08/2021 09:02, Adrian Sevcenco ha scritto:
Having just implemented some triggers i just noticed this:
NODELIST NODES PARTITION STATE CPUS S:C:T MEMORY
TMP_DISK WEIGHT AVAIL_FE REASON
alien-0-47 1 alien* draining 48 48:1:1 193324
214030 1 rack-0,4 Kill task failed
alien-0-56 1 alien* drained 48 48:1:1 193324
214030 1 rack-0,4 Kill task failed
i was wondering why a node is drained when killing of task fails
and how can i disable it? (i use cgroups)
moreover, how can the killing of task fails? (this is on slurm
19.05)
Thank you!
Adrian