Hi everybody,

for forcing a run of your config management as Tina suggested you might just add a

ExecStartPre=

line to your slurmd.service file?

This is somewhat unrelated to your problem but we are very successfully using

ExecStartPre=-/usr/bin/nvidia-smi -L

in our slurmd.service file to make sure that all the GPU-devices are visible and available on our GPU-nodes *before* slurmd starts. Of course the dash after the "=" is important to make systemd ignore potential errors when running that command.

Hermann

On 2/24/22 10:42 AM, Tina Friedrich wrote:
Hi David,

it's also not actually a problem if the slurm.conf is not exactly the same immediately on boot - really. Unless there's changes that are very fundamental, nothing bad will happen if they pick up a new copy after, say, 5 or 10 minutes.

But it should be possible to - for example - force a run of your config management on startup (or before SLURM startup)?

(Not many ideas about the Nagios check, unless you change it to something that interrogates SLURM about node states, or keep some other record somewhere that it can interrogate about nodes meant to be down.)

Tina

On 24/02/2022 09:20, David Simpson wrote:
Hi Brian,

 >>For monitoring, I use a combination of netdata+prometheus. Data is gathered whenever the nodes are up and stored for history. Yes, when the nodes are powered down, there are empty gaps, but that is interpreted as the node is powered down.

Ah time-series will cope much better - at the moment our monitoring system (for compute node health at least) is nagios-like, hence the problem. Though there’s potential the entire cluster’s stack may change at some point, so this problem will be more easy to deal with (with a change of monitoring system for node health).

 >>For the config, I have no access to DNS for configless so I use a symlink to the slurm.conf file a shared filesystem. This works great. Anytime there are changes, a simple 'scontrol reconfigure' brings all running nodes up to speed and any down nodes will automatically read the latest.

Yes, currently we use file based and config written to the compute node’s disks themselves via ansible. Perhaps we will consider moving the file to a shared fs.

regards
David

-------------

David Simpson - Senior Systems Engineer

ARCCA, Redwood Building,

King Edward VII Avenue,

Cardiff, CF10 3NB

David Simpson - peiriannydd uwch systemau

ARCCA, Adeilad Redwood,

King Edward VII Avenue,

Caerdydd, CF10 3NB

simpso...@cardiff.ac.uk <mailto:simpso...@cardiff.ac.uk>

+44 29208 74657

*From:*slurm-users <slurm-users-boun...@lists.schedmd.com> *On Behalf Of *Brian Andrus
*Sent:* 23 February 2022 15:27
*To:* slurm-users@lists.schedmd.com
*Subject:* Re: [slurm-users] monitoring and update regime for Power Saving nodes

*External email to Cardiff University - *Take care when replying/opening attachments or links.

*Nid ebost mewnol o Brifysgol Caerdydd yw hwn - *Cymerwch ofal wrth ateb/agor atodiadau neu ddolenni.

David,

For monitoring, I use a combination of netdata+prometheus. Data is gathered whenever the nodes are up and stored for history. Yes, when the nodes are powered down, there are empty gaps, but that is interpreted as the node is powered down.

For the config, I have no access to DNS for configless so I use a symlink to the slurm.conf file a shared filesystem. This works great. Anytime there are changes, a simple 'scontrol reconfigure' brings all running nodes up to speed and any down nodes will automatically read the latest.

Brian Andrus

On 2/23/2022 2:31 AM, David Simpson wrote:

    Hi all,

    Interested to know what common approaches were to:


     1. Monitoring of power saving nodes (e.g. health of the node), when
        potentially the monitoring system will see it go up and down. Do
        you limit to BMC only monitoring/health?
     2. When you want to make changes to slurm.conf (or anything else)
        to a node which is down due to power saving (during a
        maintenance/reservation) what is your approach? Do you end up
        with 2 slurm.confs (one for power saving and one that keeps
        everything up, to work on during the maintenance)?


    thanks
    David


    -------------

    David Simpson - Senior Systems Engineer

    ARCCA, Redwood Building,

    King Edward VII Avenue,

    Cardiff, CF10 3NB

    David Simpson - peiriannydd uwch systemau

    ARCCA, Adeilad Redwood,

    King Edward VII Avenue,

    Caerdydd, CF10 3NB

    simpso...@cardiff.ac.uk <mailto:simpso...@cardiff.ac.uk>

    +44 29208 74657



Reply via email to