Rafael,

Most HPC centers have scheduled downtime on a regular basis. Typically it's one day a month, but I I know that at Argonne National Lab, which is a DOE Leadership Computing Facility that house some of the largest supercomputers in the world for use by a large number of scientists, they take their systems off-line every Monday for maintenance.

Having regularly scheduled maintenance outages is pretty much necessary for any large environment. Otherwise, the users never let you take the clusters offline for maintenance. Once the system is offline for a few hours, a task like upgrading Slurm is pretty easy.

When I worked in a smaller environment, I didn't have regularly scheduled outages, but due to the small size of the environment, it was easy for me to ask/tell the users I needed to take the cluster off-line  with a few days notice w/o any complaints from the users. In larger environments, you'll always get pushback, which is why creating a policy of regularly scheduled maintenance outages is necessary.

Prentice

On 3/22/19 7:07 AM, Frava wrote:
Hi all,

I think it's not that easy to keep SLURM up to date in a cluster of more than 3k nodes with a lot of users. I mean, that cluster has only a little more than 2 years old and my today's submission got the JOBID 12711473, the queue has 9769 jobs (squeue | wc -l). In two years there were only two maintenances that impacted the users and each one was announced a few months prior. They told me that they actually plan to update SLURM but not until late 2019 because they have other things to do before that. Also, I'm the only one asking for heterogeneous jobs...

Rafael.

Le jeu. 21 mars 2019 à 22:19, Prentice Bisbal <pbis...@pppl.gov <mailto:pbis...@pppl.gov>> a écrit :

    On 3/21/19 4:40 PM, Reuti wrote:

    >> Am 21.03.2019 um 16:26 schrieb Prentice Bisbal
    <pbis...@pppl.gov <mailto:pbis...@pppl.gov>>:
    >>
    >>
    >> On 3/20/19 1:58 PM, Christopher Samuel wrote:
    >>> On 3/20/19 4:20 AM, Frava wrote:
    >>>
    >>>> Hi Chris, thank you for the reply.
    >>>> The team that manages that cluster is not very fond of
    upgrading SLURM, which I understand.
    >> As a system admin who manages clusters myself, I don't
    understand this. Our job is to provide and maintain resources for
    our users. Part of that maintenance is to provide updates for
    security, performance, and functionality (new features) reasons.
    HPC has always been a leading-edge kind if field, so I feel this
    is even more important for HPC admins.
    >>
    >> Yes, there can be issues caused by updates, but those can be
    with proper planning: Have a plan to do the actual upgrade, have a
    plan to test for issues, and have a plan to revert to an earlier
    version if issues are discovered. This is work, but it's really
    not all that much work, and this is exactly the work we are being
    paid to do as cluster admins.
    > Besides the work on the side of the admins, also the users are
    involved: exchanging libraries also means to run the test suites
    of their applications again.
    >
    > -- Reuti

    That implies the users actually wrote test suites. ;-)



Reply via email to