Thanks for the reply, Luke. I also found that with Bright they have a file
called /etc/security/pam_bright.d/pam_whitelist.conf that can be used to
allow access.

On Thu, May 9, 2024 at 5:10 AM Luke Sudbery <l.r.sudb...@bham.ac.uk> wrote:

> Draining a node will not stop someone logging on via pam_slurm_adopt.
>
>
>
> If they have a running job, and can log on when the node is not draining,
> then they can log on when it is draining.
>
>
>
> If they don’t have a running job, they can’t log on whether it is draining
> or not.
>
>
>
> If you want people to be able to log on when they don’t have a job
> running, you could put them in a group which is given access in access.conf
> and PAM, as explained here:
> https://slurm.schedmd.com/pam_slurm_adopt.html#admin_access
> <https://urldefense.proofpoint.com/v2/url?u=https-3A__slurm.schedmd.com_pam-5Fslurm-5Fadopt.html-23admin-5Faccess&d=DwMGaQ&c=aqMfXOEvEJQh2iQMCb7Wy8l0sPnURkcqADc2guUW8IM&r=X0jL9y0sL4r4iU_qVtR3lLNo4tOL1ry_m7-psV3GejY&m=930NtoLMP-HvoNP-dfQ9jhRtE5LJnxRDm9D7MJkOJnZQJRNbHHXjsP41nIQyfBxL&s=4p4zui4pf8xYjAj48y_0dCLnMEudAClm-bNhCYct-ZM&e=>
>
>
>
> Cheers,
>
>
>
> Luke
>
>
>
> --
>
> Luke Sudbery
>
> Principal Engineer (HPC and Storage).
>
> Architecture, Infrastructure and Systems
>
> Advanced Research Computing, IT Services
>
> Room 132, Computer Centre G5, Elms Road
>
>
>
> *Please note I don’t work on Monday.*
>
>
>
> *From:* Robert Kudyba via slurm-users <slurm-users@lists.schedmd.com>
> *Sent:* Friday, April 19, 2024 9:17 PM
> *To:* Slurm User Community List <slurm-users@lists.schedmd.com>
> *Subject:* [slurm-users] any way to allow interactive jobs or ssh in
> Slurm 23.02 when node is draining?
>
>
>
> *CAUTION:* This email originated from outside the organisation. Do not
> click links or open attachments unless you recognise the sender and know
> the content is safe.
>
>
>
> We use Bright Cluster Manager with SLurm 23.02 on RHEL9. I know about
> pam_slurm_adopt https://slurm.schedmd.com/pam_slurm_adopt.html
> <https://urldefense.proofpoint.com/v2/url?u=https-3A__slurm.schedmd.com_pam-5Fslurm-5Fadopt.html&d=DwMGaQ&c=aqMfXOEvEJQh2iQMCb7Wy8l0sPnURkcqADc2guUW8IM&r=X0jL9y0sL4r4iU_qVtR3lLNo4tOL1ry_m7-psV3GejY&m=930NtoLMP-HvoNP-dfQ9jhRtE5LJnxRDm9D7MJkOJnZQJRNbHHXjsP41nIQyfBxL&s=Kch4xC6o-kw7TW21LcDVPMjH1a0Zl7TL1l8FiTdvLyI&e=>
> which does not appear to come by default with the Bright 'cm' package of
> Slurm.
>
>
>
> Currently ssh to a node gets:
>
> Login not allowed: no running jobs and no WLM allocations
>
>
>
> We have 8 GPUs on a node so when we drain a node, which can have up to a 5
> day job, no new jobs can run. And since we have 20+ TB (yes TB) local
> drives, researchers have their work and files on them to retrieve.
>
>
>
> Is there a way to use /etc/security/access.conf to work around this at
> least temporarily until the reboot and then we can revert?
>
>
>
> Thanks!
>
>
>
> Rob
>
-- 
slurm-users mailing list -- slurm-users@lists.schedmd.com
To unsubscribe send an email to slurm-users-le...@lists.schedmd.com

Reply via email to