Hi Tim,
we are using the container/tmpfs plugin to map /tmp to a local NVMe
drive which works great. I did consider setting up directory quotas. I
thought the InitScript [1] option should do the trick. Alas, I didn't
get it to work. If I remember correctly, slurm complained about the
option being present. In the end we recommend our users to make
exclusive use a node if they are going to use a lot of local scratch
space. I don't think this happens very often if at all.
Regards
magnus

[1] 
https://slurm.schedmd.com/job_container.conf.html#OPT_InitScript


On Tue, 2024-02-06 at 14:39 +0100, Tim Schneider via slurm-users wrote:
> Hi,
> 
> In our SLURM cluster, we are using the job_container/tmpfs plugin to 
> ensure that each user can use /tmp and it gets cleaned up after them.
> Currently, we are mapping /tmp into the nodes RAM, which means that
> the 
> cgroups make sure that users can only use a certain amount of storage
> inside /tmp.
> 
> Now we would like to use of the node's local SSD instead of its RAM
> to 
> hold the files in /tmp. I have seen people define local storage as
> GRES, 
> but I am wondering how to make sure that users do not exceed the
> storage 
> space they requested in a job. Does anyone have an idea how to
> configure 
> local storage as a proper tracked resource?
> 
> Thanks a lot in advance!
> 
> Best,
> 
> Tim
> 
> 

-- 
Magnus Hagdorn
Charité – Universitätsmedizin Berlin
Geschäftsbereich IT | Scientific Computing
 
Campus Charité Mitte
BALTIC - Invalidenstraße 120/121
10115 Berlin
 
magnus.hagd...@charite.de
https://www.charite.de
HPC Helpdesk: sc-hpc-helpd...@charite.de

Attachment: smime.p7s
Description: S/MIME cryptographic signature

-- 
slurm-users mailing list -- slurm-users@lists.schedmd.com
To unsubscribe send an email to slurm-users-le...@lists.schedmd.com

Reply via email to