Bill - thank you for the code. For the record, our queues explicitly block any requests greater than what the queue allows - MaxMemPerNode = 64 GiB's.
If there are any other ideas, feel free to comment. Sandor ________________________________ From: slurm-users <slurm-users-boun...@lists.schedmd.com> on behalf of Bill <b...@princeton.edu> Sent: Thursday, December 8, 2022 7:52 AM To: slurm-users@lists.schedmd.com <slurm-users@lists.schedmd.com> Subject: Re: [slurm-users] srun --mem issue CAUTION: This email originated from outside of the organization. Do not click links or open attachments unless you can confirm the sender and know the content is safe. If you use a job_submit.lua script just add if job_desc.pn_min_memory == 0 or job_desc.min_mem_per_cpu == 0 then log_info("slurm_job_submit: job from uid %d invalid memory request MaxMemPerNode", job_desc.user_id) return 2044 -- signal ESLURM_INVALID_TASK_MEMORY end Bill On 12/7/22 12:03 PM, Felho, Sandor wrote: > TransUnion is running a ten-node site using slurm with multiple queues. > We have an issue with --mem parameter. The is one user who has read the > slurm manual and found the --mem=0. This is giving the maximum memory on > the node (500 GiB's) for the single job. How can I block a --mem=0 request? > > We are running: > > * OS: RHEL 7 > * cgroups version 1 > * slurm: 19.05 > > Thank you, > > *Sandor Felho * > > Sr Consultant, Data Science & Analytics > >