Hello, everyone, I'll answer everyone in a single reply because I've reached a conclusion: I'll give up on the idea of using shared nodes and will require exclusive allocation to a whole node. The final command line used will be:
sbatch -N 1 --exclusive --ntasks-per-node=1 --mem=0 pz-train.batch Thank you everyone for the discussion, On Mon, Aug 5, 2024 at 5:27 AM Daniel Letai via slurm-users <slurm-users@lists.schedmd.com> wrote: > > I think the issue is more severe than you describe. > > > Slurm juggles the needs of many jobs. Just because there are some resources > available at the exact second a job starts, doesn't mean those resource are > not pre-allocated for some future job waiting for even more resources, or > what about the use case of the opportunistic job being a backfill job, and > prevents a higher priority job from starting, or being pushed back due to > asking more resources at the last minute? > > > The request, while understandable from a user's point of view, is a > non-starter for a shared cluster. > > > Just my 2 cents. > > > On 02/08/2024 17:34, Laura Hild via slurm-users wrote: > > My read is that Henrique wants to specify a job to require a variable number > of CPUs on one node, so that when the job is at the front of the queue, it > will run opportunistically on however many happen to be available on a single > node as long as there are at least five. > > I don't personally know of a way to specify such a job, and wouldn't be > surprised if there isn't one, since as other posters have suggested, usually > there's a core-count sweet spot that should be used, achieving a performance > goal while making efficient use of resources. A cluster administrator may in > fact not want you using extra cores, even if there's a bit more speed-up to > be had, when those cores could be used more efficiently by another job. I'm > also not sure how one would set a judicious TimeLimit on a job that would > have such a variable wall-time. > > So there is the question of whether it is possible, and whether it is > advisable. > > > -- > slurm-users mailing list -- slurm-users@lists.schedmd.com > To unsubscribe send an email to slurm-users-le...@lists.schedmd.com -- Henrique Dante de Almeida hda...@gmail.com -- slurm-users mailing list -- slurm-users@lists.schedmd.com To unsubscribe send an email to slurm-users-le...@lists.schedmd.com