Hello, everyone, I'll answer everyone in a single reply because I've
reached a conclusion: I'll give up on the idea of using shared nodes
and will require exclusive allocation to a whole node. The final
command line used will be:
sbatch -N 1 --exclusive --ntasks-per-node=1 --mem=0 pz-train.batch
Thank you everyone for the discussion,
On Mon, Aug 5, 2024 at 5:27 AM Daniel Letai via slurm-users
<[email protected]> wrote:
>
> I think the issue is more severe than you describe.
>
>
> Slurm juggles the needs of many jobs. Just because there are some resources
> available at the exact second a job starts, doesn't mean those resource are
> not pre-allocated for some future job waiting for even more resources, or
> what about the use case of the opportunistic job being a backfill job, and
> prevents a higher priority job from starting, or being pushed back due to
> asking more resources at the last minute?
>
>
> The request, while understandable from a user's point of view, is a
> non-starter for a shared cluster.
>
>
> Just my 2 cents.
>
>
> On 02/08/2024 17:34, Laura Hild via slurm-users wrote:
>
> My read is that Henrique wants to specify a job to require a variable number
> of CPUs on one node, so that when the job is at the front of the queue, it
> will run opportunistically on however many happen to be available on a single
> node as long as there are at least five.
>
> I don't personally know of a way to specify such a job, and wouldn't be
> surprised if there isn't one, since as other posters have suggested, usually
> there's a core-count sweet spot that should be used, achieving a performance
> goal while making efficient use of resources. A cluster administrator may in
> fact not want you using extra cores, even if there's a bit more speed-up to
> be had, when those cores could be used more efficiently by another job. I'm
> also not sure how one would set a judicious TimeLimit on a job that would
> have such a variable wall-time.
>
> So there is the question of whether it is possible, and whether it is
> advisable.
>
>
> --
> slurm-users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
--
Henrique Dante de Almeida
[email protected]
--
slurm-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]