[slurm-users] QOS MaxTRESPU node=X intepretation

2024-08-30 Thread David Magda via slurm-users
Hello, Have a question on how to interpret a Node=X MaxTRESPU value for a QOS: If (e.g.) X=4, and each node has (say) 64 CPUs or cores: if a particular job needs 32 cores, then would two jobs count as the equivalent of one node (2*32=64)? And if X=4 for the QOS, would that mean that eight of th

[slurm-users] Re: Best practices for tracking jobs started across multiple clusters for accounting purposes.

2024-08-30 Thread Laura Hild via slurm-users
Can whatever is running those sbatch commands add a --comment with a shared identifier that AccountingStoreFlags=job_comment would make available in sacct? -- slurm-users mailing list -- slurm-users@lists.schedmd.com To unsubscribe send an email to slurm-users-le...@lists.schedmd.com

[slurm-users] Re: playing with --nodes=

2024-08-30 Thread Brian Andrus via slurm-users
Looks like it is not doing what you think it should. It does state: If the number of tasks is given and a number of requested nodes is also given, the number of nodes used from that request will be reduced to match that of the number of tasks if the number of nodes in the request is greater t

[slurm-users] Re: Best practices for tracking jobs started across multiple clusters for accounting purposes.

2024-08-30 Thread Di Bernardini, Fabio via slurm-users
For example if a job has to use different clusters with Slurm I am forced to launch it with two sbatch commands: sbatch -M cluster1 job1 sbatch -m cluster2 job2 This way I get two different jobids. Using sacct I have not found a way to know that the two jobs were launched within the same workfl