Dear SLURM experts,

we have a cluster of 56 nodes with 28 cores each. Is it possible to limit the number of jobs of a certain name which concurrently run on
one node, without blocking the node for other jobs?

For example, when I do

  for filename in runtimes/*/jobscript.sh; do
    sbatch -J iojob -n 1 $filename
  done

How can I assure that only one of these jobs runs per node? The jobs
are very lightweight computationally and only use 1 core each, but
since they are rather heavy on the I/O side, I'd like to ensure that when a job runs, it doesn't have to share the available I/O bandwidth with other jobs. (This would actually work since usually our other
jobs are not I/O intensive.)

From reading the manpage, I couldn't figure out how to do this.


Sunny greetings,
Andreas

Reply via email to