Dear all,

     Does anyone know how to set #SBATCH options to get multiple GPU cards
from different worker nodes?

     One of our users would like to apply for 16 NVIDIA V100 cards for his
job, and  there are 8 GPU cards on each worker node, I have tried the
following #SBATCH options:

      #SBATCH --partition=gpu
      #SBATCH --qos=normal
      #SBATCH --account=u07
      #SBATCH --job-name=cross
      #SBATCH --nodes=2
      #SBATCH --mem-per-cpu=1024
      #SBATCH --output=test.32^4.16gpu.log
      #SBATCH --gres=gpu:v100:16

      but got the sbatch error message :
      sbatch: error: Batch job submission failed: Requested node
configuration is not available

      And I found a similar question on stack overflow:

https://stackoverflow.com/questions/45200926/how-to-access-to-gpus-on-different-nodes-in-a-cluster-with-slurm

      And it is said that multiple GPU cards allocation on different worker
nodes are not available, the post is in 2017, is it still true at present?

      Thanks a lot for your help.

Best regards,
Ran

Reply via email to