Ask for 8 gpus on 2 nodes instead. In your script just change the 16 to 8 and it should do what you want.
You are currently asking for 2 nodes with 16 gpu each as Gres resources are per node. Antony On Mon, 15 Apr 2019, 09:08 Ran Du, <bella.ran...@gmail.com> wrote: > Dear all, > > Does anyone know how to set #SBATCH options to get multiple GPU cards > from different worker nodes? > > One of our users would like to apply for 16 NVIDIA V100 cards for his > job, and there are 8 GPU cards on each worker node, I have tried the > following #SBATCH options: > > #SBATCH --partition=gpu > #SBATCH --qos=normal > #SBATCH --account=u07 > #SBATCH --job-name=cross > #SBATCH --nodes=2 > #SBATCH --mem-per-cpu=1024 > #SBATCH --output=test.32^4.16gpu.log > #SBATCH --gres=gpu:v100:16 > > but got the sbatch error message : > sbatch: error: Batch job submission failed: Requested node > configuration is not available > > And I found a similar question on stack overflow: > > https://stackoverflow.com/questions/45200926/how-to-access-to-gpus-on-different-nodes-in-a-cluster-with-slurm > > And it is said that multiple GPU cards allocation on different > worker nodes are not available, the post is in 2017, is it still true at > present? > > Thanks a lot for your help. > > Best regards, > Ran >