[slurm-users] Re: sbatch: Node count specification invalid - when only specifying --ntasks

2024-06-11 Thread George Leaver via slurm-users
Hi Loris, > Doesn't splitting up your jobs over two partitions mean that either one of > the two partitions could be full, while the other has idle nodes? Yes, potentially, and we may move away from our current config at some point (it's a bit of a hangover from an SGE cluster.) Hasn't really b

[slurm-users] Re: sbatch: Node count specification invalid - when only specifying --ntasks

2024-06-11 Thread Loris Bennett via slurm-users
Hi George, George Leaver via slurm-users writes: > Hi Loris, > >> Doesn't splitting up your jobs over two partitions mean that either >> one of the two partitions could be full, while the other has idle >> nodes? > > Yes, potentially, and we may move away from our current config at some > point

[slurm-users] Re: srun hostname - Socket timed out on send/recv operation

2024-06-11 Thread Arnuld via slurm-users
I enabled "debug3" logging and saw this in the node log: error: mpi_conf_send_stepd: unable to resolve MPI plugin offset from plugin_id=106. This error usually results from a job being submitted against an MPI plugin which was not compiled into slurmd but was for job submission command. error: _se