Hi all,
We’re running a small slurm dev cluster on Ubuntu and are facing issues with
MPI/PMIx after upgrading slurm from 23.02.5 to 23.11.3.
The first job step to use MPI within a job fails roughly 80% of the time but
following attempts to use MPI within the same job work fine. For the failing
Hello everyone !
Recently our users bought a cplex dynamic license and want to use it on
our slurm cluster.
I've installed the paid version of cplex within modules so authorized
user can load it with a simple module load cplex/2111 command but I
don't know how to manage and ensure slurm doesn'
The simple answer is to just add a line such as
Licenses=whatever:20
and then request your users to use the -L option as described at
https://slurm.schedmd.com/licenses.html
This works very well, however it does not do enforcement like Slurm does
with other resources. You will find posts in this
Hello all,
I've used the "scontrol write batch_script" command to output the job
submission script from completed jobs in the past, but for some reason, no
matter which job I specify, it tells me it is invalid. Any way to
troubleshoot this? Alternatively, is there another way - even if a manual
da
Are you using the job_script storage option? If so then you should be
able to get at it by doing:
sacct -B j JOBID
https://slurm.schedmd.com/sacct.html#OPT_batch-script
-Paul Edmon-
On 2/16/2024 2:41 PM, Jason Simms via slurm-users wrote:
Hello all,
I've used the "scontrol write batch_scrip
Yes, that is what we are also doing and it works well.
Note that requesting a batch script for another user, one sees nothing
(rather than an error message saying that one does not have permissions)
On Fri, Feb 16, 2024 at 12:48 PM Paul Edmon via slurm-users <
slurm-users@lists.schedmd.com> wrote:
Are you absolutely certain you’ve done it before for completed jobs? I would
not expect that to work for completed jobs, with the possible exception of very
recently completed jobs (or am I thinking of Torque?).
Other replies mention the relatively new feature (21.08?) to store the job
script i