We determined that this issue was actually due to not having an unlimited
memlock for the slurm user when the slurm service started. The work-around was
to simply restart slurm subsequent to boot and the new unlimited setting would
allow infiniband usage. Moving the startup script to runlevel 3
We recently updated and rebooted Infiniband-attached nodes, and now when
trying to schedule MPI jobs with slurm, we are seeing the following:
--
No OpenFabrics connection schemes reported that they were able to be
used on a s