and Lustre file
system.
It’s easy to extend with more metrices.
If you want more information on what we are doing just send me an email and I
can give you more information.
/Magnus
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet
By sending an email to Umeå University, the University will
ad
Advanced Computing Core Laboratory
King Abdullah University of Science and Technology
Building #1, Office #0124
greg.wick...@kaust.edu.sa <mailto:greg.wick...@kaust.edu.sa> +966 544
700 330
--
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet
mobile Huawei
Message original
Objet : [slurm-users] Two jobs ends up on one GPU?
De : Magnus Jonsson
À : slurm-users@lists.schedmd.com
Cc :
Hi!
We have machines with multiple GPUs (Nvidia V100).
We allow multiple (two) jobs on the nodes.
We have a user that in
only one GPU are requested) and only the defined /dev/nvidia?
device are accessable.
We are unable to reproduce this. Have anybody seen anything like this?
/Magnus
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet
?
This is an issue in a production environment. We don't want to have to
restart all the slurmctld daemons anytime there is a change to any
associations. That could get painful
Brian Andrus
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet
then
obvious solution of restarting the controller after every time we add
new members.
We are running 17.11.5 on both our clusters but we seen this since 17.02.
/Magnus
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet
bmitted later that a similar job submitted earlier,
he/she would be able to switch the priorities.
Is it just wishful thinking on my part or does something along those
lines exist?
Cheers,
Loris
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet
ure to give a range of priorities of you jobs
without being "root".
/M
--
Magnus Jonsson, Developer, HPC2N, Umeå Universitet