On 27 July 2018 at 03:13, Michael Robbert wrote:
> The line that you list from your slurm.conf shows the "course" partition
> being set as the default partition, but on our system the sinfo command
> shows our default partition with a * at the end and your output doesn't
> show that so I'm wonder
You'll need to check the slurm.conf man page for that parameter, to see
which combination of values are appropriate to achieve your site policies,
but among them you'll need "qos".
Regards,
Lyn
On Thu, Jul 26, 2018 at 2:29 PM, Siddharth Dalmia
wrote:
> Hi Lyn,
>
> Thanks for helping out. No I h
Hi Lyn,
Thanks for helping out. No I have not enforced any other limits, including
AccountingStorageEnforce.
Thanks
Sid
On Thu, Jul 26, 2018 at 5:06 PM Lyn Gerner wrote:
> HI,
>
> Have you enforced other limits successfully? What is the value of
> AccountingStorageEnforce?
>
> Regards,
> Lyn
>
I seem to recall conversations on this list in the past (although my Googling
hasn’t turned anything up) about sites that have developed tools to sanity
check slurm.conf. Can anyone speak to this? Or even better, would anyone be
willing to share their tool(s)? :)
We occasionally do silly things
HI,
Have you enforced other limits successfully? What is the value of
AccountingStorageEnforce?
Regards,
Lyn
On Thu, Jul 26, 2018 at 1:45 PM, Siddharth Dalmia
wrote:
>
> Hi all, We wanted to try make 2 different qos (priority and normal). For
> priority QOS - 1) Each user is only allowed 1 JOB
Hi all, We wanted to try make 2 different qos (priority and normal). For
priority QOS - 1) Each user is only allowed 1 JOB. 2) Has higher priority
than normal, with the ability to preempt. For priority QOS - 1) No
restriction on the number of jobs. Overall Jobs run with
`priority/multifactor`. I am
Hello all,
My error was indeed just the comma in my gres.conf. I was confused because
I had the same file on my running nodes but that's just because slurmd
started before the erroneous comma was added to the config.
So the error message was in fact directly correct, it could not find the
device
Generally it is best that they should be. Slurm maps the users
environment into the job submission. So if things change in the OSt
under it it can lead to issues.
-Paul Edmon-
On 07/26/2018 12:39 PM, Liam Forbes wrote:
Morning All.
I'm attempting to set up a new submit host with a differe
The line that you list from your slurm.conf shows the "course" partition
being set as the default partition, but on our system the sinfo command
shows our default partition with a * at the end and your output doesn't
show that so I'm wondering if you've got another partition that is
getting def
Morning All.
I'm attempting to set up a new submit host with a different OS (major)
version than the compute cluster. Should this be doable?
I have an OpenMPI app I can submit, using sbatch and qsub, from a cluster
node and it runs fine. However, when I submit from the new host it fails.
Initiall
Hi all,
I dont´t understand why its occurs!
user: john
group: courseit
partition: course
[john@master ~]$ sinfo
PARTITION AVAIL TIMELIMIT NODES STATE NODELIST
courseup infinite 8 idle node[02-04,06,09-12]
/etc/group
courseit:x:1002:john
/etc/passwd
john:x:1001:1002::/home/
Dear slurm users,
I'm having a hard time making slurm to correctly output progress
information during my program execution.
Apparently by default srun line-caches output, which is fine. Nevertheless
if we want to display some progress without a line break it seems not
possible. I tried using flush
12 matches
Mail list logo