Strange. I set "limited" QoS with limit of cpu=560 and apply it to my queue.

[root@head ~]# sacctmgr show qos
\      Name   Priority  GraceTime    Preempt PreemptMode                        
            Flags UsageThres UsageFactor       GrpTRES   GrpTRESMins 
GrpTRESRunMin GrpJobs GrpSubmit     GrpWall       MaxTRES MaxTRESPerNode   
MaxTRESMins     MaxWall     MaxTRESPU MaxJobsPU MaxSubmitPU     MaxTRESPA 
MaxJobsPA MaxSubmitPA       MinTRES
---------- ---------- ---------- ---------- ----------- 
---------------------------------------- ---------- ----------- ------------- 
------------- ------------- ------- --------- ----------- ------------- 
-------------- ------------- ----------- ------------- --------- ----------- 
------------- --------- ----------- -------------
    normal          0   00:00:00                cluster                         
                               1.000000                                         
                                                                                
                                                                         
    limited         10   00:00:00                cluster                        
                                1.000000 cpu=560


For test, I run hostname from user on more than 560 cpu. It's run...

13655          hostname    work    limited       1232  COMPLETED      0:0

[root@head ~]# scontrol show job 13655 | grep cpu
   TRES=cpu=1232,node=22

-----Original Message-----
From: slurm-users <slurm-users-boun...@lists.schedmd.com> On Behalf Of Renfro, 
Michael
Sent: Wednesday, February 05, 2020 5:53 PM
To: Slurm User Community List <slurm-users@lists.schedmd.com>
Subject: Re: [slurm-users] Limits to partitions for users groups

If you want to rigidly define which 20 nodes are available to the one group of 
users, you could define a 20-node partition for them, and a 35-node partition 
for the priority group, and restrict access by Unix group membership:

PartitionName=restricted Nodes=node0[01-20] AllowGroups=ALL 
PartitionName=priority Nodes=node0[01-35] AllowGroups=prioritygroup

If you don’t care which of the 35 nodes get used by the first group, but want 
to restrict them to using at most 20 nodes of the 35, you could define a single 
partition and a QOS for each group:

PartitionName=restricted Nodes=node0[01-35] AllowGroups=ALL QoS=restricted 
PartitionName=priority Nodes=node0[01-35] AllowGroups=prioritygroup QoS=priority

sacctmgr add qos restricted
sacctmgr modify qos restricted set grptres=cpu=N # where N=20*(cores per node) 
sacctmgr add qos priority sacctmgr modify qos restricted set grptres=cpu=-1 # 
might not be strictly required


> On Feb 5, 2020, at 8:07 AM, Рачко Антон Сергеевич <an...@ciam.ru> wrote:
> 
> External Email Warning
> This email originated from outside the university. Please use caution when 
> opening attachments, clicking links, or responding to requests.
> I have partition with 35 nodes. Many users use it, but one group of them have 
> more priority than others. I want to set limit of max. 20 nodes for any users 
> and allow use all nodes for users in priority group.
> I can split this partition to 2: 20-node partition for all and 15-node for 
> priority group. Can I do it otherwise (sacctmg, QOS, etc.)?

Reply via email to