Hi Miguel, I modified my test configuration to evaluate the effect of NoDecay.
I modified all QOS adding NoDecay Flag. toto@login1:~/TEST$ sacctmgr show QOS Name Priority GraceTime Preempt PreemptExemptTime PreemptMode Flags UsageThres UsageFactor GrpTRES GrpTRESMins GrpTRESRunMin GrpJobs GrpSubmit GrpWall MaxTRES MaxTRESPerNode MaxTRESMins MaxWall MaxTRESPU MaxJobsPU MaxSubmitPU MaxTRESPA MaxJobsPA MaxSubmitPA MinTRES ---------- ---------- ---------- ---------- ------------------- ----------- ---------------------------------------- ---------- ----------- ------------- ------------- ------------- ------- --------- ----------- ------------- -------------- ------------- ----------- ------------- --------- ----------- ------------- --------- ----------- ------------- normal 0 00:00:00 cluster NoDecay 1.000000 interactif 10 00:00:00 cluster NoDecay 1.000000 node=50 node=22 1-00:00:00 node=50 petit 4 00:00:00 cluster NoDecay 1.000000 node=1500 node=22 1-00:00:00 node=300 gros 6 00:00:00 cluster NoDecay 1.000000 node=2106 node=700 1-00:00:00 node=700 court 8 00:00:00 cluster NoDecay 1.000000 node=1100 node=100 02:00:00 node=300 long 4 00:00:00 cluster NoDecay 1.000000 node=500 node=200 5-00:00:00 node=200 special 10 00:00:00 cluster NoDecay 1.000000 node=2106 node=2106 5-00:00:00 node=2106 support 10 00:00:00 cluster NoDecay 1.000000 node=2106 node=700 1-00:00:00 node=2106 visu 10 00:00:00 cluster NoDecay 1.000000 node=4 node=700 06:00:00 node=4 I submitted a bunch of jobs to control the NoDecay efficiency and I noticed RawUsage as well as GrpTRESRaw cpu is still decreasing. toto@login1:~/TEST$ sshare -A dci -u " " -o account,user,GrpTRESRaw%80, GrpTRESMins ,RawUsage Account User GrpTRESRaw GrpTRESMins RawUsage -------------------- ---------- ----------------------------------------------------- ------------------------------ ----------- dci cpu=6932 ,mem=12998963,energy=0,node=216,billing=6932,fs/disk=0,vmem=0,pages=0 cpu=17150 415966 toto@login1:~/TEST$ sshare -A dci -u " " -o account,user,GrpTRESRaw%80, GrpTRESMins , RawUsage Account User GrpTRESRaw GrpTRESMins RawUsage -------------------- ---------- ----------------------------------------------------- ------------------------------ ----------- dci cpu=6931 ,mem=12995835,energy=0,node=216,billing=6931,fs/disk=0,vmem=0,pages=0 cpu=17150 415866 toto@login1:~/TEST$ sshare -A dci -u " " -o account,user,GrpTRESRaw%80,GrpTRESMins,RawUsage Account User GrpTRESRaw GrpTRESMins RawUsage -------------------- ---------- ----------------------------------------------------- ------------------------------ ----------- dci cpu=6929 ,mem=12992708,energy=0,node=216,billing=6929,fs/disk=0,vmem=0,pages=0 cpu=17150 415766 Something I forgot to do ? Best, Gérard Cordialement, Gérard Gil Département Calcul Intensif Centre Informatique National de l'Enseignement Superieur 950, rue de Saint Priest 34097 Montpellier CEDEX 5 FRANCE tel : (334) 67 14 14 14 fax : (334) 67 52 37 63 web : [ http://www.cines.fr/ | http://www.cines.fr ] > De: "Gérard Gil" <gerard....@cines.fr> > À: "Slurm-users" <slurm-users@lists.schedmd.com> > Cc: "slurm-users" <slurm-us...@schedmd.com> > Envoyé: Vendredi 24 Juin 2022 14:52:12 > Objet: Re: [slurm-users] GrpTRESMins and GrpTRESRaw usage > Hi Miguel, > Good !! > I'll try this options on all existing QOS and see if everything works as > expected. > I'll inform you on the results. > Thanks a lot > Best, > Gérard > ----- Mail original ----- >> De: "Miguel Oliveira" <miguel.olive...@uc.pt> >> À: "Slurm-users" <slurm-users@lists.schedmd.com> >> Cc: "slurm-users" <slurm-us...@schedmd.com> >> Envoyé: Vendredi 24 Juin 2022 14:07:16 >> Objet: Re: [slurm-users] GrpTRESMins and GrpTRESRaw usage >> Hi Gérard, >> I believe so. All our accounts correspond to one project and all have an >> associated QoS with NoDecay and DenyOnLimit. This is enough to restrict usage >> on each individual project. >> You only need these flags on the QoS. The association will carry on as usual >> and >> fairshare will not be impacted. >> Hope that helps, >> Miguel Oliveira >>> On 24 Jun 2022, at 12:56, gerard....@cines.fr wrote: >>> Hi Miguel, >>>> Why not? You can have multiple QoSs and you have other techniques to change >>>> priorities according to your policies. >>> Is this answer my question ? >>> "If all configured QOS use NoDecay, we can take advantage of the FairShare >>> priority with Decay and all jobs GrpTRESRaw with NoDecay ?" >>> Thanks >>> Best, > > > Gérard
smime.p7s
Description: S/MIME Cryptographic Signature