Good sleuthing.

It would be nice if Slurm would say something like
Reason=Priority_Lower_Than_Job_XXXX so people will immediately find the
culprit in such situations. Has anybody with a SchedMD subscription ever
asked something like that, or is there some reasons for which it'd be
impossible (or too hard) information to gather programmatically?

On Tue, Dec 10, 2024 at 1:09 AM Diego Zuccato via slurm-users <
slurm-users@lists.schedmd.com> wrote:

> Found the problem: another job was blocking access to the reservation.
> The strangest thing is that the node (gpu03) has always been reserved
> for a project, the blocking job did not explicitly request it (and even
> if it did, it would have been denied access) but its state was:
>     JobState=PENDING Reason=ReqNodeNotAvail,_UnavailableNodes:gpu03
> Dependency=(null)
>
> Paint me surprised...
>
> Diego
>
> Il 07/12/2024 10:03, Diego Zuccato via slurm-users ha scritto:
> > Ciao Davide.
> >
> > Il 06/12/2024 16:42, Davide DelVento ha scritto:
> >
> >> I find it extremely hard to understand situations like this. I wish
> >> Slurm were more clear on how it reported what it is doing, but I
> >> digress...
> > I agree. A "scontrol explain" command could be really useful to pinpont
> > the cause :)
> >
> >> I suspect that there are other job(s) which have higher priority than
> >> this one which are supposed to run on that node but cannot start
> >> because maybe this/these high-priority jobs(s) need(s) several nodes
> >> and the other nodes are not available at the moment?
> > That partition is a single node, and it's IDLE. If another job needed
> > it, it would be in PLANNED state (IIRC).
> >
> >> Pure speculation, obviously, since I have no idea what the rest of
> >> your cluster looks like, and what the rest of the workflow is, but the
> >> clue/ hint is
> >>
> >>  > JobState=PENDING Reason=Priority Dependency=(null)
> >>
> >> You are pending because something else has higher priority. Going back
> >> to my first sentence, I wish Slurm would say which one other job
> >> (maybe there are more than one, but one would suffice for this
> >> investigation) is trumping this job priority so one could more
> >> clearly understand what is going on, without sleuthing.
> > Couldn't agree more :) Scheduler is quite opaque in its decisions. :(
> >
> > Actually the job that the user submitted is not starting and has
> > Reason=PartitionConfig . But QoS 'debug' (the one I'm using for testing)
> > does have higher priority (1000) than QoS 'long' (10, IIRC).
> >
> > Diego
> >
> >> On Fri, Dec 6, 2024 at 7:36 AM Diego Zuccato via slurm-users <slurm-
> >> us...@lists.schedmd.com <mailto:slurm-users@lists.schedmd.com>> wrote:
> >>
> >>     Hello all.
> >>     An user reported that a job wasn't starting, so I tried to replicate
> >>     the
> >>     request and I get:
> >>     -8<--
> >>     [root@ophfe1 root.old]# scontrol show job 113936
> >>     JobId=113936 JobName=test.sh
> >>          UserId=root(0) GroupId=root(0) MCS_label=N/A
> >>          Priority=1 Nice=0 Account=root QOS=long
> >>          JobState=PENDING Reason=Priority Dependency=(null)
> >>          Requeue=1 Restarts=0 BatchFlag=1 Reboot=0 ExitCode=0:0
> >>          RunTime=00:00:00 TimeLimit=2-00:00:00 TimeMin=N/A
> >>          SubmitTime=2024-12-06T13:19:36 EligibleTime=2024-12-06T13:19:36
> >>          AccrueTime=2024-12-06T13:19:36
> >>          StartTime=Unknown EndTime=Unknown Deadline=N/A
> >>          SuspendTime=None SecsPreSuspend=0
> >>     LastSchedEval=2024-12-06T13:21:32
> >>     Scheduler=Backfill:*
> >>          Partition=m3 AllocNode:Sid=ophfe1:855189
> >>          ReqNodeList=(null) ExcNodeList=(null)
> >>          NodeList=
> >>          NumNodes=1-1 NumCPUs=96 NumTasks=96 CPUs/Task=1
> >> ReqB:S:C:T=0:0:*:*
> >>          TRES=cpu=96,mem=95000M,node=1,billing=1296
> >>          Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=*
> >>          MinCPUsNode=1 MinMemoryNode=95000M MinTmpDiskNode=0
> >>          Features=(null) DelayBoot=00:00:00
> >>          OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null)
> >>          Command=/home/root.old/test.sh
> >>          WorkDir=/home/root.old
> >>          StdErr=/home/root.old/%N-%J.err
> >>          StdIn=/dev/null
> >>          StdOut=/home/root.old/%N-%J.out
> >>          Power=
> >>
> >>
> >>     [root@ophfe1 root.old]# scontrol sho partition m3
> >>     PartitionName=m3
> >>          AllowGroups=ALL DenyAccounts=formazione AllowQos=ALL
> >>          AllocNodes=ALL Default=NO QoS=N/A
> >>          DefaultTime=NONE DisableRootJobs=NO ExclusiveUser=NO
> GraceTime=0
> >>     Hidden=NO
> >>          MaxNodes=UNLIMITED MaxTime=UNLIMITED MinNodes=0 LLN=NO
> >>     MaxCPUsPerNode=UNLIMITED
> >>          Nodes=mtx20
> >>          PriorityJobFactor=1 PriorityTier=1 RootOnly=NO ReqResv=NO
> >>     OverSubscribe=NO
> >>          OverTimeLimit=NONE PreemptMode=CANCEL
> >>          State=UP TotalCPUs=192 TotalNodes=1
> >>     SelectTypeParameters=CR_SOCKET_MEMORY
> >>          JobDefaults=(null)
> >>          DefMemPerNode=UNLIMITED MaxMemPerNode=UNLIMITED
> >>          TRES=cpu=192,mem=1150000M,node=1,billing=2592
> >>          TRESBillingWeights=CPU=13.500,Mem=2.2378G
> >>
> >>     [root@ophfe1 root.old]# scontrol show node mtx20
> >>     NodeName=mtx20 Arch=x86_64 CoresPerSocket=24
> >>          CPUAlloc=0 CPUEfctv=192 CPUTot=192 CPULoad=0.00
> >>          AvailableFeatures=ib,matrix,intel,avx
> >>          ActiveFeatures=ib,matrix,intel,avx
> >>          Gres=(null)
> >>          NodeAddr=mtx20 NodeHostName=mtx20 Version=22.05.6
> >>          OS=Linux 4.18.0-372.9.1.el8.x86_64 #1 SMP Tue May 10 14:48:47
> >>     UTC 2022
> >>          RealMemory=1150000 AllocMem=0 FreeMem=1156606 Sockets=4
> Boards=1
> >>          MemSpecLimit=2048
> >>          State=IDLE ThreadsPerCore=2 TmpDisk=0 Weight=8 Owner=N/A
> >>     MCS_label=N/A
> >>          Partitions=m3
> >>          BootTime=2024-12-06T10:01:42
> SlurmdStartTime=2024-12-06T10:02:54
> >>          LastBusyTime=2024-12-06T10:51:58
> >>          CfgTRES=cpu=192,mem=1150000M,billing=2592
> >>          AllocTRES=
> >>          CapWatts=n/a
> >>          CurrentWatts=0 AveWatts=0
> >>          ExtSensorsJoules=n/s ExtSensorsWatts=0 ExtSensorsTemp=n/s
> >>
> >>     -8<--
> >>
> >>     So the node is free, the partition does not impose extra limits
> (used
> >>     only for accounting factors) but the job does not start.
> >>
> >>     Any hints?
> >>
> >>     Tks
> >>
> >>     --     Diego Zuccato
> >>     DIFA - Dip. di Fisica e Astronomia
> >>     Servizi Informatici
> >>     Alma Mater Studiorum - Università di Bologna
> >>     V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
> >>     tel.: +39 051 20 95786
> >>
> >>
> >>     --     slurm-users mailing list -- slurm-users@lists.schedmd.com
> >>     <mailto:slurm-users@lists.schedmd.com>
> >>     To unsubscribe send an email to slurm-users-le...@lists.schedmd.com
> >>     <mailto:slurm-users-le...@lists.schedmd.com>
> >>
> >
>
> --
> Diego Zuccato
> DIFA - Dip. di Fisica e Astronomia
> Servizi Informatici
> Alma Mater Studiorum - Università di Bologna
> V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
> tel.: +39 051 20 95786
>
>
> --
> slurm-users mailing list -- slurm-users@lists.schedmd.com
> To unsubscribe send an email to slurm-users-le...@lists.schedmd.com
>
-- 
slurm-users mailing list -- slurm-users@lists.schedmd.com
To unsubscribe send an email to slurm-users-le...@lists.schedmd.com

Reply via email to