you don't need to be a subscriber to search bugs.schedmd.com

On Tue, Dec 10, 2024 at 9:44 AM Davide DelVento via slurm-users
<slurm-users@lists.schedmd.com> wrote:
>
> Good sleuthing.
>
> It would be nice if Slurm would say something like 
> Reason=Priority_Lower_Than_Job_XXXX so people will immediately find the 
> culprit in such situations. Has anybody with a SchedMD subscription ever 
> asked something like that, or is there some reasons for which it'd be 
> impossible (or too hard) information to gather programmatically?
>
> On Tue, Dec 10, 2024 at 1:09 AM Diego Zuccato via slurm-users 
> <slurm-users@lists.schedmd.com> wrote:
>>
>> Found the problem: another job was blocking access to the reservation.
>> The strangest thing is that the node (gpu03) has always been reserved
>> for a project, the blocking job did not explicitly request it (and even
>> if it did, it would have been denied access) but its state was:
>>     JobState=PENDING Reason=ReqNodeNotAvail,_UnavailableNodes:gpu03
>> Dependency=(null)
>>
>> Paint me surprised...
>>
>> Diego
>>
>> Il 07/12/2024 10:03, Diego Zuccato via slurm-users ha scritto:
>> > Ciao Davide.
>> >
>> > Il 06/12/2024 16:42, Davide DelVento ha scritto:
>> >
>> >> I find it extremely hard to understand situations like this. I wish
>> >> Slurm were more clear on how it reported what it is doing, but I
>> >> digress...
>> > I agree. A "scontrol explain" command could be really useful to pinpont
>> > the cause :)
>> >
>> >> I suspect that there are other job(s) which have higher priority than
>> >> this one which are supposed to run on that node but cannot start
>> >> because maybe this/these high-priority jobs(s) need(s) several nodes
>> >> and the other nodes are not available at the moment?
>> > That partition is a single node, and it's IDLE. If another job needed
>> > it, it would be in PLANNED state (IIRC).
>> >
>> >> Pure speculation, obviously, since I have no idea what the rest of
>> >> your cluster looks like, and what the rest of the workflow is, but the
>> >> clue/ hint is
>> >>
>> >>  > JobState=PENDING Reason=Priority Dependency=(null)
>> >>
>> >> You are pending because something else has higher priority. Going back
>> >> to my first sentence, I wish Slurm would say which one other job
>> >> (maybe there are more than one, but one would suffice for this
>> >> investigation) is trumping this job priority so one could more
>> >> clearly understand what is going on, without sleuthing.
>> > Couldn't agree more :) Scheduler is quite opaque in its decisions. :(
>> >
>> > Actually the job that the user submitted is not starting and has
>> > Reason=PartitionConfig . But QoS 'debug' (the one I'm using for testing)
>> > does have higher priority (1000) than QoS 'long' (10, IIRC).
>> >
>> > Diego
>> >
>> >> On Fri, Dec 6, 2024 at 7:36 AM Diego Zuccato via slurm-users <slurm-
>> >> us...@lists.schedmd.com <mailto:slurm-users@lists.schedmd.com>> wrote:
>> >>
>> >>     Hello all.
>> >>     An user reported that a job wasn't starting, so I tried to replicate
>> >>     the
>> >>     request and I get:
>> >>     -8<--
>> >>     [root@ophfe1 root.old]# scontrol show job 113936
>> >>     JobId=113936 JobName=test.sh
>> >>          UserId=root(0) GroupId=root(0) MCS_label=N/A
>> >>          Priority=1 Nice=0 Account=root QOS=long
>> >>          JobState=PENDING Reason=Priority Dependency=(null)
>> >>          Requeue=1 Restarts=0 BatchFlag=1 Reboot=0 ExitCode=0:0
>> >>          RunTime=00:00:00 TimeLimit=2-00:00:00 TimeMin=N/A
>> >>          SubmitTime=2024-12-06T13:19:36 EligibleTime=2024-12-06T13:19:36
>> >>          AccrueTime=2024-12-06T13:19:36
>> >>          StartTime=Unknown EndTime=Unknown Deadline=N/A
>> >>          SuspendTime=None SecsPreSuspend=0
>> >>     LastSchedEval=2024-12-06T13:21:32
>> >>     Scheduler=Backfill:*
>> >>          Partition=m3 AllocNode:Sid=ophfe1:855189
>> >>          ReqNodeList=(null) ExcNodeList=(null)
>> >>          NodeList=
>> >>          NumNodes=1-1 NumCPUs=96 NumTasks=96 CPUs/Task=1
>> >> ReqB:S:C:T=0:0:*:*
>> >>          TRES=cpu=96,mem=95000M,node=1,billing=1296
>> >>          Socks/Node=* NtasksPerN:B:S:C=0:0:*:* CoreSpec=*
>> >>          MinCPUsNode=1 MinMemoryNode=95000M MinTmpDiskNode=0
>> >>          Features=(null) DelayBoot=00:00:00
>> >>          OverSubscribe=OK Contiguous=0 Licenses=(null) Network=(null)
>> >>          Command=/home/root.old/test.sh
>> >>          WorkDir=/home/root.old
>> >>          StdErr=/home/root.old/%N-%J.err
>> >>          StdIn=/dev/null
>> >>          StdOut=/home/root.old/%N-%J.out
>> >>          Power=
>> >>
>> >>
>> >>     [root@ophfe1 root.old]# scontrol sho partition m3
>> >>     PartitionName=m3
>> >>          AllowGroups=ALL DenyAccounts=formazione AllowQos=ALL
>> >>          AllocNodes=ALL Default=NO QoS=N/A
>> >>          DefaultTime=NONE DisableRootJobs=NO ExclusiveUser=NO GraceTime=0
>> >>     Hidden=NO
>> >>          MaxNodes=UNLIMITED MaxTime=UNLIMITED MinNodes=0 LLN=NO
>> >>     MaxCPUsPerNode=UNLIMITED
>> >>          Nodes=mtx20
>> >>          PriorityJobFactor=1 PriorityTier=1 RootOnly=NO ReqResv=NO
>> >>     OverSubscribe=NO
>> >>          OverTimeLimit=NONE PreemptMode=CANCEL
>> >>          State=UP TotalCPUs=192 TotalNodes=1
>> >>     SelectTypeParameters=CR_SOCKET_MEMORY
>> >>          JobDefaults=(null)
>> >>          DefMemPerNode=UNLIMITED MaxMemPerNode=UNLIMITED
>> >>          TRES=cpu=192,mem=1150000M,node=1,billing=2592
>> >>          TRESBillingWeights=CPU=13.500,Mem=2.2378G
>> >>
>> >>     [root@ophfe1 root.old]# scontrol show node mtx20
>> >>     NodeName=mtx20 Arch=x86_64 CoresPerSocket=24
>> >>          CPUAlloc=0 CPUEfctv=192 CPUTot=192 CPULoad=0.00
>> >>          AvailableFeatures=ib,matrix,intel,avx
>> >>          ActiveFeatures=ib,matrix,intel,avx
>> >>          Gres=(null)
>> >>          NodeAddr=mtx20 NodeHostName=mtx20 Version=22.05.6
>> >>          OS=Linux 4.18.0-372.9.1.el8.x86_64 #1 SMP Tue May 10 14:48:47
>> >>     UTC 2022
>> >>          RealMemory=1150000 AllocMem=0 FreeMem=1156606 Sockets=4 Boards=1
>> >>          MemSpecLimit=2048
>> >>          State=IDLE ThreadsPerCore=2 TmpDisk=0 Weight=8 Owner=N/A
>> >>     MCS_label=N/A
>> >>          Partitions=m3
>> >>          BootTime=2024-12-06T10:01:42 SlurmdStartTime=2024-12-06T10:02:54
>> >>          LastBusyTime=2024-12-06T10:51:58
>> >>          CfgTRES=cpu=192,mem=1150000M,billing=2592
>> >>          AllocTRES=
>> >>          CapWatts=n/a
>> >>          CurrentWatts=0 AveWatts=0
>> >>          ExtSensorsJoules=n/s ExtSensorsWatts=0 ExtSensorsTemp=n/s
>> >>
>> >>     -8<--
>> >>
>> >>     So the node is free, the partition does not impose extra limits (used
>> >>     only for accounting factors) but the job does not start.
>> >>
>> >>     Any hints?
>> >>
>> >>     Tks
>> >>
>> >>     --     Diego Zuccato
>> >>     DIFA - Dip. di Fisica e Astronomia
>> >>     Servizi Informatici
>> >>     Alma Mater Studiorum - Università di Bologna
>> >>     V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>> >>     tel.: +39 051 20 95786
>> >>
>> >>
>> >>     --     slurm-users mailing list -- slurm-users@lists.schedmd.com
>> >>     <mailto:slurm-users@lists.schedmd.com>
>> >>     To unsubscribe send an email to slurm-users-le...@lists.schedmd.com
>> >>     <mailto:slurm-users-le...@lists.schedmd.com>
>> >>
>> >
>>
>> --
>> Diego Zuccato
>> DIFA - Dip. di Fisica e Astronomia
>> Servizi Informatici
>> Alma Mater Studiorum - Università di Bologna
>> V.le Berti-Pichat 6/2 - 40127 Bologna - Italy
>> tel.: +39 051 20 95786
>>
>>
>> --
>> slurm-users mailing list -- slurm-users@lists.schedmd.com
>> To unsubscribe send an email to slurm-users-le...@lists.schedmd.com
>
>
> --
> slurm-users mailing list -- slurm-users@lists.schedmd.com
> To unsubscribe send an email to slurm-users-le...@lists.schedmd.com

-- 
slurm-users mailing list -- slurm-users@lists.schedmd.com
To unsubscribe send an email to slurm-users-le...@lists.schedmd.com

Reply via email to