On Mon, 14 Jun 2021 09:33:02 +0200 (CEST)
Arthur Gilly wrote:
> Hi all,
>
> A related question, on my setup, scontrol show job displays the
> standard output, standard error redirections as well as the wd,
> whereas this info is lost after completion when sacct is required. Is
> this something t
On our slurm clusters the module system (Lmod) works without extra init
in job scripts due to the environment-forwarding in slurm. "module" in
the submitting context (in bash) on the login node is an "exported"
function and as such makes it across.
/Peter
On Fri, 22 Jan 2021 10:41:06 +
Gestió
On Wed, 18 Nov 2020 09:15:59 -0500
Jason Simms wrote:
> Dear Diego,
>
> A while back, I attempted to make some edits locally to see whether I
> could produce "better" results. Here is a comparison of the output of
> your latest version, and then mine:
I'm not sure what bug or behavior you're se
On Wed, 6 May 2020 10:42:46 +0100
Killian Murphy wrote:
> Hi all.
>
> I'm probably making a rookie error here...which 'megabyte' (powers of
> 1000 or 1024) does the Slurm documentation refer to in, for example,
> the slurm.conf documentation for RealMemory and the sbatch
> documentation for `--m
On Mon, 27 Apr 2020 14:51:01 +0530
Sudeep Narayan Banerjee wrote:
> Dear All,
>
> I have 360 cpu cores in my cluster; 9 compute nodes with 20core x 2
> sockets each.
>
> I have slurm.18.08.7 version and have multifactor (fair share) and
> backfill enabled.
>
> I am running jobs with less nta
:15 +0530
Sudeep Narayan Banerjee wrote:
> Dear Peter: I am trying with *sacct* and multiple flags.. but am not
> getting the desired output as per the query...
>
> Thanks & Regards,
> Sudeep Narayan Banerjee
>
> On 02/04/20 5:23 pm, Peter Kjellström wrote:
> > On
On Thu, 2 Apr 2020 16:57:46 +0530
Sudeep Narayan Banerjee wrote:
> any help in getting the right flags ?
You may need to clarify that question a bit...
How many users ran jobs on each day? (weekly, monthly average?)
How many jobs/per day did each user run? (weekly, monthly average?)
And what
On Mon, 29 Apr 2019 15:47:49 +
Michael Robbert wrote:
> I was curious what startup method other sites are using with Intel
> MPI?
We (nsc.liu.se nat. level academic HPC) have used IntelMPI+slurm as a
primary platform for several generations.
Our setup to this day is "mpiexec.hydra -bootstra
On Thu, 22 Nov 2018 01:51:59 +0800 (GMT+08:00)
宋亚磊 wrote:
> Hello everyone,
>
> How to check the percent cpu of a job in slurm? I tried sacct, sstat,
> squeue, but I can't find that how to check. Can someone help me?
I've written a small tool, jobload, that takes a jobid and outputs
current per
On Tue, 18 Sep 2018 15:00:58 -0400
David Rhey wrote:
> Hello, All,
>
> First time caller, long-time listener. Does anyone use any sort of
> external tool (e.g. a form submission) that generates accounts for
> their Slurm environment (notably for new accounts/allocations)? An
> example of this wo
On Thu, 19 Jul 2018 18:57:09 +0800
"Bill" wrote:
> Hi ,
>
>
> I just found the way , set "DefMemPerCPU=4096" for partition in
> slurm.conf
>
> It will use 4G memory request.
That is how we do it too (except not for a specific partition but
globally).
You can also add custom logic to a submi
On Wed, 11 Jul 2018 14:10:51 +0200
Alexander Grund wrote:
> Hi all,
>
> is it expected/intended that the env variable SLURM_NTASKS is not
> defined after salloc? It only gets defined after the an srun command.
> The number of tasks appear in `scontrol -d show job ` though.
> So is it a bug in o
On Fri, 13 Apr 2018 13:49:56 +0430
Mahmood Naderan wrote:
> Hi,
> I see some old posts on the web about performance comparison of srun
> vs. mpirun. Is that still an issue? Both the following scripts works
> for test programs and surely the performance concerns is not visible
> here.
...
> #SBAT
On Wed, 28 Feb 2018 06:51:15 +1100
Chris Samuel wrote:
> On Wednesday, 28 February 2018 2:13:41 AM AEDT Miguel Gila wrote:
>
> > Microcode patches were not applied to the physical system, only the
> > kernel was upgraded, so I'm not sure whether the performance hit
> > could come from that or no
On Tue, 14 Nov 2017 14:58:00 +
Zohar Roe MLM wrote:
> Hello,
> Trying again with the slurm.conf This time.
>
> I have a cluster name: Autobot
> In this cluster I have servers:
> Optimus[1-10] and
> Megatron[1-10].
>
> I sent 3000 jobs with feature Optimus and part are running while part
> a
15 matches
Mail list logo