Hi,
if by "share the GPU" you mean exclusive allocation to a single job then, I
believe, you are missing cgroup configuration for isolating access to the GPU.
Below the relevant parts (I believe) of our configuration.
There also is a way of time- and space-slice GPUs but I guess you should g
s(slurm.reservations) do
slurm.log_user("i = %s", inspect(a))
slurm.log_user(inspect(b.start_time))
slurm.log_user("--")
end
might print
srun: i = "root_1"
srun: 1630598400
srun: --
Maybe that's useful for someone.
Best wishes,
Manuel
_
Hi,
I have trouble finding examples for job_submit.lua code that uses the global
Slurm reservations. I want to adjust the maximal running time of jobs with
respect to the closest reservation flagged as "maintenance". However, I cannot
find any job_submit.lua coda that accesses the Slurm reserva
Dear all,
I'm using the slurm.conf file from the attachment. I have some partitions,
e.g., "long", that apply to the same list of nodes but that are disjoint to the
partition "highmem". "highmem" only allows a user to use one node at a time. I
have a user who has submitted many jobs to highmem.
tart, it’ll get pushed back like you see here.
On Aug 31, 2020, at 12:13 PM, Holtgrewe, Manuel
wrote:
Dear all,
I'm seeing some user's job getting a StartTime 3 days in the future although
there are plenty of resources available in the the partition (and the user is
well below m
Dear all,
I'm seeing some user's job getting a StartTime 3 days in the future although
there are plenty of resources available in the the partition (and the user is
well below maxTRESPU of the partition).
Attached is our slurm.conf and the dump of "sacctmgr list qos -P". I'd be
grateful for an
Dear all,
I have a Slurm setup that has a couple of partitions and I want to limit the
number of CPUs that each user can use in each partition. I'm trying to do this
by attaching a qos to each partition and setting "MaxTRESPerUser" in each to,
e.g., cpu=1000. It looks like this setting is ignor
Hi,
I have some trouble understanding the "Oversubscribe" setting completely. What
I would like is to oversubscribe nodes to increase overall throughput.
- Is there a way to oversubscribe by a certain fraction, e.g. +20% or +50%?
- Is there a way to stop if a node reaches 100% "Load"?
Is there
dmon wrote:
>>>
>>> Generally the way we've solved this is to set aside a specific set of
>>> nodes in a partition for interactive sessions. We deliberately scale
>>> the size of the resources so that users will always run immediately and
>>> we al
Hi,
is there a way to make interactive logins where users will use almost no
resources "always succeed"?
In most of these interactive sessions, users will have mostly idle shells
running and do some batch job submissions. Is there a way to allocate "infinite
virtual cpus" on each node that can
pread amongst multiple nodes.
Mike
From: slurm-users on behalf of
"Holtgrewe, Manuel"
Reply-To: Slurm User Community List
Date: Friday, May 8, 2020 at 03:28
To: "slurm-users@lists.schedmd.com"
Subject: [External] [slurm-users] Defining a default --nodes=1
CAUTION: This emai
Dear all,
we're running a cluster where the large majority of jobs will use
multi-threading and no message passing. Sometimes CPU>1 jobs are scheduled to
run on more than one node (which would be fine for MPI jobs of course...)
Is it possible to automatically set "--nodes=1" for all jobs outsid
Hi,
is it possible to display the effective QOS of a job?
I need to investigate some unexpected behaviour of the scheduler (at least
unexpected to me at the moment). I want to limit maximum number of CPUs per
user in each partition. It is my understanding from the documentation that
partition
Dear all,
is it possible to find out which GPU was assigned to which job through squeue
or sacct?
My motivation is as follows: some users write jobs with bad resource usage
(e.g., 1h CPU to precompute, followed by 1h GPU to process, and so on). I don't
care so much about CPUs at the moment as
Dear all,
I would like to limit the number of cores used in by each user in a partition.
Is this possible?
Thanks,
--
Dr. Manuel Holtgrewe, Dipl.-Inform.
Bioinformatician
Core Unit Bioinformatics – CUBI
Berlin Institute of Health / Max Delbrück Center for Molecular Medicine in the
Helmholtz As
15 matches
Mail list logo