Hi Bill,

if I understand the concept of fairshare correctly, this could result in
a situation where one institute uses all resources.

Because of this fairshare is out of the question as I have to enforce
the ratio between the institutes - I cannot allow usage that would
result in one institute using more than what they paied for. If an
institute doesn't use the resources they have to run idle (or power down).

You could compare my situation with running two clusters that use the
same base infrastructure. What I want to do is enable users of both
institutes to use both clusters - but for each point in time use a
maximum of nodes that belong to "their" cluster.


Regards,

        Uwe


Am 14.08.2014 um 14:34 schrieb Bill Barth:
> 
> Why not make one partition and use fairshare to balance the usage over
> time? That way both institutes can run large jobs that span the whole
> machine when others are not using it.
> 
> Bill.
> --
> Bill Barth, Ph.D., Director, HPC
> bba...@tacc.utexas.edu        |   Phone: (512) 232-7069
> Office: ROC 1.435             |   Fax:   (512) 475-9445
> 
> 
> 
> 
> 
> 
> 
> On 8/14/14, 4:11 AM, "Uwe Sauter" <uwe.sauter...@gmail.com> wrote:
> 
>>
>> Hi all,
>>
>> I got a question about a configuration detail: "dynamic partitions"
>>
>> Situation:
>> I operate a Linux cluster of currently 54 nodes for a cooperation of two
>> different institutes at the university. To reflect the ratio of cash
>> those institutes invested I configured SLURM with two partition, one for
>> each institute. Those partitions have assigned different numbers of
>> nodes in a hard way, e.g.
>>
>> PartitionName=InstA Nodes=n[01-20]
>> PartitionName=InstB Nodes=n[21-54]
>>
>> To improve availability in case nodes break (and perhaps save some
>> power) I'd like to configure SLURM in a way that jobs can be assigned
>> nodes from the whole pool, respecting the number of nodes each institute
>> bought.
>>
>>
>> Research so far:
>> There is an option for partition configuration called "MaxNodes" but the
>> man pages state that this restricts the maximum number of nodes PER JOB.
>> It probably is possible to get something similar working using limit
>> enforcment through accounting, but I haven't understood that part of
>> SLURM 100% yet.
>> BlueGene systems seem to have the ability for something alike but then
>> this is for IBM systems only.
>>
>>
>> Question:
>> Is it possible to configure SLURM so that both partitions could utilize
>> all nodes but respect a maximum number of nodes that may be used the
>> same time? Something like:
>>
>> PartitionName=InstA Nodes=n[01-54] MaxPartNodes=20
>> PartitionName=InstB Nodes=n[01-54] MaxPartNodes=34
>>
>>
>> So is there a way to achieve this using the confg file? Do I have to use
>> accounting to enfoce the limits? Or is there another way that I don't see?
>>
>>
>> Best regards,
>>
>>      Uwe Sauter

Reply via email to