Hi,

Yes, cpu == core. The legacy naming is somewhat unfortunate. I'll update
the page to clarify.

The 1 slot - 1 core approach does not work for us because:
* each task may use more than one core (e.g., openmp, multithread)
* specifying ncpus allows for grouping that better matches the application
needs (e.g., geometry, grid size)
* we want the user to be able to select based on various tradeoffs

Our job requests take the form of:
#$ -l pe dev <nslots>
#$ -l ncpus=<count>,mem=<Mbytes>,tmpfs=<Mbytes>

This way each slot gets a specific amount of cpus, memory, and dedicated
tmpfs. Each of the slots is supported by an LXC container with its own
contstraints, networking, etc.

Where we do the above, the cpus are dedicated. So, there is no
overallocation. The cpus are either available or not.

John

On Fri, 2018-08-31 at 12:58 +0200, Reuti wrote:

Hi John,


Am 31.08.2018 um 12:27 schrieb Marshall2, John (SSC/SPC) 
<john.marsha...@canada.ca<mailto:john.marsha...@canada.ca>>:


Hi,


When gridengine calculates cpu usage (based on wallclock) it uses:

cpu usage = wallclock * nslots


This does not account for the number of cpus that may be used for

each slot, which is problematic.


What was the motivation for implementing it this way? I mean: usually one SLOT 
represents one CORE in GirdEngine. Hence, to get a proper allocation and 
accounting while not oversubscribe the nodes you have to request the overall 
amount of cores in case you want to combine processes (like for MPI) and 
threads (like for Open MP).


In your case, it looks to me that you assume that necessary cores are 
available, independent from the actual usage of each node?


-- Reuti


PS: I assume with CPUS you refer to CORES.



I have written up an article at:

https://expl.info/display/MISC/Slot+Multiplier+for+Calculating+CPU+Usage+in+Gridengine


which explains the issue and provides a patch (against sge-8.1.9)

so that:

cpu usage = wallclock * nslots * ncpus_per_slot


This makes the usage information much more useful/accurate

when using the fair share.


Have others encountered this issue? Feedback is welcome.


Thanks,

John

_______________________________________________

users mailing list

users@gridengine.org<mailto:users@gridengine.org>

https://gridengine.org/mailman/listinfo/users


_______________________________________________
users mailing list
users@gridengine.org
https://gridengine.org/mailman/listinfo/users

Reply via email to