Thank you, for the reply I was beginning to wonder if my message was seen.
While I understand how batch systems work, if you have a system daemon that develops a memory leak and consumes the memory outside of allocation. Not checking the used memory on the box before dispatch seems like a good way to black hole a bunch of jobs. On Sat, May 5, 2018 at 7:21 AM, Chris Samuel <ch...@csamuel.org> wrote: > On Thursday, 26 April 2018 3:28:19 AM AEST Cory Holcomb wrote: > > > It appears that I have a configuration that only takes into account the > > allocated memory before dispatching. > > With batch systems the idea is for the users to set constraints for their > jobs > so the scheduler can backfill other jobs onto nodes knowing how much > memory > they can rely on. > > So really the emphasis is on making the users set good resource requests > (such > as memory) and for the batch system to terminate jobs that exceed them (or > to > arrange for the kernel to constrain them to that amount via cgroups). > > Hope that helps! > Chris > -- > Chris Samuel : http://www.csamuel.org/ : Melbourne, VIC > > >