I’ve suspected for some time that this matters in our environment, though we /do/ use GPFS. Maybe any use of local scratch (XFS, local drive) could figure in here?
Are there any tips for how to determine easily where the extra money is coming from, for example when the user has specifically constrained application to a certain amount of memory with its own flags, or – to put another way – to prove that it’s not this sort of phenomenon happening? On Jun 21, 2019, at 13:04, Christopher Samuel <ch...@csamuel.org<mailto:ch...@csamuel.org>> wrote: On 6/13/19 5:27 PM, Kilian Cavalotti wrote: I would take a look at the various *KmemSpace options in cgroups.conf, they can certainly help with this. Specifically I think you'll want: ConstrainKmemSpace=no to fix this. This happens for NFS and Lustre based systems, I don't think it's a problem for GPFS as mmfsd has its own pagepool separate to the processes address space. All the best, Chris -- Chris Samuel : http://www.csamuel.org/ : Berkeley, CA, USA -- ____ || \\UTGERS, |---------------------------*O*--------------------------- ||_// the State | Ryan Novosielski - novos...@rutgers.edu<mailto:novos...@rutgers.edu> || \\ University | Sr. Technologist - 973/972.0922 (2x0922) ~*~ RBHS Campus || \\ of NJ | Office of Advanced Research Computing - MSB C630, Newark `'