If you would like the high watermark memory utilization after the job completes, https://github.com/NCAR/peak_memusage is a great tool. Of course it has the limitation that you need to know that you want that information *before* starting the job, which might or might not a problem for your use case
On Fri, Feb 9, 2024 at 10:07 AM Gerhard Strangar via slurm-users < slurm-users@lists.schedmd.com> wrote: > Hello, > > I'm wondering if there's a way to tell how much memory my job is using > per node. I'm doing > > #SBATCH -n 256 > srun solver inputfile > > When I run sacct -o maxvmsize, the result apparently is the maxmimum VSZ > of the largest solver process, not the maximum of the sum of them all > (unlike when calling mpirun instead). When I sstat -o TresUsageInMax, I > get the memory summed up over all nodes being used. Can I get the > maximum VSZ per node? > > > Gerhard > > -- > slurm-users mailing list -- slurm-users@lists.schedmd.com > To unsubscribe send an email to slurm-users-le...@lists.schedmd.com >
-- slurm-users mailing list -- slurm-users@lists.schedmd.com To unsubscribe send an email to slurm-users-le...@lists.schedmd.com