On Wed, 2017-08-23 at 01:26 -0600, Sema Atasever wrote: > > > Computing predictions by SVM... > slurmstepd: Job 3469 exceeded memory limit (4235584 > 2048000), being > killed > slurmstepd: Exceeded job memory limit > > > How can i fix this problem. >
Error messages often give useful information. In this case you haven't requested enough memory in your Slurm script. Memory can be set with `#SBATCH --mem=[mem][M|G|T]` directive (entire job) or `#SBATCH --mem-per-cpu=[mem][M|G|T]` (per core). As a rule of thumb, the maximum request per node should be based around total cores -1 (for system processes). All the best, -- Lev Lafayette, BA (Hons), GradCertTerAdEd (Murdoch), GradCertPM, MBA (Tech Mngmnt) (Chifley) HPC Support and Training Officer +61383444193 +61432255208 Department of Infrastructure Services, University of Melbourne
