There are two points here 1. slurm(stepd) is unable to put the processes in the (null) cgroup. at first glance, this looks more of a slurm jus configuration 2. the MPI process forking. though this has a much better support than in the past, that might not always work, especially with fast interconnects. since you are running on a single node, you should be fine. simply export OMPI_MCA_mpi_warn_on_fork=0 before invoking srun, in order to silence this message.
Cheers, Gilles On Monday, June 20, 2016, Ahmed Rizwan <rizwan.ah...@aalto.fi <javascript:_e(%7B%7D,'cvml','rizwan.ah...@aalto.fi');>> wrote: > Dear MPI users, > > I am getting the errors below while submitting/executing following script, > > #!/bin/sh > #SBATCH -p short > #SBATCH -J layers > #SBATCH -n 12 > #SBATCH -N 1 > #SBATCH -t 01:30:00 > #SBATCH --mem-per-cpu=2500 > #SBATCH --exclusive > #SBATCH --mail-type=END > #SBATCH --mail-user=rizwan.ah...@aalto.fi > #SBATCH -o output_%j.out > #SBATCH -e errors_%j.err > > srun --mpi=pmi2 gpaw-python layers.py > > -------------------------------------------------------------------------- > slurmstepd: error: task/cgroup: unable to add task[pid=126453] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=80379] to memory cg > '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124258] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124259] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124261] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124266] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124264] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124262] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124260] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124265] to memory > cg '(null)' > slurmstepd: error: task/cgroup: unable to add task[pid=124263] to memory > cg '(null)' > -------------------------------------------------------------------------- > An MPI process has executed an operation involving a call to the > "fork()" system call to create a child process. Open MPI is currently > operating in a condition that could result in memory corruption or > other system errors; your MPI job may hang, crash, or produce silent > data corruption. The use of fork() (or system() or other calls that > create child processes) is strongly discouraged. > > The process that invoked fork was: > > Local host: pe38 (PID 80379) > MPI_COMM_WORLD rank: 1 > > If you are *absolutely sure* that your application will successfully > and correctly survive a call to fork(), you may disable this warning > by setting the mpi_warn_on_fork MCA parameter to 0. > -------------------------------------------------------------------------- > > Is this error fatal or should it be ignored? Thanks > Regards, > Rizwan >