“4 threads” In MPI, we refer to this as 4 ranks or 4 processes. So what is your question? Are you getting errors with PBS?
-Tom From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Mahmood Naderan Sent: Monday, July 31, 2017 9:27 AM To: Open MPI Users <users@lists.open-mpi.org> Subject: Re: [OMPI users] -host vs -hostfile Excuse me, my fault.. I meant nodes=2:ppn=2 is 4 threads. Regards, Mahmood On Mon, Jul 31, 2017 at 8:49 PM, r...@open-mpi.org<mailto:r...@open-mpi.org> <r...@open-mpi.org<mailto:r...@open-mpi.org>> wrote: ?? Doesn't that tell pbs to allocate 1 node with 2 slots on it? I don't see where you get 4 Sent from my iPad On Jul 31, 2017, at 10:00 AM, Mahmood Naderan <mahmood...@gmail.com<mailto:mahmood...@gmail.com>> wrote: OK. The next question is how touse it with torque (PBS)? currently we write this directive Nodes=1:ppn=2 which means 4 threads. Then we omit -np and -hostfile in the mpirun command. On 31 Jul 2017 20:24, "Elken, Tom" <tom.el...@intel.com<mailto:tom.el...@intel.com>> wrote: Hi Mahmood, With the -hostfile case, Open MPI is trying to helpfully run things faster by keeping both processes on one host. Ways to avoid this… On the mpirun command line add: -pernode (runs 1 process per node), oe -npernode 1 , but these two has been deprecated in favor of the wonderful syntax: --map-by ppr:1:node Or you could change your hostfile to: cluster slots=1 compute-0-0 slots=1 -Tom From: users [mailto:users-boun...@lists.open-mpi.org<mailto:users-boun...@lists.open-mpi.org>] On Behalf Of Mahmood Naderan Sent: Monday, July 31, 2017 6:47 AM To: Open MPI Users <users@lists.open-mpi.org<mailto:users@lists.open-mpi.org>> Subject: [OMPI users] -host vs -hostfile Hi, I have stuck at a problem which I don't remember that on previous versions. when I run a test program with -host, it works. I mean, the process spans to the hosts I specified. However, when I specify -hostfile, it doesn't work!! mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun -host compute-0-0,cluster -np 2 a.out **************************************************************************** * hwloc 1.11.2 has encountered what looks like an error from the operating system. * * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset 0xff00ffff) without inclusion! * Error occurred in topology.c line 1048 * * The following FAQ entry in the hwloc documentation may help: * What should I do when hwloc reports "operating system" warnings? * Otherwise please report this error message to the hwloc user's mailing list, * along with the output+tarball generated by the hwloc-gather-topology script. **************************************************************************** Hello world from processor cluster.hpc.org<http://cluster.hpc.org>, rank 1 out of 2 processors Hello world from processor compute-0-0.local, rank 0 out of 2 processors mahmood@cluster:mpitest$ cat hosts cluster compute-0-0 mahmood@cluster:mpitest$ /share/apps/computer/openmpi-2.0.1/bin/mpirun -hostfile hosts -np 2 a.out **************************************************************************** * hwloc 1.11.2 has encountered what looks like an error from the operating system. * * Package (P#1 cpuset 0xffff0000) intersects with NUMANode (P#1 cpuset 0xff00ffff) without inclusion! * Error occurred in topology.c line 1048 * * The following FAQ entry in the hwloc documentation may help: * What should I do when hwloc reports "operating system" warnings? * Otherwise please report this error message to the hwloc user's mailing list, * along with the output+tarball generated by the hwloc-gather-topology script. **************************************************************************** Hello world from processor cluster.hpc.org<http://cluster.hpc.org>, rank 0 out of 2 processors Hello world from processor cluster.hpc.org<http://cluster.hpc.org>, rank 1 out of 2 processors how can I resolve that? Regards, Mahmood _______________________________________________ users mailing list users@lists.open-mpi.org<mailto:users@lists.open-mpi.org> https://rfd.newmexicoconsortium.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org<mailto:users@lists.open-mpi.org> https://rfd.newmexicoconsortium.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org<mailto:users@lists.open-mpi.org> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users