Thanks, Bennet. I made the modification to the Torque submission file and got “20 n388”, which confirms (like you said) that for my cluster runs I am requesting 20 cores on a single node.
Best regards, Andy On Feb 1, 2017, at 5:15 PM, Bennet Fauber <ben...@umich.edu> wrote: You may want to run this by Penguin support, too. I believe that Penguin on Demand use Torque, in which case the nodes=1:ppn=20 is requesting 20 cores on a single node. If this is Torque, then you should get a host list, with counts by inserting uniq -c $PBS_NODEFILE after the last #PBS directive. That should print the host name and the number 20. MPI should resort to whatever it uses when it is on the same node. -- bennet On Wed, Feb 1, 2017 at 6:04 PM, r...@open-mpi.org <r...@open-mpi.org> wrote: > Simple test: replace your executable with “hostname”. If you see multiple > hosts come out on your cluster, then you know why the performance is > different. > > On Feb 1, 2017, at 2:46 PM, Andy Witzig <cap1...@icloud.com> wrote: > > Honestly, I’m not exactly sure what scheme is being used. I am using the > default template from Penguin Computing for job submission. It looks like: > > #PBS -S /bin/bash > #PBS -q T30 > #PBS -l walltime=24:00:00,nodes=1:ppn=20 > #PBS -j oe > #PBS -N test > #PBS -r n > > mpirun $EXECUTABLE $INPUT_FILE > > I’m not configuring OpenMPI anywhere else. It is possible the Penguin > Computing folks have pre-configured my MPI environment. I’ll see what I can > find. > > Best regards, > Andy > > On Feb 1, 2017, at 4:32 PM, Douglas L Reeder <d...@centurylink.net> wrote: > > Andy, > > What allocation scheme are you using on the cluster. For some codes we see > noticeable differences using fillup vs round robin, not 4x though. Fillup is > more shared memory use while round robin uses more infinniband. > > Doug > > On Feb 1, 2017, at 3:25 PM, Andy Witzig <cap1...@icloud.com> wrote: > > Hi Tom, > > The cluster uses an Infiniband interconnect. On the cluster I’m requesting: > #PBS -l walltime=24:00:00,nodes=1:ppn=20. So technically, the run on the > cluster should be SMP on the node, since there are 20 cores/node. On the > workstation I’m just using the command: mpirun -np 20 …. I haven’t finished > setting Torque/PBS up yet. > > Best regards, > Andy > > On Feb 1, 2017, at 4:10 PM, Elken, Tom <tom.el...@intel.com> wrote: > > For this case: " a cluster system with 2.6GHz Intel Haswell with 20 cores / > node and 128GB RAM/node. " > > are you running 5 ranks per node on 4 nodes? > What interconnect are you using for the cluster? > > -Tom > > -----Original Message----- > From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Andrew > Witzig > Sent: Wednesday, February 01, 2017 1:37 PM > To: Open MPI Users > Subject: Re: [OMPI users] Performance Issues on SMP Workstation > > By the way, the workstation has a total of 36 cores / 72 threads, so using > mpirun > -np 20 is possible (and should be equivalent) on both platforms. > > Thanks, > cap79 > > On Feb 1, 2017, at 2:52 PM, Andy Witzig <cap1...@icloud.com> wrote: > > Hi all, > > I’m testing my application on a SMP workstation (dual Intel Xeon E5-2697 V4 > > 2.3 GHz Intel Broadwell (boost 2.8-3.1GHz) processors 128GB RAM) and am > seeing a 4x performance drop compared to a cluster system with 2.6GHz Intel > Haswell with 20 cores / node and 128GB RAM/node. Both applications have > been compiled using OpenMPI 1.6.4. I have tried running: > > > mpirun -np 20 $EXECUTABLE $INPUT_FILE > mpirun -np 20 --mca btl self,sm $EXECUTABLE $INPUT_FILE > > and others, but cannot achieve the same performance on the workstation as is > > seen on the cluster. The workstation outperforms on other non-MPI but > multi- > threaded applications, so I don’t think it’s a hardware issue. > > > Any help you can provide would be appreciated. > > Thanks, > cap79 > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users > > > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://rfd.newmexicoconsortium.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users _______________________________________________ users mailing list users@lists.open-mpi.org https://rfd.newmexicoconsortium.org/mailman/listinfo/users