The same TCP connection if I run on the other cluster UINTA, I get the following error log
mca/pls/tm/pls_tm_module.c at line 572 [uinta-0039:14508] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../../orte/\ mca/errmgr/hnp/errmgr_hnp.c at line 90 [uinta-0039:14508] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../orte/mca\ /pls/base/pls_base_orted_cmds.c at line 188 [uinta-0039:14508] [0,0,0] ORTE_ERROR_LOG: Timeout in file ../../../../../orte/\ mca/pls/tm/pls_tm_module.c at line 603 -------------------------------------------------------------------------- mpirun was unable to cleanly terminate the daemons for this job. Returned value\ Timeout instead of ORTE_SUCCESS. -------------------------------------------------------------------------- [uinta-0039:14510] OOB: Connection to HNP lost [uinta-0038:15165] OOB: Connection to HNP lost On Sat, Nov 20, 2010 at 5:35 PM, Tushar Andriyas <thugnomic...@gmail.com>wrote: > I tried out the TCP connection and here is what the error file came out as. > > > > [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file > ../../../../orte/mca\ > /pls/base/pls_base_orted_cmds.c at line 275 > [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file > ../../../../../orte/\ > mca/pls/tm/pls_tm_module.c at line 572 > [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file > ../../../../../orte/\ > mca/errmgr/hnp/errmgr_hnp.c at line 90 > [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file > ../../../../orte/mca\ > /pls/base/pls_base_orted_cmds.c at line 188 > [wasatch-29:05042] [0,0,0] ORTE_ERROR_LOG: Timeout in file > ../../../../../orte/\ > mca/pls/tm/pls_tm_module.c at line 603 > -------------------------------------------------------------------------- > mpirun was unable to cleanly terminate the daemons for this job. Returned > value\ > Timeout instead of ORTE_SUCCESS. > -------------------------------------------------------------------------- > [wasatch-29:05044] OOB: Connection to HNP lost > > > > > On Sat, Nov 20, 2010 at 2:03 PM, Addepalli, Srirangam V < > srirangam.v.addepa...@ttu.edu> wrote: > >> mpirun --mca btl self,sm,tcp --mca btl_base_verbose 30 -np 8 >> /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log >> >> to run using tcp interface in job submission script. >> >> Rangam >> _______________________________________ >> From: users-boun...@open-mpi.org [users-boun...@open-mpi.org] On Behalf >> Of Tushar Andriyas [thugnomic...@gmail.com] >> Sent: Saturday, November 20, 2010 1:36 PM >> To: Open MPI Users >> Subject: Re: [OMPI users] Unable to find the following executable >> >> Ya sure, here is the list >> >> >> Open MPI: 1.2.7 >> Open MPI SVN revision: r19401 >> Open RTE: 1.2.7 >> Open RTE SVN revision: r19401 >> OPAL: 1.2.7 >> OPAL SVN revision: r19401 >> Prefix: /opt/libraries/openmpi/openmpi-1.2.7-pgi >> Configured architecture: x86_64-unknown-linux-gnu >> Configured by: A00017402 >> Configured on: Thu Sep 18 15:00:05 MDT 2008 >> Configure host: volvox.hpc.usu.edu<http://volvox.hpc.usu.edu> >> Built by: A00017402 >> Built on: Thu Sep 18 15:20:06 MDT 2008 >> Built host: volvox.hpc.usu.edu<http://volvox.hpc.usu.edu> >> C bindings: yes >> C++ bindings: yes >> Fortran77 bindings: yes (all) >> Fortran90 bindings: yes >> Fortran90 bindings size: large >> C compiler: pgcc >> C compiler absolute: /opt/apps/pgi/linux86-64/7.2/bin/pgcc >> C++ compiler: pgCC >> C++ compiler absolute: /opt/apps/pgi/linux86-64/7.2/bin/pgCC >> Fortran77 compiler: pgf77 >> Fortran77 compiler abs: /opt/apps/pgi/linux86-64/7.2/bin/pgf77 >> Fortran90 compiler: pgf90 >> Fortran90 compiler abs: /opt/apps/pgi/linux86-64/7.2/bin/pgf90 >> C profiling: yes >> C++ profiling: yes >> Fortran77 profiling: yes >> Fortran90 profiling: yes >> C++ exceptions: no >> Thread support: posix (mpi: no, progress: no) >> Internal debug support: no >> MPI parameter check: runtime >> Memory profiling support: no >> Memory debugging support: no >> libltdl support: yes >> Heterogeneous support: yes >> mpirun default --prefix: no >> MCA backtrace: execinfo (MCA v1.0, API v1.0, Component v1.2.7) >> MCA memory: ptmalloc2 (MCA v1.0, API v1.0, Component v1.2.7) >> MCA paffinity: linux (MCA v1.0, API v1.0, Component v1.2.7) >> MCA maffinity: first_use (MCA v1.0, API v1.0, Component v1.2.7) >> MCA maffinity: libnuma (MCA v1.0, API v1.0, Component v1.2.7) >> MCA timer: linux (MCA v1.0, API v1.0, Component v1.2.7) >> MCA installdirs: env (MCA v1.0, API v1.0, Component v1.2.7) >> MCA installdirs: config (MCA v1.0, API v1.0, Component v1.2.7) >> MCA allocator: basic (MCA v1.0, API v1.0, Component v1.0) >> MCA allocator: bucket (MCA v1.0, API v1.0, Component v1.0) >> MCA coll: basic (MCA v1.0, API v1.0, Component v1.2.7) >> MCA coll: self (MCA v1.0, API v1.0, Component v1.2.7) >> MCA coll: sm (MCA v1.0, API v1.0, Component v1.2.7) >> MCA coll: tuned (MCA v1.0, API v1.0, Component v1.2.7) >> MCA io: romio (MCA v1.0, API v1.0, Component v1.2.7) >> MCA mpool: rdma (MCA v1.0, API v1.0, Component v1.2.7) >> MCA mpool: sm (MCA v1.0, API v1.0, Component v1.2.7) >> MCA pml: cm (MCA v1.0, API v1.0, Component v1.2.7) >> MCA pml: ob1 (MCA v1.0, API v1.0, Component v1.2.7) >> MCA bml: r2 (MCA v1.0, API v1.0, Component v1.2.7) >> MCA rcache: vma (MCA v1.0, API v1.0, Component v1.2.7) >> MCA btl: gm (MCA v1.0, API v1.0.1, Component v1.2.7) >> MCA btl: self (MCA v1.0, API v1.0.1, Component v1.2.7) >> MCA btl: sm (MCA v1.0, API v1.0.1, Component v1.2.7) >> MCA btl: tcp (MCA v1.0, API v1.0.1, Component v1.0) >> MCA topo: unity (MCA v1.0, API v1.0, Component v1.2.7) >> MCA osc: pt2pt (MCA v1.0, API v1.0, Component v1.2.7) >> MCA errmgr: hnp (MCA v1.0, API v1.3, Component v1.2.7) >> MCA errmgr: orted (MCA v1.0, API v1.3, Component v1.2.7) >> MCA errmgr: proxy (MCA v1.0, API v1.3, Component v1.2.7) >> MCA gpr: null (MCA v1.0, API v1.0, Component v1.2.7) >> MCA gpr: proxy (MCA v1.0, API v1.0, Component v1.2.7) >> MCA gpr: replica (MCA v1.0, API v1.0, Component v1.2.7) >> MCA iof: proxy (MCA v1.0, API v1.0, Component v1.2.7) >> MCA iof: svc (MCA v1.0, API v1.0, Component v1.2.7) >> MCA ns: proxy (MCA v1.0, API v2.0, Component v1.2.7) >> MCA ns: replica (MCA v1.0, API v2.0, Component v1.2.7) >> MCA oob: tcp (MCA v1.0, API v1.0, Component v1.0) >> MCA ras: dash_host (MCA v1.0, API v1.3, Component v1.2.7) >> MCA ras: gridengine (MCA v1.0, API v1.3, Component v1.2.7) >> MCA ras: localhost (MCA v1.0, API v1.3, Component v1.2.7) >> MCA ras: slurm (MCA v1.0, API v1.3, Component v1.2.7) >> MCA ras: tm (MCA v1.0, API v1.3, Component v1.2.7) >> MCA rds: hostfile (MCA v1.0, API v1.3, Component v1.2.7) >> MCA rds: proxy (MCA v1.0, API v1.3, Component v1.2.7) >> MCA rds: resfile (MCA v1.0, API v1.3, Component v1.2.7) >> MCA rmaps: round_robin (MCA v1.0, API v1.3, Component >> v1.2.7) >> MCA rmgr: proxy (MCA v1.0, API v2.0, Component v1.2.7) >> MCA rmgr: urm (MCA v1.0, API v2.0, Component v1.2.7) >> MCA rml: oob (MCA v1.0, API v1.0, Component v1.2.7) >> MCA pls: gridengine (MCA v1.0, API v1.3, Component v1.2.7) >> MCA pls: proxy (MCA v1.0, API v1.3, Component v1.2.7) >> MCA pls: rsh (MCA v1.0, API v1.3, Component v1.2.7) >> MCA pls: slurm (MCA v1.0, API v1.3, Component v1.2.7) >> MCA pls: tm (MCA v1.0, API v1.3, Component v1.2.7) >> MCA sds: env (MCA v1.0, API v1.0, Component v1.2.7) >> MCA sds: pipe (MCA v1.0, API v1.0, Component v1.2.7) >> MCA sds: seed (MCA v1.0, API v1.0, Component v1.2.7) >> MCA sds: singleton (MCA v1.0, API v1.0, Component v1.2.7) >> MCA sds: slurm (MCA v1.0, API v1.0, Component v1.2.7) >> >> How do you invoke tcp? I no for sure that the launcher on the clusters is >> torque. >> >> Tushar >> >> On Sat, Nov 20, 2010 at 11:28 AM, Addepalli, Srirangam V < >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu>> >> wrote: >> Hello Tushar, >> Can you send me the output of ompi_info. >> Have you tried using just tcp instead of IB to narrow down. >> Rangam >> >> #!/bin/sh >> #PBS -V >> #PBS -q wasatch >> #PBS -N SWMF >> #PBS -l nodes=1:ppn=8 >> # change to the run directory >> #cd $SWMF_v2.3/run >> cat `echo ${PBS_NODEFILE}` > list_of_nodes >> >> mpirun --mca btl self,sm,tcp --mca btl_base_verbose 30 -np 8 >> /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log >> >> >> ________________________________________ >> From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org> [ >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>] On Behalf >> Of Tushar Andriyas [thugnomic...@gmail.com<mailto:thugnomic...@gmail.com >> >] >> Sent: Saturday, November 20, 2010 12:11 PM >> To: Open MPI Users >> Subject: Re: [OMPI users] Unable to find the following executable >> >> Rangam, >> >> It does not want to run at all. Attached is the log file from the batch >> file run u sent. >> >> On Sat, Nov 20, 2010 at 10:32 AM, Addepalli, Srirangam V < >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> ><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu>>> wrote: >> Hello Tushar, >> MPIRUN is not able to spawn processes on the node allocated. This should >> help >> >> #!/bin/sh >> #PBS -V >> #PBS -q wasatch >> #PBS -N SWMF >> #PBS -l nodes=2:ppn=8 >> # change to the run directory >> #cd $SWMF_v2.3/run >> cat `echo ${PBS_NODEFILE}` > list_of_nodes >> mpirun -np 8 /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log >> >> >> Rangam >> >> >> ________________________________________ >> From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>> [ >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>] On Behalf >> Of Tushar Andriyas [thugnomic...@gmail.com<mailto:thugnomic...@gmail.com >> ><mailto:thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>] >> Sent: Saturday, November 20, 2010 10:48 AM >> To: Open MPI Users >> Subject: Re: [OMPI users] Unable to find the following executable >> >> Hi Rangam, >> >> I ran the batch file that you gave and have attached the error file. Also, >> since the WASATCH cluster is kind of small, people usually run on UINTA. So, >> if possible could you look at the uinta error files? >> Tushar >> >> On Fri, Nov 19, 2010 at 12:31 PM, Addepalli, Srirangam V < >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> ><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu>><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu>>>> >> wrote: >> Hello Tushar, >> After looking at the log files you attached it appears that there are >> multiple issues. >> >> [0,1,11]: Myrinet/GM on host wasatch-55 was unable to find any NICs. >> Another transport will be used instead, although this may result in >> lower performance. >> >> Usually they occur if there is a mismatch in mpirun version and mca blt >> selection. I suggest the following order to check if the job actually works >> on a single node >> >> #!/bin/sh >> #PBS -V >> #PBS -q wasatch >> #PBS -N SWMF >> #PBS -l nodes=2:ppn=8 >> # change to the run directory >> #cd $SWMF_v2.3/run >> cat `echo ${PBS_NODEFILE}` > list_of_nodes >> mpirun -np 8 -machinefile list_of_nodes >> /home/A00945081/SWMF_v2.3/run/SWMF.exe > run.log >> >> >> Rangam >> >> >> ________________________________________ >> From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>> >> [users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>] On >> Behalf Of Tushar Andriyas [thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>>] >> Sent: Friday, November 19, 2010 1:11 PM >> To: Open MPI Users >> Subject: Re: [OMPI users] Unable to find the following executable >> >> Hey Rangam, >> >> I tried out the batch script and the error file comes out empty and the >> output file has /home/A00945081/SWM_v2.3/run/SWMF.exe (WHEN RUN ON A SINGLE >> MACHINE) and the same with multiple machines in the run. So, does that mean >> that the exe is auto mounted ? What should I do next? >> >> Tushar >> >> On Fri, Nov 19, 2010 at 10:05 AM, Addepalli, Srirangam V < >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> ><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu>><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> >>><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu>><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> ><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu>>>>> wrote: >> Hello Tushar, >> >> Try the following script. >> >> #!/bin/sh >> #PBS -V >> #PBS -q wasatch >> #PBS -N SWMF >> #PBS -l nodes=1:ppn=8 >> # change to the run directory >> #cd $SWMF_v2.3/run >> cat `echo ${PBS_NODEFILE}` > list_of_nodes >> >> >> >> >> The objective is to check if your user directories are auto mounted on >> compute nodes and are available during run time. >> >> If the job returns information about SWMF.exe then it can be safely >> assumed that user directories are being auto mounted. >> >> Rangam >> >> >> >> ________________________________________ >> From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>> >> [users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: >> users-bounce! >> s...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>>] On >> Behalf Of Tushar Andriyas [thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>>>] >> Sent: Friday, November 19, 2010 8:35 AM >> To: Open MPI Users >> Subject: Re: [OMPI users] Unable to find the following executable >> >> It just gives back the info on folders in my home directory. Dont get me >> wrong but i m kinda new in this. So, could u type out d full command which i >> need to give? >> >> Tushar >> >> On Thu, Nov 18, 2010 at 8:35 AM, Ralph Castain <r...@open-mpi.org<mailto: >> r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org >> <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>>><mailto:r...@open-mpi.org >> <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org >> <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>>>>> wrote: >> You can qsub a simple "ls" on that path - that will tell you if the path >> is valid on all machines in that allocation. >> >> What typically happens is that home directories aren't remotely mounted, >> or are mounted on a different location. >> >> >> On Thu, Nov 18, 2010 at 8:31 AM, Tushar Andriyas <thugnomic...@gmail.com >> <mailto:thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomics28@! >> gmail.com<mailto:thugnomic...@gmail.com>>>>>> wrote: >> no its not in the same directory as SWMF. I guess the path is the same >> since all the machines in a cluster are configured d same way. How do I know >> if this is not the case? >> >> >> On Thu, Nov 18, 2010 at 8:25 AM, Ralph Castain <r...@open-mpi.org<mailto: >> r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org >> <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>>><mailto:r...@open-mpi.org >> <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>><mailto:r...@open-mpi.org >> <mailto:r...@open-mpi.org><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org >> >><mailto:r...@open-mpi.org<mailto:r...@open-mpi.org><mailto: >> r...@open-mpi.org<mailto:r...@open-mpi.org>>>>>> wrote: >> Is you "hello world" test program in the same directory as SWMF? Is it >> possible that the path you are specifying is not available on all of the >> remote machines? That's the most common problem we see. >> >> >> On Thu, Nov 18, 2010 at 7:59 AM, Tushar Andriyas <thugnomic...@gmail.com >> <mailto:thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com>><mailto:thugnomic...@gmail.com<mailto: >> thugnomic...@gmail.com><mailto:thugnomics28@! >> gmail.com<mailto:thugnomic...@gmail.com>>>>>> wrote: >> Hi there, >> >> Thanks for the expedite reply. The thing is that although the mpirun is >> setup correctly (since a simple hello world works), when I run the main >> SWMF.exe executable, the cluster machines somehow fail to find the >> executable (SWMF.exe). >> >> So, I have attached the sample error file from one of the runs >> (SWMF.e143438) and also the MAKEFILES so that you could better gauge the >> problem. The makefiles have Linux as the OS and pgf90 as compiler with >> mpif90 as the linker. I am using openmpi-1.2.7-pgi. Job is submitted using a >> batch file (job.bats) and the scheduler is Torque (version I am not sure but >> I can see three on the machines viz 2.0.0, 2.2.1, 2.5.2). >> >> I have also attached an error file from one of the clusters (WASATCH viz >> SWMF.e143439) and UINTA (SWMF.e143440) with the whole path of the exe as >> Srirangam mentioned as follows (in the batch file). >> >> mpirun --prefix /opt/libraries/openmpi/openmpi-1.2.7-pgi >> /home/A00945081/SWMF_v2.3/run/SWMF.exe > runlog_`date +%y%m%d%H%M` >> >> I have tried both mpirun and mpiexec but nothing seems to work. >> >> Tushar >> >> >> On Wed, Nov 17, 2010 at 8:12 PM, Addepalli, Srirangam V < >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> ><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu>><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> >>><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu>><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> ><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu>>>><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> >><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu>>><mailto:srirangam.v.addepalli@! >> ttu.edu<mailto:srirangam.v.addepa...@ttu.edu><mailto: >> srirangam.v.addepa...@ttu.edu<mailto:srirangam.v.addepa...@ttu.edu >> >><mailto:srirangam.v.addepa...@ttu.edu<mailto: >> srirangam.v.addepa...@ttu.edu><mailto:srirangam.v.addepa...@ttu.edu >> <mailto:srirangam.v.addepa...@ttu.edu>>>>>> wrote: >> Hello Tushar, >> Have you tried supplying the full path of the executable just to check ? >> Rangam >> ________________________________________ >> From: users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >>>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >>><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> ><mailto:users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org >> >><mailto:users-! >> boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>>> [ >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org><mailto: >> users-boun...@open-mpi.org<mailto:users-boun...@open-mpi.org>>><mailto: >> users-bounce<mailto:! >> users-bounce>! >> >> s...@open-mpi.org<mailto:s...@open-mpi.org><mailto:users-boun...@open-mpi.org >> <mailto:users-boun...@open-mpi.org>><mailto:users-boun...@open-mpi.org >> <mailto:users-boun...@open-mpi.org><mailto:users-boun...@open-mpi.org >> <mailto:users-boun...@open-mpi.org>>>>>] On Behalf Of Tushar Andriyas [ >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailt! >> o:thugnomic...@gmail.com <o%3athugnomic...@gmail.com>>>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com><mailto: >> thugnomic...@gmail.com<mailto:thugnomic...@gmail.com>>>>>] >> Sent: Wednesday, November 17, 2010 8:49 PM >> To: us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> >> Subject: [OMPI users] Unable to find the following executable >> >> Hi there, >> >> I am new to using mpi commands and was stuck in problem with running a >> code. When I submit my job through a batch file, the job exits with the >> message that the executable could not be found on the machines. I have tried >> a lot of options such as PBS -V and so on on but the problem persists. If >> someone is interested, I can send the full info on the cluster, the compiler >> and openmpi settings and other stuff. BTW the launcher is torque (which you >> might have guessed). The code does not have a forum so I am in a deep mire. >> >> Thanks, >> Tushar >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> ><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org>>>>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org><mailto: >> us...@open-mpi.org<mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org>>>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>><mailto:us...@open-mpi.org<mailto: >> us...@open-mpi.org><mailto:us...@open-mpi.org<mailto:us...@open-mpi.org >> >>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org><mailto:us...@open-mpi.org >> <mailto:us...@open-mpi.org>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org<mailto:us...@open-mpi.org> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > >