On Sun, Feb 1, 2009 at 10:37 PM, Reuti <re...@staff.uni-marburg.de> wrote: > Am 01.02.2009 um 16:00 schrieb Sangamesh B: > >> On Sat, Jan 31, 2009 at 6:27 PM, Reuti <re...@staff.uni-marburg.de> wrote: >>> >>> Am 31.01.2009 um 08:49 schrieb Sangamesh B: >>> >>>> On Fri, Jan 30, 2009 at 10:20 PM, Reuti <re...@staff.uni-marburg.de> >>>> wrote: >>>>> >>>>> Am 30.01.2009 um 15:02 schrieb Sangamesh B: >>>>> >>>>>> Dear Open MPI, >>>>>> >>>>>> Do you have a solution for the following problem of Open MPI (1.3) >>>>>> when run through Grid Engine. >>>>>> >>>>>> I changed global execd params with H_MEMORYLOCKED=infinity and >>>>>> restarted the sgeexecd in all nodes. >>>>>> >>>>>> But still the problem persists: >>>>>> >>>>>> $cat err.77.CPMD-OMPI >>>>>> ssh_exchange_identification: Connection closed by remote host >>>>> >>>>> I think this might already be the reason why it's not working. A >>>>> mpihello >>>>> program is running fine through SGE? >>>>> >>>> No. >>>> >>>> Any Open MPI parallel job thru SGE runs only if its running on a >>>> single node (i.e. 8processes on 8 cores of a single node). If number >>>> of processes is more than 8, then SGE will schedule it on 2 nodes - >>>> the job will fail with the above error. >>>> >>>> Now I did a loose integration of Open MPI 1.3 with SGE. The job runs, >>>> but all 16 processes run on a single node. >>> >>> What are the entries in `qconf -sconf`for: >>> >>> rsh_command >>> rsh_daemon >>> >> $ qconf -sconf >> global: >> execd_spool_dir /opt/gridengine/default/spool >> ... >> ..... >> qrsh_command /usr/bin/ssh >> rsh_command /usr/bin/ssh >> rlogin_command /usr/bin/ssh >> rsh_daemon /usr/sbin/sshd >> qrsh_daemon /usr/sbin/sshd >> reprioritize 0 > > Do you must use ssh? Often in a private cluster the rsh based one is ok, or > with SGE 6.2 the built-in mechanism of SGE. Otherwise please follow this: > > http://gridengine.sunsource.net/howto/qrsh_qlogin_ssh.html > > >> I think its better to check once with Open MPI 1.2.8 >> >>> What is your mpirun command in the jobscript - you are getting there the >>> mpirun from Open MPI? According to the output below, it's not a loose >>> integration, but you prepare alraedy a machinefile, which is superfluous >>> for >>> Open MPI. >>> >> No. I've not prepared the machinefile for Open MPI. >> For Tight integartion job: >> >> /opt/mpi/openmpi/1.3/intel/bin/mpirun -np $NSLOTS >> $CPMDBIN/cpmd311-ompi-mkl.x wf1.in $PP_LIBRARY > >> wf1.out_OMPI$NSLOTS.$JOB_ID >> >> For loose integration job: >> >> /opt/mpi/openmpi/1.3/intel/bin/mpirun -np $NSLOTS -hostfile >> $TMPDIR/machines $CPMDBIN/cpmd311-ompi-mkl.x wf1.in $PP_LIBRARY > >> wf1.out_OMPI_$JOB_ID.$NSLOTS > > a) you compiled Open MPI with "--with-sge"? > Yes. But ompi_info shows only one component of sge
$ /opt/mpi/openmpi/1.3/intel/bin/ompi_info | grep gridengine MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.3) > b) when the $SGE_ROOT variable is set, Open MPI will use a Tight Integration > automatically. > In SGE job submit script, I set SGE_ROOT= <nothing> And run a loose integration job. It failed to run with following error: $ cat err.87.Hello-OMPI [node-0-18.local:08252] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found in file ess_hnp_module.c at line 126 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): orte_plm_base_select failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- [node-0-18.local:08252] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found in file runtime/orte_init.c at line 132 -------------------------------------------------------------------------- It looks like orte_init failed for some reason; your parallel process is likely to abort. There are many reasons that a parallel process can fail during orte_init; some of which are due to configuration or environment problems. This failure appears to be an internal failure; here's some additional information (which may only be relevant to an Open MPI developer): orte_ess_set_name failed --> Returned value Not found (-13) instead of ORTE_SUCCESS -------------------------------------------------------------------------- [node-0-18.local:08252] [[INVALID],INVALID] ORTE_ERROR_LOG: Not found in file orterun.c at line 454 $ cat out.87.Hello-OMPI /opt/gridengine/default/spool/node-0-18/active_jobs/87.1/pe_hostfile ibc18 ibc18 ibc18 ibc18 ibc18 ibc18 ibc18 ibc18 ibc17 ibc17 ibc17 ibc17 ibc17 ibc17 ibc17 ibc17 > c) The machine file you presented looks like being for MPICH(1), the syntax > for Open MPI in the machine is different: > > ibc17 slots=8 > ibc12 slots=8 > I tested a helloworld program with Open MPI with machinefile of style MPICH(1). It works. So in a loose integration job, Open MPI may not be able to find $TMPDIR/machines file Or it might be running in a Tight integration style. > So you would have to adjust the format of the generated file and reset > SGE_ROOT inside your jobscript, to force Open MPI to do a loose integration > only. > > -- Reuti > > >> I think I should check with Open MPI 1.2.8. That may work.. >> >> Thanks, >> Sangamesh >>>> >>>> $ cat out.83.Hello-OMPI >>>> /opt/gridengine/default/spool/node-0-17/active_jobs/83.1/pe_hostfile >>>> ibc17 >>>> ibc17 >>>> ibc17 >>>> ibc17 >>>> ibc17 >>>> ibc17 >>>> ibc17 >>>> ibc17 >>>> ibc12 >>>> ibc12 >>>> ibc12 >>>> ibc12 >>>> ibc12 >>>> ibc12 >>>> ibc12 >>>> ibc12 >>>> Greetings: 1 of 16 from the node node-0-17.local >>>> Greetings: 10 of 16 from the node node-0-17.local >>>> Greetings: 15 of 16 from the node node-0-17.local >>>> Greetings: 9 of 16 from the node node-0-17.local >>>> Greetings: 14 of 16 from the node node-0-17.local >>>> Greetings: 8 of 16 from the node node-0-17.local >>>> Greetings: 11 of 16 from the node node-0-17.local >>>> Greetings: 12 of 16 from the node node-0-17.local >>>> Greetings: 6 of 16 from the node node-0-17.local >>>> Greetings: 0 of 16 from the node node-0-17.local >>>> Greetings: 5 of 16 from the node node-0-17.local >>>> Greetings: 3 of 16 from the node node-0-17.local >>>> Greetings: 13 of 16 from the node node-0-17.local >>>> Greetings: 4 of 16 from the node node-0-17.local >>>> Greetings: 7 of 16 from the node node-0-17.local >>>> Greetings: 2 of 16 from the node node-0-17.local >>>> >>>> But qhost -u <user name> shows that it is scheduled/running on two >>>> nodes. >>>> >>>> Any body successful in running Open MPI 1.3 tightly integrated with SGE? >>> >>> For a Tight Integration there's a FAQ: >>> >>> http://www.open-mpi.org/faq/?category=running#run-n1ge-or-sge >>> >>> -- Reuti >>> >>>> >>>> Thanks, >>>> Sangamesh >>>> >>>>> -- Reuti >>>>> >>>>> >>>>>> >>>>>> >>>>>> -------------------------------------------------------------------------- >>>>>> A daemon (pid 31947) died unexpectedly with status 129 while >>>>>> attempting >>>>>> to launch so we are aborting. >>>>>> >>>>>> There may be more information reported by the environment (see above). >>>>>> >>>>>> This may be because the daemon was unable to find all the needed >>>>>> shared >>>>>> libraries on the remote node. You may set your LD_LIBRARY_PATH to have >>>>>> the >>>>>> location of the shared libraries on the remote nodes and this will >>>>>> automatically be forwarded to the remote nodes. >>>>>> >>>>>> >>>>>> -------------------------------------------------------------------------- >>>>>> >>>>>> >>>>>> -------------------------------------------------------------------------- >>>>>> mpirun noticed that the job aborted, but has no info as to the process >>>>>> that caused that situation. >>>>>> >>>>>> >>>>>> -------------------------------------------------------------------------- >>>>>> ssh_exchange_identification: Connection closed by remote host >>>>>> >>>>>> >>>>>> -------------------------------------------------------------------------- >>>>>> mpirun was unable to cleanly terminate the daemons on the nodes shown >>>>>> below. Additional manual cleanup may be required - please refer to >>>>>> the "orte-clean" tool for assistance. >>>>>> >>>>>> >>>>>> -------------------------------------------------------------------------- >>>>>> node-0-19.local - daemon did not report back when launched >>>>>> node-0-20.local - daemon did not report back when launched >>>>>> node-0-21.local - daemon did not report back when launched >>>>>> node-0-22.local - daemon did not report back when launched >>>>>> >>>>>> The hostnames for infiniband interfaces are ibc0, ibc1, ibc2 .. ibc23. >>>>>> May be Open MPI is not able to identify hosts as it is using node-0-.. >>>>>> . Is this causing open mpi to fail? >>>>>> >>>>>> Thanks, >>>>>> Sangamesh >>>>>> >>>>>> >>>>>> On Mon, Jan 26, 2009 at 5:09 PM, mihlon <vacl...@fel.cvut.cz> wrote: >>>>>>> >>>>>>> Hi, >>>>>>> >>>>>>>> Hello SGE users, >>>>>>>> >>>>>>>> The cluster is installed with Rocks-4.3, SGE 6.0 & Open MPI 1.3. >>>>>>>> Open MPI is configured with "--with-sge". >>>>>>>> ompi_info shows only one component: >>>>>>>> # /opt/mpi/openmpi/1.3/intel/bin/ompi_info | grep gridengine >>>>>>>> MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.3) >>>>>>>> >>>>>>>> Is this acceptable? >>>>>>> >>>>>>> maybe yes >>>>>>> >>>>>>> see: http://www.open-mpi.org/faq/?category=building#build-rte-sge >>>>>>> >>>>>>> shell$ ompi_info | grep gridengine >>>>>>> MCA ras: gridengine (MCA v1.0, API v1.3, Component v1.3) >>>>>>> MCA pls: gridengine (MCA v1.0, API v1.3, Component v1.3) >>>>>>> >>>>>>> (Specific frameworks and version numbers may vary, depending on your >>>>>>> version of Open MPI.) >>>>>>> >>>>>>>> The Open MPI parallel jobs run successfully through command line, >>>>>>>> but >>>>>>>> fail when run thru SGE(with -pe orte <slots>). >>>>>>>> >>>>>>>> The error is: >>>>>>>> >>>>>>>> $ cat err.26.Helloworld-PRL >>>>>>>> ssh_exchange_identification: Connection closed by remote host >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> A daemon (pid 8462) died unexpectedly with status 129 while >>>>>>>> attempting >>>>>>>> to launch so we are aborting. >>>>>>>> >>>>>>>> There may be more information reported by the environment (see >>>>>>>> above). >>>>>>>> >>>>>>>> This may be because the daemon was unable to find all the needed >>>>>>>> shared >>>>>>>> libraries on the remote node. You may set your LD_LIBRARY_PATH to >>>>>>>> have >>>>>>>> the >>>>>>>> location of the shared libraries on the remote nodes and this will >>>>>>>> automatically be forwarded to the remote nodes. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> mpirun noticed that the job aborted, but has no info as to the >>>>>>>> process >>>>>>>> that caused that situation. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> mpirun: clean termination accomplished >>>>>>>> >>>>>>>> But the same job runs well, if it runs on a single node but with an >>>>>>>> error: >>>>>>>> >>>>>>>> $ cat err.23.Helloworld-PRL >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> WARNING: There was an error initializing an OpenFabrics device. >>>>>>>> >>>>>>>> Local host: node-0-4.local >>>>>>>> Local device: mthca0 >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -------------------------------------------------------------------------- >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> libibverbs: Warning: RLIMIT_MEMLOCK is 32768 bytes. >>>>>>>> This will severely limit memory registrations. >>>>>>>> [node-0-4.local:07869] 7 more processes have sent help message >>>>>>>> help-mpi-btl-openib.txt / error in device init >>>>>>>> [node-0-4.local:07869] Set MCA parameter "orte_base_help_aggregate" >>>>>>>> to >>>>>>>> 0 to see all help / error messages >>>>>>>> >>>>>>>> The following link explains the same problem: >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> http://gridengine.sunsource.net/ds/viewMessage.do?dsForumId=38&dsMessageId=72398 >>>>>>>> >>>>>>>> With this reference, I put 'ulimit -l unlimited' into >>>>>>>> /etc/init.d/sgeexecd in all nodes. Restarted the services. >>>>>>> >>>>>>> Do not set 'ulimit -l unlimited' in /etc/init.d/sgeexecd >>>>>>> but set it in the SGE: >>>>>>> >>>>>>> Run qconf -mconf and set execd_params >>>>>>> >>>>>>> >>>>>>> frontend$> qconf -sconf >>>>>>> ... >>>>>>> execd_params H_MEMORYLOCKED=infinity >>>>>>> ... >>>>>>> >>>>>>> >>>>>>> Then restart all your sgeexecd hosts. >>>>>>> >>>>>>> >>>>>>> Milan >>>>>>> >>>>>>>> But still the problem persists. >>>>>>>> >>>>>>>> What could be the way out for this? >>>>>>>> >>>>>>>> Thanks, >>>>>>>> Sangamesh >>>>>>>> >>>>>>>> ------------------------------------------------------ >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> http://gridengine.sunsource.net/ds/viewMessage.do?dsForumId=38&dsMessageId=99133 >>>>>>>> >>>>>>>> To unsubscribe from this discussion, e-mail: >>>>>>>> [users-unsubscr...@gridengine.sunsource.net]. >>>>>>>> >>>>>>> >>>>>>> ------------------------------------------------------ >>>>>>> >>>>>>> >>>>>>> >>>>>>> http://gridengine.sunsource.net/ds/viewMessage.do?dsForumId=38&dsMessageId=99461 >>>>>>> >>>>>>> To unsubscribe from this discussion, e-mail: >>>>>>> [users-unsubscr...@gridengine.sunsource.net]. >>>>>>> >>>>>> _______________________________________________ >>>>>> users mailing list >>>>>> us...@open-mpi.org >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >