Thank you, i m reading up on de tools u suggested.
I am facing another problem, my cluster is working fine with 2 hosts (1
master + 1 compute node) but when i tried 2 add another node (1 master + 2
compute node) its not working. it works fine when i give de command
mpirun -host <hostname> /work/Pi

but when i try to run
mpirun  /work/Pi it gives following error:

root@192.168.45.65's password: root@192.168.67.241's password:

Permission denied, please try again. <The password i provide is correct>

root@192.168.45.65's password:

Permission denied, please try again.

root@192.168.45.65's password:

Permission denied (publickey,gssapi-with-mic,password).



Permission denied, please try again.

root@192.168.67.241's password: [ccomp1.cluster:03503] [0,0,0]
ORTE_ERROR_LOG: Timeout in file base/pls_base_orted_cmds.c at line 275

[ccomp1.cluster:03503] [0,0,0] ORTE_ERROR_LOG: Timeout in file
pls_rsh_module.c at line 1166

[ccomp1.cluster:03503] [0,0,0] ORTE_ERROR_LOG: Timeout in file errmgr_hnp.c
at line 90

[ccomp1.cluster:03503] ERROR: A daemon on node 192.168.45.65 failed to start
as expected.

[ccomp1.cluster:03503] ERROR: There may be more information available from

[ccomp1.cluster:03503] ERROR: the remote shell (see above).

[ccomp1.cluster:03503] ERROR: The daemon exited unexpectedly with status
255.

[ccomp1.cluster:03503] [0,0,0] ORTE_ERROR_LOG: Timeout in file
base/pls_base_orted_cmds.c at line 188

[ccomp1.cluster:03503] [0,0,0] ORTE_ERROR_LOG: Timeout in file
pls_rsh_module.c at line 1198


What is the problem here?

--------------------------------------------------------------------------

mpirun was unable to cleanly terminate the daemons for this job. Returned
value Timeout instead of ORTE_SUCCESS

On Tue, Apr 14, 2009 at 7:15 PM, Eugene Loh <eugene....@sun.com> wrote:

> Ankush Kaul wrote:
>
>  Finally, after mentioning the hostfiles the cluster is working fine. We
>> downloaded few benchmarking softwares but i would like to know if there is
>> any GUI based benchmarking software so that its easier to demonstrate the
>> working of our cluster while displaying our cluster.
>>
>
> I'm confused what you're looking for here, but thought I'd venture a
> suggestion.
>
> There are GUI-based performance analysis and tracing tools.  E.g., run a
> program, [[semi-]automatically] collect performance data, run a GUI-based
> analysis tool on the data, visualize what happened on your cluster.  Would
> this suit your purposes?
>
> If so, there are a variety of tools out there you could try.  Some are
> platform-specific or cost money.  Some are widely/freely available.
>  Examples of these tools include Intel Trace Analyzer, Jumpshot, Vampir,
> TAU, etc.  I do know that Sun Studio (Performance Analyzer) is available via
> free download on x86 and SPARC and Linux and Solaris and works with OMPI.
>  Possibly the same with Jumpshot.  VampirTrace instrumentation is already in
> OMPI, but then you need to figure out the analysis-tool part.  (I think the
> Vampir GUI tool requires a license, but I'm not sure.  Maybe you can convert
> to TAU, which is probably available for free download.)
>
> Anyhow, I don't even know if that sort of thing fits your requirements.
>  Just an idea.
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to