Hi Ralph,

> Just committed a potential fix to the trunk - please let me know
> if it worked for you

Now I get the hostnames but also a segmentation fault.

tyr fd1026 101 which mpiexec
/usr/local/openmpi-1.9_64_cc/bin/mpiexec
tyr fd1026 102 mpiexec -np 3 --host tyr,sunpc1,linpc1 hostname
tyr.informatik.hs-fulda.de
linpc1
sunpc1
[tyr:22835] *** Process received signal ***
[tyr:22835] Signal: Segmentation Fault (11)
[tyr:22835] Signal code: Address not mapped (1)
[tyr:22835] Failing at address: ffffffff7bf16de0
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:opal_backtrace_print+0x1c
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:0x183960
/lib/sparcv9/libc.so.1:0xd8b98
/lib/sparcv9/libc.so.1:0xcc70c
/lib/sparcv9/libc.so.1:0xcc918
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:0x1ce0e8 
[ Signal 2125151224 (?)]
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:0x1ccde4
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:opal_libevent2021_event_del+0x88
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:opal_libevent2021_event_base_free+0x154
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:0x1bb9e8
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:mca_base_framework_close+0x1a0
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-pal.so.0.0.0:opal_finalize+0xcc
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/lib64/libopen-rte.so.0.0.0:orte_finalize+0x168
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/bin/orterun:orterun+0x23e0
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/bin/orterun:main+0x24
/export2/prog/SunOS_sparc/openmpi-1.9_64_cc/bin/orterun:_start+0x12c
[tyr:22835] *** End of error message ***
Segmentation fault
tyr fd1026 103 ompi_info | grep "revision:"
  Open MPI repo revision: r31769
  Open RTE repo revision: r31769
      OPAL repo revision: r31769
tyr fd1026 104 



I get the following output in "dbx".

tyr fd1026 104 /opt/solstudio12.3/bin/sparcv9/dbx 
/usr/local/openmpi-1.9_64_cc/bin/mpiexec 
For information about new features see `help changes'
To remove this message, put `dbxenv suppress_startup_message 7.9' in your .dbxrc
Reading mpiexec
Reading ld.so.1
Reading libopen-rte.so.0.0.0
Reading libopen-pal.so.0.0.0
Reading libsendfile.so.1
Reading libpicl.so.1
Reading libkstat.so.1
Reading liblgrp.so.1
Reading libsocket.so.1
Reading libnsl.so.1
Reading librt.so.1
Reading libm.so.2
Reading libthread.so.1
Reading libc.so.1
Reading libdoor.so.1
Reading libaio.so.1
Reading libmd.so.1
(dbx) run -np 3 --host tyr,sunpc1,linpc1 hostname
Running: mpiexec -np 3 --host tyr,sunpc1,linpc1 hostname 
(process id 23328)
Reading libc_psr.so.1
Reading mca_shmem_mmap.so
Reading libmp.so.2
Reading libscf.so.1
Reading libuutil.so.1
Reading libgen.so.1
Reading mca_shmem_posix.so
Reading mca_shmem_sysv.so
Reading mca_sec_basic.so
Reading mca_ess_env.so
Reading mca_ess_hnp.so
Reading mca_ess_singleton.so
Reading mca_ess_tool.so
Reading mca_pstat_test.so
Reading mca_state_app.so
Reading mca_state_hnp.so
Reading mca_state_novm.so
Reading mca_state_orted.so
Reading mca_state_staged_hnp.so
Reading mca_state_staged_orted.so
Reading mca_state_tool.so
Reading mca_errmgr_default_app.so
Reading mca_errmgr_default_hnp.so
Reading mca_errmgr_default_orted.so
Reading mca_errmgr_default_tool.so
Reading mca_plm_isolated.so
Reading mca_plm_rsh.so
Reading mca_oob_tcp.so
Reading mca_rml_oob.so
Reading mca_routed_binomial.so
Reading mca_routed_debruijn.so
Reading mca_routed_direct.so
Reading mca_routed_radix.so
Reading mca_dstore_hash.so
Reading mca_grpcomm_bad.so
Reading mca_ras_simulator.so
Reading mca_rmaps_lama.so
Reading mca_rmaps_mindist.so
Reading mca_rmaps_ppr.so
Reading mca_rmaps_rank_file.so
Reading mca_rmaps_resilient.so
Reading mca_rmaps_round_robin.so
Reading mca_rmaps_seq.so
Reading mca_rmaps_staged.so
Reading mca_odls_default.so
Reading mca_rtc_hwloc.so
Reading mca_iof_hnp.so
Reading mca_iof_mr_hnp.so
Reading mca_iof_mr_orted.so
Reading mca_iof_orted.so
Reading mca_iof_tool.so
Reading mca_filem_raw.so
Reading mca_dfs_app.so
Reading mca_dfs_orted.so
Reading mca_dfs_test.so
tyr.informatik.hs-fulda.de
linpc1
sunpc1
t@1 (l@1) signal SEGV (no mapping at the fault address) in event_queue_remove 
at 0xffffffff7e9ce0e8
0xffffffff7e9ce0e8: event_queue_remove+0x01a8:  stx      %l0, [%l3 + 24]
Current function is opal_event_base_close
   62       opal_event_base_free (opal_event_base);

(dbx) check -all
dbx: warning: check -all will be turned on in the next run of the process
access checking - OFF
memuse checking - OFF

(dbx) run -np 3 --host tyr,sunpc1,linpc1 hostname
Running: mpiexec -np 3 --host tyr,sunpc1,linpc1 hostname 
(process id 23337)
Reading rtcapihook.so
Reading libdl.so.1
Reading rtcaudit.so
Reading libmapmalloc.so.1
Reading rtcboot.so
Reading librtc.so
Reading libmd_psr.so.1
RTC: Enabling Error Checking...
RTC: Using UltraSparc trap mechanism
RTC: See `help rtc showmap' and `help rtc limitations' for details.
RTC: Running program...
Write to unallocated (wua) on thread 1:
Attempting to write 1 byte at address 0xffffffff79f04000
t@1 (l@1) stopped in _readdir at 0xffffffff56574da0
0xffffffff56574da0: _readdir+0x0064:    call     
_PROCEDURE_LINKAGE_TABLE_+0x2380 [PLT] ! 0xffffffff56742a80
Current function is find_dyn_components
  393                       if (0 != lt_dlforeachfile(dir, save_filename, 
NULL)) {
(dbx) 



Do you need anything else?


KInd regards

Siegmar




 On May 14, 2014, at 11:44 AM, Siegmar Gross 
<siegmar.gr...@informatik.hs-fulda.de> wrote:
> 
> > Hi Ralph,
> > 
> >> Hmmm...well, that's an interesting naming scheme :-)
> >> 
> >> Try adding "-mca oob_base_verbose 10 --report-uri -" on your cmd line
> >> and let's see what it thinks is happening
> > 
> > 
> > tyr fd1026 105 mpiexec -np 3 --host tyr,sunpc1,linpc1 --mca 
> > oob_base_verbose 10 --report-uri - hostname
> > [tyr.informatik.hs-fulda.de:06877] mca: base: components_register: 
> > registering oob components
> > [tyr.informatik.hs-fulda.de:06877] mca: base: components_register: found 
> > loaded component tcp
> > [tyr.informatik.hs-fulda.de:06877] mca: base: components_register: 
> > component tcp register function successful
> > [tyr.informatik.hs-fulda.de:06877] mca: base: components_open: opening oob 
> > components
> > [tyr.informatik.hs-fulda.de:06877] mca: base: components_open: found loaded 
> > component tcp
> > [tyr.informatik.hs-fulda.de:06877] mca: base: components_open: component 
> > tcp open function successful
> > [tyr.informatik.hs-fulda.de:06877] mca:oob:select: checking available 
> > component tcp
> > [tyr.informatik.hs-fulda.de:06877] mca:oob:select: Querying component [tcp]
> > [tyr.informatik.hs-fulda.de:06877] oob:tcp: component_available called
> > [tyr.informatik.hs-fulda.de:06877] WORKING INTERFACE 1 KERNEL INDEX 1 
> > FAMILY: V4
> > [tyr.informatik.hs-fulda.de:06877] WORKING INTERFACE 2 KERNEL INDEX 2 
> > FAMILY: V4
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] oob:tcp:init creating 
> > module for V4 address on interface bge0
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] creating OOB-TCP module 
> > for interface bge0
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] oob:tcp:init adding 
> > 193.174.24.39 to our list of V4 connections
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] TCP STARTUP
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] attempting to bind to IPv4 
> > port 0
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] assigned IPv4 port 55567
> > [tyr.informatik.hs-fulda.de:06877] mca:oob:select: Adding component to end
> > [tyr.informatik.hs-fulda.de:06877] mca:oob:select: Found 1 active transports
> > 3170566144.0;tcp://193.174.24.39:55567
> > [sunpc1:07690] mca: base: components_register: registering oob components
> > [sunpc1:07690] mca: base: components_register: found loaded component tcp
> > [sunpc1:07690] mca: base: components_register: component tcp register 
> > function successful
> > [sunpc1:07690] mca: base: components_open: opening oob components
> > [sunpc1:07690] mca: base: components_open: found loaded component tcp
> > [sunpc1:07690] mca: base: components_open: component tcp open function 
> > successful
> > [sunpc1:07690] mca:oob:select: checking available component tcp
> > [sunpc1:07690] mca:oob:select: Querying component [tcp]
> > [sunpc1:07690] oob:tcp: component_available called
> > [sunpc1:07690] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
> > [sunpc1:07690] WORKING INTERFACE 2 KERNEL INDEX 2 FAMILY: V4
> > [sunpc1:07690] [[48379,0],1] oob:tcp:init creating module for V4 address on 
> > interface nge0
> > [sunpc1:07690] [[48379,0],1] creating OOB-TCP module for interface nge0
> > [sunpc1:07690] [[48379,0],1] oob:tcp:init adding 193.174.26.210 to our list 
> > of V4 connections
> > [sunpc1:07690] [[48379,0],1] TCP STARTUP
> > [sunpc1:07690] [[48379,0],1] attempting to bind to IPv4 port 0
> > [sunpc1:07690] [[48379,0],1] assigned IPv4 port 39616
> > [sunpc1:07690] mca:oob:select: Adding component to end
> > [sunpc1:07690] mca:oob:select: Found 1 active transports
> > [sunpc1:07690] [[48379,0],1]: set_addr to uri 
> > 3170566144.0;tcp://193.174.24.39:55567
> > [sunpc1:07690] [[48379,0],1]:set_addr checking if peer [[48379,0],0] is 
> > reachable via component tcp
> > [sunpc1:07690] [[48379,0],1] oob:tcp: working peer [[48379,0],0] address 
> > tcp://193.174.24.39:55567
> > [sunpc1:07690] [[48379,0],1] UNFOUND KERNEL INDEX -13 FOR ADDRESS 
> > 193.174.24.39
> > [sunpc1:07690] [[48379,0],1] PEER [[48379,0],0] MAY BE REACHABLE BY ROUTING 
> > - ASSIGNING MODULE AT KINDEX 2 
INTERFACE nge0
> > [sunpc1:07690] [[48379,0],1] PASSING ADDR 193.174.24.39 TO INTERFACE nge0 
> > AT KERNEL INDEX 2
> > [sunpc1:07690] [[48379,0],1]:tcp set addr for peer [[48379,0],0]
> > [sunpc1:07690] [[48379,0],1]: peer [[48379,0],0] is reachable via component 
> > tcp
> > [sunpc1:07690] [[48379,0],1] OOB_SEND: 
> > ../../../../../openmpi-1.8.2a1r31742/orte/mca/rml/oob/rml_oob_send.c:199
> > [sunpc1:07690] [[48379,0],1]:tcp:processing set_peer cmd for interface nge0
> > [sunpc1:07690] [[48379,0],1] oob:base:send to target [[48379,0],0]
> > [sunpc1:07690] [[48379,0],1] oob:tcp:send_nb to peer [[48379,0],0]:10
> > [sunpc1:07690] [[48379,0],1] tcp:send_nb to peer [[48379,0],0]
> > [sunpc1:07690] 
> > [[48379,0],1]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:508]
> >  post send to 
[[48379,0],0]
> > [sunpc1:07690] 
> > [[48379,0],1]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:442]
> >  processing 
send to peer 
> > [[48379,0],0]:10
> > [sunpc1:07690] 
> > [[48379,0],1]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:476]
> >  queue pending 
to [[48379,0],0]
> > [sunpc1:07690] [[48379,0],1] tcp:send_nb: initiating connection to 
> > [[48379,0],0]
> > [sunpc1:07690] 
> > [[48379,0],1]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:490]
> >  connect to 
[[48379,0],0]
> > [sunpc1:07690] [[48379,0],1] orte_tcp_peer_try_connect: attempting to 
> > connect to proc [[48379,0],0] via interface 
nge0
> > [sunpc1:07690] [[48379,0],1] oob:tcp:peer creating socket to [[48379,0],0]
> > [sunpc1:07690] [[48379,0],1] orte_tcp_peer_try_connect: attempting to 
> > connect to proc [[48379,0],0] via interface 
nge0 on socket 10
> > [sunpc1:07690] [[48379,0],1] orte_tcp_peer_try_connect: attempting to 
> > connect to proc [[48379,0],0] on 
193.174.24.39:55567 - 0 retries
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] mca_oob_tcp_listen_thread: 
> > new connection: (15, 0) 
193.174.26.210:39617
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] connection_handler: 
> > working connection (15, 11) 
193.174.26.210:39617
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] CONNECTION REQUEST ON 
> > UNKNOWN INTERFACE
> > [sunpc1:07690] [[48379,0],1] waiting for connect completion to 
> > [[48379,0],0] - activating send event
> > [sunpc1:07690] [[48379,0],1] tcp:send_handler called to send to peer 
> > [[48379,0],0]
> > [sunpc1:07690] [[48379,0],1] tcp:send_handler CONNECTING
> > [sunpc1:07690] [[48379,0],1]:tcp:complete_connect called for peer 
> > [[48379,0],0] on socket 10
> > [sunpc1:07690] [[48379,0],1] tcp_peer_complete_connect: sending ack to 
> > [[48379,0],0]
> > [sunpc1:07690] [[48379,0],1] SEND CONNECT ACK
> > [sunpc1:07690] [[48379,0],1] send blocking of 48 bytes to socket 10
> > [sunpc1:07690] [[48379,0],1] connect-ack sent to socket 10
> > [sunpc1:07690] [[48379,0],1] tcp_peer_complete_connect: setting read event 
> > on connection to [[48379,0],0]
> > [linpc1:21511] mca: base: components_register: registering oob components
> > [linpc1:21511] mca: base: components_register: found loaded component tcp
> > [linpc1:21511] mca: base: components_register: component tcp register 
> > function successful
> > [linpc1:21511] mca: base: components_open: opening oob components
> > [linpc1:21511] mca: base: components_open: found loaded component tcp
> > [linpc1:21511] mca: base: components_open: component tcp open function 
> > successful
> > [linpc1:21511] mca:oob:select: checking available component tcp
> > [linpc1:21511] mca:oob:select: Querying component [tcp]
> > [linpc1:21511] oob:tcp: component_available called
> > 
> > [linpc1:21511] WORKING INTERFACE 1 KERNEL INDEX 1 FAMILY: V4
> > [linpc1:21511] WORKING INTERFACE 2 KERNEL INDEX 2 FAMILY: V4
> > [linpc1:21511] [[48379,0],2] oob:tcp:init creating module for V4 address on 
> > interface eth0
> > [linpc1:21511] [[48379,0],2] creating OOB-TCP module for interface eth0
> > [linpc1:21511] [[48379,0],2] oob:tcp:init adding 193.174.26.208 to our list 
> > of V4 connections
> > [linpc1:21511] [[48379,0],2] TCP STARTUP
> > [linpc1:21511] [[48379,0],2] attempting to bind to IPv4 port 0
> > [linpc1:21511] [[48379,0],2] assigned IPv4 port 39724
> > [linpc1:21511] mca:oob:select: Adding component to end
> > [linpc1:21511] mca:oob:select: Found 1 active transports
> > [linpc1:21511] [[48379,0],2]: set_addr to uri 
> > 3170566144.0;tcp://193.174.24.39:55567
> > [linpc1:21511] [[48379,0],2]:set_addr checking if peer [[48379,0],0] is 
> > reachable via component tcp
> > [linpc1:21511] [[48379,0],2] oob:tcp: working peer [[48379,0],0] address 
> > tcp://193.174.24.39:55567
> > [linpc1:21511] [[48379,0],2] UNFOUND KERNEL INDEX -13 FOR ADDRESS 
> > 193.174.24.39
> > [linpc1:21511] [[48379,0],2] PEER [[48379,0],0] MAY BE REACHABLE BY ROUTING 
> > - ASSIGNING MODULE AT KINDEX 2 
INTERFACE eth0
> > [linpc1:21511] [[48379,0],2] PASSING ADDR 193.174.24.39 TO INTERFACE eth0 
> > AT KERNEL INDEX 2
> > [linpc1:21511] [[48379,0],2]:tcp set addr for peer [[48379,0],0]
> > [linpc1:21511] [[48379,0],2]: peer [[48379,0],0] is reachable via component 
> > tcp
> > [linpc1:21511] [[48379,0],2] OOB_SEND: 
> > ../../../../../openmpi-1.8.2a1r31742/orte/mca/rml/oob/rml_oob_send.c:199
> > [linpc1:21511] [[48379,0],2]:tcp:processing set_peer cmd for interface eth0
> > [linpc1:21511] [[48379,0],2] oob:base:send to target [[48379,0],0]
> > [linpc1:21511] [[48379,0],2] oob:tcp:send_nb to peer [[48379,0],0]:10
> > [linpc1:21511] [[48379,0],2] tcp:send_nb to peer [[48379,0],0]
> > [linpc1:21511] 
> > [[48379,0],2]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:508]
> >  post send to 
[[48379,0],0]
> > [linpc1:21511] 
> > [[48379,0],2]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:442]
> >  processing 
send to peer 
> > [[48379,0],0]:10
> > [linpc1:21511] 
> > [[48379,0],2]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:476]
> >  queue pending 
to [[48379,0],0]
> > [linpc1:21511] [[48379,0],2] tcp:send_nb: initiating connection to 
> > [[48379,0],0]
> > [linpc1:21511] 
> > [[48379,0],2]:[../../../../../openmpi-1.8.2a1r31742/orte/mca/oob/tcp/oob_tcp.c:490]
> >  connect to 
[[48379,0],0]
> > [linpc1:21511] [[48379,0],2] orte_tcp_peer_try_connect: attempting to 
> > connect to proc [[48379,0],0] via interface 
eth0
> > [linpc1:21511] [[48379,0],2] oob:tcp:peer creating socket to [[48379,0],0]
> > [linpc1:21511] [[48379,0],2] orte_tcp_peer_try_connect: attempting to 
> > connect to proc [[48379,0],0] via interface 
eth0 on socket 9
> > [linpc1:21511] [[48379,0],2] orte_tcp_peer_try_connect: attempting to 
> > connect to proc [[48379,0],0] on 
193.174.24.39:55567 - 0 retries
> > [linpc1:21511] [[48379,0],2] waiting for connect completion to 
> > [[48379,0],0] - activating send event
> > [linpc1:21511] [[48379,0],2] tcp:send_handler called to send to peer 
> > [[48379,0],0]
> > [linpc1:21511] [[48379,0],2] tcp:send_handler CONNECTING
> > [linpc1:21511] [[48379,0],2]:tcp:complete_connect called for peer 
> > [[48379,0],0] on socket 9
> > [linpc1:21511] [[48379,0],2] tcp_peer_complete_connect: sending ack to 
> > [[48379,0],0]
> > [linpc1:21511] [[48379,0],2] SEND CONNECT ACK
> > [linpc1:21511] [[48379,0],2] send blocking of 48 bytes to socket 9
> > [linpc1:21511] [[48379,0],2] connect-ack sent to socket 9
> > [linpc1:21511] [[48379,0],2] tcp_peer_complete_connect: setting read event 
> > on connection to [[48379,0],0]
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] mca_oob_tcp_listen_thread: 
> > new connection: (16, 11) 
193.174.26.208:53741
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] connection_handler: 
> > working connection (16, 11) 
193.174.26.208:53741
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] CONNECTION REQUEST ON 
> > UNKNOWN INTERFACE
> > ^CKilled by signal 2.
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] OOB_SEND: 
../../../../../openmpi-1.8.2a1r31742/orte/mca/rml/oob/rml_oob_send.c:199
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] OOB_SEND: 
../../../../../openmpi-1.8.2a1r31742/orte/mca/rml/oob/rml_oob_send.c:199
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] oob:base:send to target 
> > [[48379,0],1]
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] oob:base:send unknown peer 
> > [[48379,0],1]
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] is NOT reachable by TCP
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] oob:base:send to target 
> > [[48379,0],2]
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] oob:base:send unknown peer 
> > [[48379,0],2]
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] is NOT reachable by TCP
> > Killed by signal 2.
> > [tyr.informatik.hs-fulda.de:06877] [[48379,0],0] TCP SHUTDOWN
> > [tyr.informatik.hs-fulda.de:06877] mca: base: close: component tcp closed
> > [tyr.informatik.hs-fulda.de:06877] mca: base: close: unloading component tcp
> > tyr fd1026 106 
> > 
> > 
> > Thank you very much for your help in advance. Do you need anything else?
> > 
> > 
> > Kind regards
> > 
> > Siegmar
> > 
> > 
> > 
> >> On May 14, 2014, at 9:06 AM, Siegmar Gross 
> >> <siegmar.gr...@informatik.hs-fulda.de> wrote:
> >> 
> >>> Hi Ralph,
> >>> 
> >>>> What are the interfaces on these machines?
> >>> 
> >>> tyr fd1026 111 ifconfig -a
> >>> lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu 
> >>> 8232 index 1
> >>>       inet 127.0.0.1 netmask ff000000 
> >>> bge0: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 2
> >>>       inet 193.174.24.39 netmask ffffffe0 broadcast 193.174.24.63
> >>> tyr fd1026 112 
> >>> 
> >>> 
> >>> tyr fd1026 112 ssh sunpc1 ifconfig -a
> >>> lo0: flags=2001000849<UP,LOOPBACK,RUNNING,MULTICAST,IPv4,VIRTUAL> mtu 
> >>> 8232 index 1
> >>>       inet 127.0.0.1 netmask ff000000 
> >>> nge0: flags=1000843<UP,BROADCAST,RUNNING,MULTICAST,IPv4> mtu 1500 index 2
> >>>       inet 193.174.26.210 netmask ffffffc0 broadcast 193.174.26.255
> >>> tyr fd1026 113 
> >>> 
> >>> 
> >>> tyr fd1026 113 ssh linpc1 /sbin/ifconfig -a
> >>> eth0      Link encap:Ethernet  HWaddr 00:14:4F:23:FD:A8  
> >>>         inet addr:193.174.26.208  Bcast:193.174.26.255  
> >>> Mask:255.255.255.192
> >>>         UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
> >>>         RX packets:18052524 errors:127 dropped:0 overruns:0 frame:127
> >>>         TX packets:15917888 errors:0 dropped:0 overruns:0 carrier:0
> >>>         collisions:0 txqueuelen:1000 
> >>>         RX bytes:4158294157 (3965.6 Mb)  TX bytes:12060556809 (11501.8 Mb)
> >>>         Interrupt:23 Base address:0x4000 
> >>> 
> >>> eth1      Link encap:Ethernet  HWaddr 00:14:4F:23:FD:A9  
> >>>         BROADCAST MULTICAST  MTU:1500  Metric:1
> >>>         RX packets:0 errors:0 dropped:0 overruns:0 frame:0
> >>>         TX packets:0 errors:0 dropped:0 overruns:0 carrier:0
> >>>         collisions:0 txqueuelen:1000 
> >>>         RX bytes:0 (0.0 b)  TX bytes:0 (0.0 b)
> >>>         Interrupt:45 Base address:0xa000 
> >>> 
> >>> lo        Link encap:Local Loopback  
> >>>         inet addr:127.0.0.1  Mask:255.0.0.0
> >>>         UP LOOPBACK RUNNING  MTU:16436  Metric:1
> >>>         RX packets:1083 errors:0 dropped:0 overruns:0 frame:0
> >>>         TX packets:1083 errors:0 dropped:0 overruns:0 carrier:0
> >>>         collisions:0 txqueuelen:0 
> >>>         RX bytes:329323 (321.6 Kb)  TX bytes:329323 (321.6 Kb)
> >>> 
> >>> tyr fd1026 114 
> >>> 
> >>> 
> >>> Do you need something else?
> >>> 
> >>> 
> >>> Kind regards
> >>> 
> >>> Siegmar
> >>> 
> >>> 
> >>> 
> >>> 
> >>>> On May 14, 2014, at 7:45 AM, Siegmar Gross 
> >>>> <siegmar.gr...@informatik.hs-fulda.de> wrote:
> >>>> 
> >>>>> Hi,
> >>>>> 
> >>>>> I just installed openmpi-1.8.2a1r31742 on my machines (Solaris 10
> >>>>> Sparc, Solaris 10 x86_64, and openSUSE Linux 12.1 x86_64) with
> >>>>> Sun C5.12 and still have the following problem.
> >>>>> 
> >>>>> tyr fd1026 102 which mpiexec
> >>>>> /usr/local/openmpi-1.8.2_64_cc/bin/mpiexec
> >>>>> tyr fd1026 103 mpiexec -np 3 --host tyr,sunpc1,linpc1 hostname
> >>>>> [tyr.informatik.hs-fulda.de:12827] [[37949,0],0] CONNECTION
> >>>>> REQUEST ON UNKNOWN INTERFACE
> >>>>> [tyr.informatik.hs-fulda.de:12827] [[37949,0],0] CONNECTION
> >>>>> REQUEST ON UNKNOWN INTERFACE
> >>>>> ^CKilled by signal 2.
> >>>>> Killed by signal 2.
> >>>>> tyr fd1026 104 
> >>>>> 
> >>>>> 
> >>>>> The command works fine with openmpi-1.6.6rc1.
> >>>>> 
> >>>>> tyr fd1026 102 which mpiexec
> >>>>> /usr/local/openmpi-1.6.6_64_cc/bin/mpiexec
> >>>>> tyr fd1026 103 mpiexec -np 3 --host tyr,sunpc1,linpc1 hostname
> >>>>> tyr.informatik.hs-fulda.de
> >>>>> linpc1
> >>>>> sunpc1
> >>>>> tyr fd1026 104 
> >>>>> 
> >>>>> 
> >>>>> I have reported the problem before and I would be grateful, if
> >>>>> somebody could solve it. Please let me know if I can provide any
> >>>>> other information.
> >>>>> 
> >>>>> 
> >>>>> Kind regards
> >>>>> 
> >>>>> Siegmar
> >>>>> 
> >>>>> _______________________________________________
> >>>>> users mailing list
> >>>>> us...@open-mpi.org
> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>>> 
> >>>> 
> >>> 
> >>> _______________________________________________
> >>> users mailing list
> >>> us...@open-mpi.org
> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >> 
> > 
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 

Reply via email to