I got no response for this question. Is Open-MX no longer supported in Open MPI? Or is there someplace else I should submit this information? I also attached my ompi_info and omx_info output
-- Doug > I built open-mpi 1.6.1 using the open-mx libraries. > This worked previously and now I get the following > error. Here is my system: > > kernel: 2.6.32-279.5.1.el6.x86_64 > open-mx: 1.5.2 > > BTW, open-mx worked previously with open-mpi and the current > version works with mpich2 > > > $ mpiexec -np 8 -machinefile machines cpi > Process 0 on limulus > FatalError: Failed to lookup peer by addr, driver replied Bad file > descriptor > cpi: ../omx_misc.c:89: omx__ioctl_errno_to_return_checked: Assertion `0' > failed. > [limulus:04448] *** Process received signal *** > [limulus:04448] Signal: Aborted (6) > [limulus:04448] Signal code: (-6) > [limulus:04448] [ 0] /lib64/libpthread.so.0() [0x3324e0f500] > [limulus:04448] [ 1] /lib64/libc.so.6(gsignal+0x35) [0x33246328a5] > [limulus:04448] [ 2] /lib64/libc.so.6(abort+0x175) [0x3324634085] > [limulus:04448] [ 3] /lib64/libc.so.6() [0x332462ba1e] > [limulus:04448] [ 4] /lib64/libc.so.6(__assert_perror_fail+0) > [0x332462bae0] > [limulus:04448] [ 5] > /usr/open-mx/lib/libopen-mx.so.0(omx__ioctl_errno_to_return_checked+0x197) > [0x7fb587418b37] > [limulus:04448] [ 6] > /usr/open-mx/lib/libopen-mx.so.0(omx__peer_addr_to_index+0x55) > [0x7fb58741a5d5] > [limulus:04448] [ 7] /usr/open-mx/lib/libopen-mx.so.0(+0xdc7a) > [0x7fb587419c7a] > [limulus:04448] [ 8] /usr/open-mx/lib/libopen-mx.so.0(omx_connect+0x8c) > [0x7fb58741a27c] > [limulus:04448] [ 9] /usr/open-mx/lib/libopen-mx.so.0(mx_connect+0x15) > [0x7fb587425865] > [limulus:04448] [10] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_btl_mx_proc_connect+0x5e) > [0x7fb5876fe40e] > [limulus:04448] [11] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_btl_mx_send+0x2d4) > [0x7fb5876fbd94] > [limulus:04448] [12] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_pml_ob1_send_request_start_prepare+0xcb) > [0x7fb58777d6fb] > [limulus:04448] [13] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_pml_ob1_isend+0x4cb) > [0x7fb58777509b] > [limulus:04448] [14] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(ompi_coll_tuned_bcast_intra_generic+0x37b) > [0x7fb58770b55b] > [limulus:04448] [15] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(ompi_coll_tuned_bcast_intra_binomial+0xd8) > [0x7fb58770b8b8] > [limulus:04448] [16] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(ompi_coll_tuned_bcast_intra_dec_fixed+0xcc) > [0x7fb587702d8c] > [limulus:04448] [17] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(mca_coll_sync_bcast+0x78) > [0x7fb587712e88] > [limulus:04448] [18] > /opt/mpi/openmpi-gnu4/lib64/libmpi.so.1(MPI_Bcast+0x130) [0x7fb5876ce1b0] > [limulus:04448] [19] cpi(main+0x10b) [0x400cc4] > [limulus:04448] [20] /lib64/libc.so.6(__libc_start_main+0xfd) > [0x332461ecdd] > [limulus:04448] [21] cpi() [0x400ac9] > [limulus:04448] *** End of error message *** > Process 2 on limulus > Process 4 on limulus > Process 6 on limulus > Process 1 on n0 > Process 7 on n0 > Process 3 on n0 > Process 5 on n0 > -------------------------------------------------------------------------- > mpiexec noticed that process rank 0 with PID 4448 on node limulus exited > on signal 6 (Aborted). > -------------------------------------------------------------------------- > [root@limulus ~]# ompi_info Package: Open MPI deadline@limulus Distribution Open MPI: 1.6.1 Open MPI SVN revision: r27106 Open MPI release date: Aug 22, 2012 Open RTE: 1.6.1 Open RTE SVN revision: r27106 Open RTE release date: Aug 22, 2012 OPAL: 1.6.1 OPAL SVN revision: r27106 OPAL release date: Aug 22, 2012 MPI API: 2.1 Ident string: 1.6.1 Prefix: /opt/mpi/openmpi-gnu4-1.6.1 Configured architecture: x86_64-unknown-linux-gnu Configure host: limulus Configured by: deadline Configured on: Wed Sep 5 15:03:09 EDT 2012 Configure host: limulus Built by: deadline Built on: Wed Sep 5 15:21:19 EDT 2012 Built host: limulus C bindings: yes C++ bindings: yes Fortran77 bindings: yes (all) Fortran90 bindings: yes Fortran90 bindings size: small C compiler: gcc C compiler absolute: /usr/bin/gcc C compiler family name: GNU C compiler version: 4.4.6 C++ compiler: g++ C++ compiler absolute: /usr/bin/g++ Fortran77 compiler: gfortran Fortran77 compiler abs: /usr/bin/gfortran Fortran90 compiler: /usr/bin/gfortran Fortran90 compiler abs: C profiling: yes C++ profiling: yes Fortran77 profiling: yes Fortran90 profiling: yes C++ exceptions: no Thread support: posix (MPI_THREAD_MULTIPLE: no, progress: no) Sparse Groups: no Internal debug support: no MPI interface warnings: no MPI parameter check: runtime Memory profiling support: no Memory debugging support: no libltdl support: yes Heterogeneous support: no mpirun default --prefix: no MPI I/O support: yes MPI_WTIME support: gettimeofday Symbol vis. support: yes Host topology support: yes MPI extensions: affinity example FT Checkpoint support: no (checkpoint thread: no) VampirTrace support: yes MPI_MAX_PROCESSOR_NAME: 256 MPI_MAX_ERROR_STRING: 256 MPI_MAX_OBJECT_NAME: 64 MPI_MAX_INFO_KEY: 36 MPI_MAX_INFO_VAL: 256 MPI_MAX_PORT_NAME: 1024 MPI_MAX_DATAREP_STRING: 128 MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.6.1) MCA memory: linux (MCA v2.0, API v2.0, Component v1.6.1) MCA paffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.1) MCA carto: auto_detect (MCA v2.0, API v2.0, Component v1.6.1) MCA carto: file (MCA v2.0, API v2.0, Component v1.6.1) MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.6.1) MCA shmem: posix (MCA v2.0, API v2.0, Component v1.6.1) MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.6.1) MCA maffinity: first_use (MCA v2.0, API v2.0, Component v1.6.1) MCA maffinity: hwloc (MCA v2.0, API v2.0, Component v1.6.1) MCA timer: linux (MCA v2.0, API v2.0, Component v1.6.1) MCA installdirs: env (MCA v2.0, API v2.0, Component v1.6.1) MCA installdirs: config (MCA v2.0, API v2.0, Component v1.6.1) MCA sysinfo: linux (MCA v2.0, API v2.0, Component v1.6.1) MCA hwloc: hwloc132 (MCA v2.0, API v2.0, Component v1.6.1) MCA dpm: orte (MCA v2.0, API v2.0, Component v1.6.1) MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.6.1) MCA allocator: basic (MCA v2.0, API v2.0, Component v1.6.1) MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: basic (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: inter (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: self (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: sm (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: sync (MCA v2.0, API v2.0, Component v1.6.1) MCA coll: tuned (MCA v2.0, API v2.0, Component v1.6.1) MCA io: romio (MCA v2.0, API v2.0, Component v1.6.1) MCA mpool: fake (MCA v2.0, API v2.0, Component v1.6.1) MCA mpool: rdma (MCA v2.0, API v2.0, Component v1.6.1) MCA mpool: sm (MCA v2.0, API v2.0, Component v1.6.1) MCA pml: bfo (MCA v2.0, API v2.0, Component v1.6.1) MCA pml: cm (MCA v2.0, API v2.0, Component v1.6.1) MCA pml: csum (MCA v2.0, API v2.0, Component v1.6.1) MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.6.1) MCA pml: v (MCA v2.0, API v2.0, Component v1.6.1) MCA bml: r2 (MCA v2.0, API v2.0, Component v1.6.1) MCA rcache: vma (MCA v2.0, API v2.0, Component v1.6.1) MCA btl: self (MCA v2.0, API v2.0, Component v1.6.1) MCA btl: mx (MCA v2.0, API v2.0, Component v1.6.1) MCA btl: sm (MCA v2.0, API v2.0, Component v1.6.1) MCA btl: tcp (MCA v2.0, API v2.0, Component v1.6.1) MCA mtl: mx (MCA v2.0, API v2.0, Component v1.6.1) MCA topo: unity (MCA v2.0, API v2.0, Component v1.6.1) MCA osc: pt2pt (MCA v2.0, API v2.0, Component v1.6.1) MCA osc: rdma (MCA v2.0, API v2.0, Component v1.6.1) MCA iof: hnp (MCA v2.0, API v2.0, Component v1.6.1) MCA iof: orted (MCA v2.0, API v2.0, Component v1.6.1) MCA iof: tool (MCA v2.0, API v2.0, Component v1.6.1) MCA oob: tcp (MCA v2.0, API v2.0, Component v1.6.1) MCA odls: default (MCA v2.0, API v2.0, Component v1.6.1) MCA ras: cm (MCA v2.0, API v2.0, Component v1.6.1) MCA ras: gridengine (MCA v2.0, API v2.0, Component v1.6.1) MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.6.1) MCA ras: slurm (MCA v2.0, API v2.0, Component v1.6.1) MCA rmaps: load_balance (MCA v2.0, API v2.0, Component v1.6.1) MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.6.1) MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.6.1) MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.6.1) MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.6.1) MCA rmaps: topo (MCA v2.0, API v2.0, Component v1.6.1) MCA rml: oob (MCA v2.0, API v2.0, Component v1.6.1) MCA routed: binomial (MCA v2.0, API v2.0, Component v1.6.1) MCA routed: cm (MCA v2.0, API v2.0, Component v1.6.1) MCA routed: direct (MCA v2.0, API v2.0, Component v1.6.1) MCA routed: linear (MCA v2.0, API v2.0, Component v1.6.1) MCA routed: radix (MCA v2.0, API v2.0, Component v1.6.1) MCA routed: slave (MCA v2.0, API v2.0, Component v1.6.1) MCA plm: rsh (MCA v2.0, API v2.0, Component v1.6.1) MCA plm: slurm (MCA v2.0, API v2.0, Component v1.6.1) MCA filem: rsh (MCA v2.0, API v2.0, Component v1.6.1) MCA errmgr: default (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: env (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: hnp (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: singleton (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: slave (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: slurm (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: slurmd (MCA v2.0, API v2.0, Component v1.6.1) MCA ess: tool (MCA v2.0, API v2.0, Component v1.6.1) MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.6.1) MCA grpcomm: basic (MCA v2.0, API v2.0, Component v1.6.1) MCA grpcomm: hier (MCA v2.0, API v2.0, Component v1.6.1) MCA notifier: command (MCA v2.0, API v1.0, Component v1.6.1) MCA notifier: syslog (MCA v2.0, API v1.0, Component v1.6.1) [root@limulus ~]# /usr/open-mx/bin/omx_info Open-MX version 1.5.2 build: deadline@limulus:/raid1/home/deadline/rpms-sl6/BUILD/open-mx-1.5.2 Mon Sep 10 08:44:16 EDT 2012 Found 1 boards (32 max) supporting 32 endpoints each: limulus:0 (board #0 name eth0 addr e0:69:95:35:d7:71) managed by driver 'e1000e' Peer table is ready, mapper is 00:00:00:00:00:00 ================================================ 0) e0:69:95:35:d7:71 limulus:0 1) 00:1c:c0:9b:66:d0 n1:0 2) 00:1a:4d:4a:bf:85 n0:0 3) 00:1a:4d:4a:bf:83 n2:0 -- Doug -- Mailscanner: Clean