Can you try using send/recv with the datatype in question? It could be a problem with either the one-sided code or the datatype code. Could you also give master a try?
-Nathan On Tue, Apr 14, 2015 at 06:43:31AM +0000, MOHR STEPHAN 239883 wrote: > Hi Howard, > > I tried with 1.8.5rc1, but it doesn't work either. > > The output of ompi_info is attached. > > Thanks, > Stephan > > ---------------------------------------------------------------------- > > From: users [users-boun...@open-mpi.org] on behalf of Howard Pritchard > [hpprit...@gmail.com] > Sent: Monday, April 13, 2015 19:41 > To: Open MPI Users > Subject: Re: [OMPI users] mpi_type_create_struct not working for large > counts > HI Stephan, > For starters, would you mind sending the output you get when you run the > ompi_info command? > If you could, it would be great if you could try running the test against > the latest 1.8.5rc1? > The test appears to work without problem using mpich, at least with 4 > ranks. > Thanks, > Howard > 2015-04-13 10:40 GMT-06:00 MOHR STEPHAN 239883 <stephan.m...@cea.fr>: > > Hi there, > > I've got an issue when using a derived data type created by > mpi_type_create_struct in a one-sided communication. > > The problem can be reproduced using the small standalone program which I > attached. It just creates a type which should be equivalent to n > contiguous elements. This type is then used in a mpi_get. With a value > n=100 it works fine, but with n=1000 it either hangs (version 1.8.1) or > crashes (version 1.6.5). > > Any help is appreciated. > > Best regards, > Stephan > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/04/26695.php > Package: Open MPI stephanm@girofle Distribution > Open MPI: 1.8.5rc1 > Open MPI repo revision: v1.8.4-184-g481d751 > Open MPI release date: Apr 05, 2015 > Open RTE: 1.8.5rc1 > Open RTE repo revision: v1.8.4-184-g481d751 > Open RTE release date: Apr 05, 2015 > OPAL: 1.8.5rc1 > OPAL repo revision: v1.8.4-184-g481d751 > OPAL release date: Apr 05, 2015 > MPI API: 3.0 > Ident string: 1.8.5rc1 > Prefix: /local/stephanm/openmpi-1.8.5rc1_intel > Configured architecture: x86_64-unknown-linux-gnu > Configure host: girofle > Configured by: stephanm > Configured on: Tue Apr 14 07:32:10 CEST 2015 > Configure host: girofle > Built by: stephanm > Built on: Tue Apr 14 08:05:43 CEST 2015 > Built host: girofle > C bindings: yes > C++ bindings: yes > Fort mpif.h: yes (all) > Fort use mpi: yes (full: ignore TKR) > Fort use mpi size: deprecated-ompi-info-value > Fort use mpi_f08: yes > Fort mpi_f08 compliance: The mpi_f08 module is available, but due to > limitations in the ifort compiler, does not support the following: array > subsections, direct passthru (where possible) to underlying Open MPI's C > functionality > Fort mpi_f08 subarrays: no > Java bindings: no > Wrapper compiler rpath: runpath > C compiler: icc > C compiler absolute: > /local/stephanm/composer_xe_2013_sp1.0.080/bin/intel64/icc > C compiler family name: INTEL > C compiler version: 1400.20130728 > C++ compiler: icpc > C++ compiler absolute: > /local/stephanm/composer_xe_2013_sp1.0.080/bin/intel64/icpc > Fort compiler: ifort > Fort compiler abs: > /local/stephanm/composer_xe_2013_sp1.0.080/bin/intel64/ifort > Fort ignore TKR: yes (!DEC$ ATTRIBUTES NO_ARG_CHECK ::) > Fort 08 assumed shape: no > Fort optional args: yes > Fort INTERFACE: yes > Fort ISO_FORTRAN_ENV: yes > Fort STORAGE_SIZE: yes > Fort BIND(C) (all): yes > Fort ISO_C_BINDING: yes > Fort SUBROUTINE BIND(C): yes > Fort TYPE,BIND(C): yes > Fort T,BIND(C,name="a"): yes > Fort PRIVATE: yes > Fort PROTECTED: yes > Fort ABSTRACT: yes > Fort ASYNCHRONOUS: yes > Fort PROCEDURE: yes > Fort C_FUNLOC: yes > Fort f08 using wrappers: yes > Fort MPI_SIZEOF: yes > C profiling: yes > C++ profiling: yes > Fort mpif.h profiling: yes > Fort use mpi profiling: yes > Fort use mpi_f08 prof: yes > C++ exceptions: no > Thread support: posix (MPI_THREAD_MULTIPLE: no, OPAL support: yes, > OMPI progress: no, ORTE progress: yes, Event lib: yes) > Sparse Groups: no > Internal debug support: no > MPI interface warnings: yes > MPI parameter check: runtime > Memory profiling support: no > Memory debugging support: no > libltdl support: yes > Heterogeneous support: no > mpirun default --prefix: no > MPI I/O support: yes > MPI_WTIME support: gettimeofday > Symbol vis. support: yes > Host topology support: yes > MPI extensions: > FT Checkpoint support: no (checkpoint thread: no) > C/R Enabled Debugging: no > VampirTrace support: yes > MPI_MAX_PROCESSOR_NAME: 256 > MPI_MAX_ERROR_STRING: 256 > MPI_MAX_OBJECT_NAME: 64 > MPI_MAX_INFO_KEY: 36 > MPI_MAX_INFO_VAL: 256 > MPI_MAX_PORT_NAME: 1024 > MPI_MAX_DATAREP_STRING: 128 > MCA backtrace: execinfo (MCA v2.0, API v2.0, Component v1.8.5) > MCA compress: bzip (MCA v2.0, API v2.0, Component v1.8.5) > MCA compress: gzip (MCA v2.0, API v2.0, Component v1.8.5) > MCA crs: none (MCA v2.0, API v2.0, Component v1.8.5) > MCA db: hash (MCA v2.0, API v1.0, Component v1.8.5) > MCA db: print (MCA v2.0, API v1.0, Component v1.8.5) > MCA event: libevent2021 (MCA v2.0, API v2.0, Component v1.8.5) > MCA hwloc: hwloc191 (MCA v2.0, API v2.0, Component v1.8.5) > MCA if: posix_ipv4 (MCA v2.0, API v2.0, Component v1.8.5) > MCA if: linux_ipv6 (MCA v2.0, API v2.0, Component v1.8.5) > MCA installdirs: env (MCA v2.0, API v2.0, Component v1.8.5) > MCA installdirs: config (MCA v2.0, API v2.0, Component v1.8.5) > MCA memory: linux (MCA v2.0, API v2.0, Component v1.8.5) > MCA pstat: linux (MCA v2.0, API v2.0, Component v1.8.5) > MCA sec: basic (MCA v2.0, API v1.0, Component v1.8.5) > MCA shmem: mmap (MCA v2.0, API v2.0, Component v1.8.5) > MCA shmem: posix (MCA v2.0, API v2.0, Component v1.8.5) > MCA shmem: sysv (MCA v2.0, API v2.0, Component v1.8.5) > MCA timer: linux (MCA v2.0, API v2.0, Component v1.8.5) > MCA dfs: app (MCA v2.0, API v1.0, Component v1.8.5) > MCA dfs: orted (MCA v2.0, API v1.0, Component v1.8.5) > MCA dfs: test (MCA v2.0, API v1.0, Component v1.8.5) > MCA errmgr: default_app (MCA v2.0, API v3.0, Component v1.8.5) > MCA errmgr: default_hnp (MCA v2.0, API v3.0, Component v1.8.5) > MCA errmgr: default_orted (MCA v2.0, API v3.0, Component v1.8.5) > MCA errmgr: default_tool (MCA v2.0, API v3.0, Component v1.8.5) > MCA ess: env (MCA v2.0, API v3.0, Component v1.8.5) > MCA ess: hnp (MCA v2.0, API v3.0, Component v1.8.5) > MCA ess: singleton (MCA v2.0, API v3.0, Component v1.8.5) > MCA ess: slurm (MCA v2.0, API v3.0, Component v1.8.5) > MCA ess: tool (MCA v2.0, API v3.0, Component v1.8.5) > MCA filem: raw (MCA v2.0, API v2.0, Component v1.8.5) > MCA grpcomm: bad (MCA v2.0, API v2.0, Component v1.8.5) > MCA iof: hnp (MCA v2.0, API v2.0, Component v1.8.5) > MCA iof: mr_hnp (MCA v2.0, API v2.0, Component v1.8.5) > MCA iof: mr_orted (MCA v2.0, API v2.0, Component v1.8.5) > MCA iof: orted (MCA v2.0, API v2.0, Component v1.8.5) > MCA iof: tool (MCA v2.0, API v2.0, Component v1.8.5) > MCA odls: default (MCA v2.0, API v2.0, Component v1.8.5) > MCA oob: tcp (MCA v2.0, API v2.0, Component v1.8.5) > MCA plm: isolated (MCA v2.0, API v2.0, Component v1.8.5) > MCA plm: rsh (MCA v2.0, API v2.0, Component v1.8.5) > MCA plm: slurm (MCA v2.0, API v2.0, Component v1.8.5) > MCA ras: loadleveler (MCA v2.0, API v2.0, Component v1.8.5) > MCA ras: simulator (MCA v2.0, API v2.0, Component v1.8.5) > MCA ras: slurm (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: lama (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: mindist (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: ppr (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: rank_file (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: resilient (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: round_robin (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: seq (MCA v2.0, API v2.0, Component v1.8.5) > MCA rmaps: staged (MCA v2.0, API v2.0, Component v1.8.5) > MCA rml: oob (MCA v2.0, API v2.0, Component v1.8.5) > MCA routed: binomial (MCA v2.0, API v2.0, Component v1.8.5) > MCA routed: debruijn (MCA v2.0, API v2.0, Component v1.8.5) > MCA routed: direct (MCA v2.0, API v2.0, Component v1.8.5) > MCA routed: radix (MCA v2.0, API v2.0, Component v1.8.5) > MCA state: app (MCA v2.0, API v1.0, Component v1.8.5) > MCA state: hnp (MCA v2.0, API v1.0, Component v1.8.5) > MCA state: novm (MCA v2.0, API v1.0, Component v1.8.5) > MCA state: orted (MCA v2.0, API v1.0, Component v1.8.5) > MCA state: staged_hnp (MCA v2.0, API v1.0, Component v1.8.5) > MCA state: staged_orted (MCA v2.0, API v1.0, Component v1.8.5) > MCA state: tool (MCA v2.0, API v1.0, Component v1.8.5) > MCA allocator: basic (MCA v2.0, API v2.0, Component v1.8.5) > MCA allocator: bucket (MCA v2.0, API v2.0, Component v1.8.5) > MCA bcol: basesmuma (MCA v2.0, API v2.0, Component v1.8.5) > MCA bcol: ptpcoll (MCA v2.0, API v2.0, Component v1.8.5) > MCA bml: r2 (MCA v2.0, API v2.0, Component v1.8.5) > MCA btl: openib (MCA v2.0, API v2.0, Component v1.8.5) > MCA btl: self (MCA v2.0, API v2.0, Component v1.8.5) > MCA btl: sm (MCA v2.0, API v2.0, Component v1.8.5) > MCA btl: tcp (MCA v2.0, API v2.0, Component v1.8.5) > MCA btl: vader (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: basic (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: hierarch (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: inter (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: libnbc (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: ml (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: self (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: sm (MCA v2.0, API v2.0, Component v1.8.5) > MCA coll: tuned (MCA v2.0, API v2.0, Component v1.8.5) > MCA dpm: orte (MCA v2.0, API v2.0, Component v1.8.5) > MCA fbtl: posix (MCA v2.0, API v2.0, Component v1.8.5) > MCA fcoll: dynamic (MCA v2.0, API v2.0, Component v1.8.5) > MCA fcoll: individual (MCA v2.0, API v2.0, Component v1.8.5) > MCA fcoll: static (MCA v2.0, API v2.0, Component v1.8.5) > MCA fcoll: two_phase (MCA v2.0, API v2.0, Component v1.8.5) > MCA fcoll: ylib (MCA v2.0, API v2.0, Component v1.8.5) > MCA fs: ufs (MCA v2.0, API v2.0, Component v1.8.5) > MCA io: ompio (MCA v2.0, API v2.0, Component v1.8.5) > MCA io: romio (MCA v2.0, API v2.0, Component v1.8.5) > MCA mpool: grdma (MCA v2.0, API v2.0, Component v1.8.5) > MCA mpool: sm (MCA v2.0, API v2.0, Component v1.8.5) > MCA osc: rdma (MCA v2.0, API v3.0, Component v1.8.5) > MCA osc: sm (MCA v2.0, API v3.0, Component v1.8.5) > MCA pml: v (MCA v2.0, API v2.0, Component v1.8.5) > MCA pml: bfo (MCA v2.0, API v2.0, Component v1.8.5) > MCA pml: cm (MCA v2.0, API v2.0, Component v1.8.5) > MCA pml: ob1 (MCA v2.0, API v2.0, Component v1.8.5) > MCA pubsub: orte (MCA v2.0, API v2.0, Component v1.8.5) > MCA rcache: vma (MCA v2.0, API v2.0, Component v1.8.5) > MCA rte: orte (MCA v2.0, API v2.0, Component v1.8.5) > MCA sbgp: basesmsocket (MCA v2.0, API v2.0, Component v1.8.5) > MCA sbgp: basesmuma (MCA v2.0, API v2.0, Component v1.8.5) > MCA sbgp: p2p (MCA v2.0, API v2.0, Component v1.8.5) > MCA sharedfp: individual (MCA v2.0, API v2.0, Component v1.8.5) > MCA sharedfp: lockedfile (MCA v2.0, API v2.0, Component v1.8.5) > MCA sharedfp: sm (MCA v2.0, API v2.0, Component v1.8.5) > MCA topo: basic (MCA v2.0, API v2.1, Component v1.8.5) > MCA vprotocol: pessimist (MCA v2.0, API v2.0, Component v1.8.5) > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2015/04/26713.php
pgpLjt7J02Q7q.pgp
Description: PGP signature