Thanks for all the info!

But still, can we get a copy of the test in C?  That would make it 
significantly easier for us to tell if there is a problem with Open MPI -- 
mainly because we don't know anything about the internals of mpi4py.


On May 23, 2012, at 5:43 PM, Bennet Fauber wrote:

> Thanks, Ralph,
> 
> On Wed, 23 May 2012, Ralph Castain wrote:
> 
>> I don't honestly think many of us have any knowledge of mpi4py. Does this 
>> test work with other MPIs?
> 
> The mpi4py developers have said they've never seen this using mpich2.  I have 
> not been able to test that myself.
> 
>> MPI_Allgather seems to be passing our tests, so I suspect it is something in 
>> the binding. If you can provide the actual test, I'm willing to take a look 
>> at it.
> 
> The actual test is included in the install bundle for mpi4py, along with the 
> C source code used to create the bindings.
> 
>       http://code.google.com/p/mpi4py/downloads/list
> 
> The install is straightforward and simple.  Unpack the tarball, make sure 
> that mpicc is in your path
> 
>       $ cd mpi4py-1.3
>       $ python setup.py build
>       $ python setup.py install --prefix=/your/install
>       $ export PYTHONPATH=/your/install/lib/pythonN.M/site-packages
>       $ mpirun -np 5 python test/runtests.py \
>                --verbose --no-threads --include cco_obj_inter
> 
> where N.M are the major.minor numbers of your python distribution.
> 
> What I find most puzzling is that, maybe, 1 out of 10 times it will run to 
> completion with -np 5, and it runs with all other numbers of processors I've 
> tested always.
> 
>                       -- bennet
> 
>> On May 23, 2012, at 2:52 PM, Bennet Fauber wrote:
>> 
>>> I've installed the latest mpi4py-1.3 on several systems, and there is a 
>>> repeated bug when running
>>> 
>>>     $ mpirun -np 5 python test/runtests.py
>>> 
>>> where it throws an error on mpigather with openmpi-1.4.4 and hangs with 
>>> openmpi-1.3.
>>> 
>>> It runs to completion and passes all tests when run with -np of 2, 3, 4, 6, 
>>> 7, 8, 9, 10, 11, and 12.
>>> 
>>> There is a thread on this at
>>> 
>>> http://groups.google.com/group/mpi4py/browse_thread/thread/509ac46af6f79973
>>> 
>>> where others report being able to replicate, too.
>>> 
>>> The compiler used first was gcc-4.6.2, with openmpi-1.4.4.
>>> 
>>> These are all Red Hat machines, RHEL 5 or 6 and with multiple compilers and 
>>> versions of openmpi 1.3.0 and 1.4.4.
>>> 
>>> Lisandro who is the primary developer of mpi4py is able to replicate on 
>>> Fedora 16.
>>> 
>>> Someone else is able to reproduce with
>>> 
>>> [ quoting from the groups.google.com page... ]
>>> ===============================================================
>>> It also happens with the current hg version of mpi4py and
>>> $ rpm -qa openmpi gcc python
>>> python-2.7.3-6.fc17.x86_64
>>> gcc-4.7.0-5.fc17.x86_64
>>> openmpi-1.5.4-5.fc17.1.x86_64
>>> ===============================================================
>>> 
>>> So, I believe this is a bug to be reported.  Per the advice at
>>> 
>>>     http://www.open-mpi.org/community/help/bugs.php
>>> 
>>>     If you feel that you do have a definite bug to report but are
>>>     unsure which list to post to, then post to the user's list.
>>> 
>>> Please let me know if there is additional information that you need to 
>>> replicate.
>>> 
>>> Some output is included below the signature in case it is useful.
>>> 
>>>                     -- bennet
>>> --
>>> East Hall Technical Services
>>> Mathematics and Psychology Research Computing
>>> University of Michigan
>>> (734) 763-1182
>>> 
>>> On RHEL 5, openmpi 1.3, gcc 4.1.2, python 2.7
>>> 
>>> $ mpirun -np 5 --mca btl ^sm python test/runtests.py --verbose --no-threads 
>>> --include cco_obj_inter
>>> [0...@sirocco.math.lsa.umich.edu] Python 2.7 
>>> (/home/bennet/epd7.2.2/bin/python)
>>> [0...@sirocco.math.lsa.umich.edu] MPI 2.0 (Open MPI 1.3.0)
>>> [0...@sirocco.math.lsa.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.7/mpi4py)
>>> [1...@sirocco.math.lsa.umich.edu] Python 2.7 
>>> (/home/bennet/epd7.2.2/bin/python)
>>> [1...@sirocco.math.lsa.umich.edu] MPI 2.0 (Open MPI 1.3.0)
>>> [1...@sirocco.math.lsa.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.7/mpi4py)
>>> [2...@sirocco.math.lsa.umich.edu] Python 2.7 
>>> (/home/bennet/epd7.2.2/bin/python)
>>> [2...@sirocco.math.lsa.umich.edu] MPI 2.0 (Open MPI 1.3.0)
>>> [2...@sirocco.math.lsa.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.7/mpi4py)
>>> [3...@sirocco.math.lsa.umich.edu] Python 2.7 
>>> (/home/bennet/epd7.2.2/bin/python)
>>> [3...@sirocco.math.lsa.umich.edu] MPI 2.0 (Open MPI 1.3.0)
>>> [3...@sirocco.math.lsa.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.7/mpi4py)
>>> [4...@sirocco.math.lsa.umich.edu] Python 2.7 
>>> (/home/bennet/epd7.2.2/bin/python)
>>> [4...@sirocco.math.lsa.umich.edu] MPI 2.0 (Open MPI 1.3.0)
>>> [4...@sirocco.math.lsa.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.7/mpi4py)
>>> testAllgather (test_cco_obj_inter.TestCCOObjInter) ... testAllgather 
>>> (test_cco_obj_inter.TestCCOObjInter) ... testAllgather 
>>> (test_cco_obj_inter.TestCCOObjInter) ... testAllgather 
>>> (test_cco_obj_inter.TestCCOObjInter) ... testAllgather 
>>> (test_cco_obj_inter.TestCCOObjInter) ...
>>> [ hangs ]
>>> 
>>> RHEL5
>>> ===================================================
>>> $ python
>>> Python 2.6.6 (r266:84292, Sep 12 2011, 14:03:14)
>>> [GCC 4.4.5 20110214 (Red Hat 4.4.5-6)] on linux2
>>> 
>>> $ gcc -v
>>> Using built-in specs.
>>> COLLECT_GCC=gcc
>>> COLLECT_LTO_WRAPPER=/home/software/rhel6/gcc/4.7.0/libexec/gcc/x86_64-
>>> unknown-linux-gnu/4.7.0/lto-wrapper
>>> Target: x86_64-unknown-linux-gnu
>>> Configured with: ../gcc-4.7.0/configure --prefix=/home/software/rhel6/
>>> gcc/4.7.0 --with-mpfr=/home/software/rhel6/gcc/mpfr-3.1.0/ --with-mpc=/
>>> home/software/rhel6/gcc/mpc-0.9/ --with-gmp=/home/software/rhel6/gcc/
>>> gmp-5.0.5/ --disable-multilib
>>> Thread model: posix
>>> gcc version 4.7.0 (GCC)
>>> 
>>> $ mpirun -np 5 python test/runtests.py --verbose --no-threads --include 
>>> cco_obj_inter
>>> [4...@host-rh6.engin.umich.edu] Python 2.6 (/usr/bin/python)
>>> [4...@host-rh6.engin.umich.edu] MPI 2.1 (Open MPI 1.6.0)
>>> [4...@host-rh6.engin.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.6/mpi4py)
>>> [2...@host-rh6.engin.umich.edu] Python 2.6 (/usr/bin/python)
>>> [2...@host-rh6.engin.umich.edu] MPI 2.1 (Open MPI 1.6.0)
>>> [2...@host-rh6.engin.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.6/mpi4py)
>>> [1...@host-rh6.engin.umich.edu] Python 2.6 (/usr/bin/python)
>>> [1...@host-rh6.engin.umich.edu] MPI 2.1 (Open MPI 1.6.0)
>>> [1...@host-rh6.engin.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.6/mpi4py)
>>> [0...@host-rh6.engin.umich.edu] Python 2.6 (/usr/bin/python)
>>> [0...@host-rh6.engin.umich.edu] MPI 2.1 (Open MPI 1.6.0)
>>> [0...@host-rh6.engin.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.6/mpi4py)
>>> [3...@host-rh6.engin.umich.edu] Python 2.6 (/usr/bin/python)
>>> [3...@host-rh6.engin.umich.edu] MPI 2.1 (Open MPI 1.6.0)
>>> [3...@host-rh6.engin.umich.edu] mpi4py 1.3 
>>> (build/lib.linux-x86_64-2.6/mpi4py)
>>> testAllgather (test_cco_obj_inter.TestCCOObjInter) ... testAllgather
>>> (test_cco_obj_inter.TestCCOObjInter) ... testAllgather
>>> (test_cco_obj_inter.TestCCOObjInter) ... testAllgather
>>> (test_cco_obj_inter.TestCCOObjInter) ... testAllgather
>>> (test_cco_obj_inter.TestCCOObjInter) ... ERROR
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to