I haven't had time to do the openmpi build from the nightly yet, but
my user has run some more tests and now has a simple program and
algorithm to "break" openmpi.  His notes:

hey, just fyi, I can reproduce the error readily in a simple test case
my "way to break mpi" is as follows: Master proc runs MPI_Send 1000
times to each child, then waits for a "I got it" ack from each child.
Each child receives 1000 numbers from the Master, then sends "I got
it" to the master
running this on 25 nodes causes it to break about 60% of the time
interestingly, it usually breaks on the same process number each time

ah. It looks like if I let it sit for about 5 minutes, sometimes it
will work. From my log
rank: 23 Mon Feb 23 13:29:44 2009 recieved 816
rank: 23 Mon Feb 23 13:29:44 2009 recieved 817
rank: 23 Mon Feb 23 13:29:44 2009 recieved 818
rank: 23 Mon Feb 23 13:33:08 2009 recieved 819
rank: 23 Mon Feb 23 13:33:08 2009 recieved 820

Any thoughts on this problem?
(this is the only reason I'm currently working on upgrading openmpi)

--Jim

On Fri, Feb 20, 2009 at 1:59 PM, Jeff Squyres <jsquy...@cisco.com> wrote:
> There won't be an official SRPM until 1.3.1 is released.
>
> But to test if 1.3.1 is on-track to deliver a proper solution to you, can
> you try a nightly tarball, perhaps in conjunction with our "buildrpm.sh"
> script?
>
>
>  https://svn.open-mpi.org/source/xref/ompi_1.3/contrib/dist/linux/buildrpm.sh
>
> It should build a trivial SRPM for you from the tarball.  You'll likely need
> to get the specfile, too, and put it in the same dir as buildrpm.sh.  The
> specfile is in the same SVN directory:
>
>
>  https://svn.open-mpi.org/source/xref/ompi_1.3/contrib/dist/linux/openmpi.spec
>
>
>
> On Feb 20, 2009, at 3:51 PM, Jim Kusznir wrote:
>
>> As long as I can still build the rpm for it and install it via rpm.
>> I'm running it on a ROCKS cluster, so it needs to be an RPM to get
>> pushed out to the compute nodes.
>>
>> --Jim
>>
>> On Fri, Feb 20, 2009 at 11:30 AM, Jeff Squyres <jsquy...@cisco.com> wrote:
>>>
>>> On Feb 20, 2009, at 2:20 PM, Jim Kusznir wrote:
>>>
>>>> I just went to www.open-mpi.org, went to download, then source rpm.
>>>> Looks like it was actually 1.3-1.  Here's the src.rpm that I pulled
>>>> in:
>>>>
>>>>
>>>> http://www.open-mpi.org/software/ompi/v1.3/downloads/openmpi-1.3-1.src.rpm
>>>
>>> Ah, gotcha.  Yes, that's 1.3.0, SRPM version 1.  We didn't make up this
>>> nomenclature.  :-(
>>>
>>>> The reason for this upgrade is it seems a user found some bug that may
>>>> be in the OpenMPI code that results in occasionally an MPI_Send()
>>>> message getting lost.  He's managed to reproduce it multiple times,
>>>> and we can't find anything in his code that can cause it...He's got
>>>> logs of mpi_send() going out, but the matching mpi_receive() never
>>>> getting anything, thus killing his code.  We're currently running
>>>> 1.2.8 with ofed support (Haven't tried turning off ofed, etc. yet).
>>>
>>> Ok.  1.3.x is much mo' betta' then 1.2 in many ways.  We could probably
>>> help
>>> track down the problem, but if you're willing to upgrade to 1.3.x, it'll
>>> hopefully just make the problem go away.
>>>
>>> Can you try a 1.3.1 nightly tarball?
>>>
>>> --
>>> Jeff Squyres
>>> Cisco Systems
>>>
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> Cisco Systems
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to