Hi,

you are right, I should have provided C++ and Fortran example, so I am doing
now


Here is "cplusplus.cpp"

#include <mpi.h>
#include <iostream>
using namespace std;
int main()
{
    MPI::Init();
    char command[] = "./a.out";
    MPI::Info info;
    MPI::Intercomm child = MPI::COMM_WORLD.Spawn(command, NULL, 8,info, 0);
    int a[8]={0,0,0,0,0,0,0,0};
    int dummy;
    child.Allgather(&dummy, 0, MPI::INT, a, 1, MPI::INT);
    child.Disconnect();
    cout << "a[";
    for ( int i = 0; i < 7; i++ )
        cout << a[i] << ",";
    cout << a[7] << "]" << endl;

    MPI::Finalize();
}


Here is again "fortran.f90"

program main
 use mpi
 implicit none
 integer :: parent, rank, val, dummy, ierr
 call MPI_Init(ierr)
 call MPI_Comm_get_parent(parent, ierr)
 call MPI_Comm_rank(parent, rank, ierr)
 val = rank + 1
 call MPI_Allgather(val,   1, MPI_INTEGER, &
                    dummy, 0, MPI_INTEGER, &
                    parent, ierr)
 call MPI_Comm_disconnect(parent, ierr)
 call MPI_Finalize(ierr)
end program main

here is how you build and run

-bash-3.2$ mpif90 fortran.f90
-bash-3.2$ mpiCC -o parent cplusplus.cpp
-bash-3.2$ ./parent
a[0,0,0,0,0,0,0,0]



If I use mpich2,
-bash-3.2$ mpif90 fortran.f90
-bash-3.2$ mpiCC -o parent cplusplus.cpp
-bash-3.2$ ./parent
a[1,2,3,4,5,6,7,8]

I hope that you can repeat this problem to see problem with OPENMPI,

Thanks,


On Thu, May 20, 2010 at 10:09 AM, Jeff Squyres <jsquy...@cisco.com> wrote:

> Can you send us an all-C or all-Fortran example that shows the problem?
>
> We don't have easy access to test through the python bindings.  ...ok, I
> admit it, it's laziness on my part.  :-)  But having a pure Open MPI test
> app would also remove some possible variables and possible sources of error.
>
>
> On May 20, 2010, at 9:43 AM, Battalgazi YILDIRIM wrote:
>
> > Hi Jody,
> >
> > I think that it is correct, you can  test this example in your desktop,
> >
> > thanks,
> >
> > On Thu, May 20, 2010 at 3:18 AM, jody <jody....@gmail.com> wrote:
> > Hi
> > I am really no python expert, but it looks to me as if you were
> > gathering arrays filled with zeroes:
> >  a = array('i', [0]) * n
> >
> > Shouldn't this line be
> >  a = array('i', [r])*n
> > where r is the rank of the process?
> >
> > Jody
> >
> >
> > On Thu, May 20, 2010 at 12:00 AM, Battalgazi YILDIRIM
> > <yildiri...@gmail.com> wrote:
> > > Hi,
> > >
> > >
> > > I am trying to use intercommunicator ::Allgather between two child
> process.
> > > I have fortran and Python code,
> > > I am using mpi4py for python. It seems that ::Allgather is not working
> > > properly in my desktop.
> > >
> > >  I have contacted first mpi4py developers (Lisandro Dalcin), he
> simplified
> > > my problem and provided two example files (python.py and fortran.f90,
> > > please see below).
> > >
> > > We tried with different MPI vendors, the following example worked
> correclty(
> > > it means the final print out should be array('i', [1, 2, 3, 4, 5, 6, 7,
> 8])
> > > )
> > >
> > > However, it is not giving correct answer in my two desktop (Redhat and
> > > ubuntu) both
> > > using OPENMPI
> > >
> > > Could yo look at this problem please?
> > >
> > > If you want to follow our discussion before you, you can go to
> following
> > > link:
> > >
> http://groups.google.com/group/mpi4py/browse_thread/thread/c17c660ae56ff97e
> > >
> > > yildirim@memosa:~/python_intercomm$ more python.py
> > > from mpi4py import MPI
> > > from array import array
> > > import os
> > >
> > > progr = os.path.abspath('a.out')
> > > child = MPI.COMM_WORLD.Spawn(progr,[], 8)
> > > n = child.remote_size
> > > a = array('i', [0]) * n
> > > child.Allgather([None,MPI.INT],[a,MPI.INT])
> > > child.Disconnect()
> > > print a
> > >
> > > yildirim@memosa:~/python_intercomm$ more fortran.f90
> > > program main
> > >  use mpi
> > >  implicit none
> > >  integer :: parent, rank, val, dummy, ierr
> > >  call MPI_Init(ierr)
> > >  call MPI_Comm_get_parent(parent, ierr)
> > >  call MPI_Comm_rank(parent, rank, ierr)
> > >  val = rank + 1
> > >  call MPI_Allgather(val,   1, MPI_INTEGER, &
> > >                     dummy, 0, MPI_INTEGER, &
> > >                     parent, ierr)
> > >  call MPI_Comm_disconnect(parent, ierr)
> > >  call MPI_Finalize(ierr)
> > > end program main
> > >
> > > yildirim@memosa:~/python_intercomm$ mpif90 fortran.f90
> > >
> > > yildirim@memosa:~/python_intercomm$ python python.py
> > > array('i', [0, 0, 0, 0, 0, 0, 0, 0])
> > >
> > >
> > > --
> > > B. Gazi YILDIRIM
> > >
> > > _______________________________________________
> > > users mailing list
> > > us...@open-mpi.org
> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > >
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> >
> >
> > --
> > B. Gazi YILDIRIM
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>
> --
> Jeff Squyres
> jsquy...@cisco.com
> For corporate legal information go to:
> http://www.cisco.com/web/about/doing_business/legal/cri/
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>



-- 
B. Gazi YILDIRIM

Reply via email to