I'm afraid there's on news so far. :-\ On May 14, 2012, at 4:21 PM, Aurélien Esnard wrote:
> > Hi, > > No news, good news... ? > > Aurélien Esnard :-) > > On 11/04/2011 01:17 PM, Jeff Squyres wrote: >> After some discussion on the devel list, I opened >> https://svn.open-mpi.org/trac/ompi/ticket/2904 to track the issue. >> >> >> On Oct 25, 2011, at 12:08 PM, Ralph Castain wrote: >> >>> FWIW: I have tracked this problem down. The fix is a little more >>> complicated then I'd like, so I'm going to have to ping some other folks to >>> ensure we concur on the approach before doing something. >>> >>> On Oct 25, 2011, at 8:20 AM, Ralph Castain wrote: >>> >>>> I still see it failing the test George provided on the trunk. I'm unaware >>>> of anyone looking further into it, though, as the prior discussion seemed >>>> to just end. >>>> >>>> On Oct 25, 2011, at 7:01 AM, orel wrote: >>>> >>>>> Dears, >>>>> >>>>> I try from several days to use advanced MPI2 features in the following >>>>> scenario : >>>>> >>>>> 1) a master code A (of size NPA) spawns (MPI_Comm_spawn()) two slave >>>>> codes B (of size NPB) and C (of size NPC), providing intercomms A-B and >>>>> A-C ; >>>>> 2) i create intracomm AB and AC by merging intercomms ; >>>>> 3) then i create intercomm AB-C by calling MPI_Intercomm_create() by >>>>> using AC as bridge... >>>>> >>>>> MPI_Comm intercommABC; A: MPI_Intercomm_create(intracommAB, 0, >>>>> intracommAC, NPA, TAG,&intercommABC); >>>>> B: MPI_Intercomm_create(intracommAB, 0, MPI_COMM_NULL, >>>>> 0,TAG,&intercommABC); >>>>> C: MPI_Intercomm_create(intracommC, 0, intracommAC, 0, TAG,&intercommABC); >>>>> >>>>> In these calls, A0 and C0 play the role of local leader for AB and C >>>>> respectively. >>>>> C0 and A0 play the roles of remote leader in bridge intracomm AC. >>>>> >>>>> 3) MPI_Barrier(intercommABC); >>>>> 4) i merge intercomm AB-C into intracomm ABC$ >>>>> 5) MPI_Barrier(intracommABC); >>>>> >>>>> My BUG: These calls success, but when i try to use intracommABC for a >>>>> collective communication like MPI_Barrier(), >>>>> i got the following error : >>>>> >>>>> *** An error occurred in MPI_Barrier >>>>> *** on communicator >>>>> *** MPI_ERR_INTERN: internal error >>>>> *** MPI_ERRORS_ARE_FATAL: your MPI job will now abort >>>>> >>>>> >>>>> I try with OpenMPI trunk, 1.5.3, 1.5.4 and Mpich2-1.4.1p1 >>>>> >>>>> My code works perfectly if intracomm A, B and C are obtained by >>>>> MPI_Comm_split() instead of MPI_Comm_spawn() !!!! >>>>> >>>>> >>>>> I found same problem in a previous thread of the OMPI Users mailing list : >>>>> >>>>> => http://www.open-mpi.org/community/lists/users/2011/06/16711.php >>>>> >>>>> Is that bug/problem is currently under investigation ? :-) >>>>> >>>>> i can give detailed code, but the one provided by George Bosilca in this >>>>> previous thread provides same error... >>>>> >>>>> Thank you to help me... >>>>> >>>>> -- >>>>> Aurélien Esnard >>>>> University Bordeaux 1 / LaBRI / INRIA (France) >>>>> _______________________________________________ >>>>> users mailing list >>>>> us...@open-mpi.org >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/