Hello Brian, Actually, I am not modifying the local communication buffer that contains the data. I am modifying the the buffer that contains the indices of the data buffer(source_disp and target_disp).
in MPICH2 this is not a problem. I am not sure about Open MPI. Thanks, Ziaul On Wed, Jun 6, 2012 at 1:05 PM, Barrett, Brian W <bwba...@sandia.gov> wrote: > Ziaul - > > Your program is erroneous; you can not modify the local communication > buffer of an MPI_ACCUMULATE call until after the next synchronization call > (Section 11.3 of MPI 2.2). In your example, that would be after the > MPI_FENCE call following the call to MPI_ACCUMULATE. > > Brian > > On 6/6/12 9:44 AM, "Ziaul Haque Olive" <mzh.ol...@gmail.com> wrote: > > >Hello, > > > >I am not sure, if my code is correct according to Open MPI(v1.6). the > > code is given as follows, I am doing MPI one-sided communication inside > > a function - data_transfer. this function is being called inside a > >fence epoch. inside data_transfer, I am allocating memory for > >non-contiguous data, creating derived data type, using this datatype in > >MPI_Accumulate, and after calling MPI_Accumulate, freeing the indexed > >data type and also freeing the memory containing indices for indexed > >data type. is it okay that I am freeing memory for derived datatype > >before the closing > >fence? > > > >I am getting segmentation fault with this code. if I comment out the > >MPI_Accumulate call, then no seg-fault occurs. > > > > > > > >void data_transfer(void *data, int > >*sources_disp, int *targets_disp, int *target, int size, int *blength, > >int func, MPI_Op op, MPI_Win win, MPI_Datatype dtype){ > > > > int i,j, index; > > int tmp_target; > > int *flag; > > int *source_disp; > > int *target_disp; > > MPI_Datatype source_type, target_type; > > > > > > MPI_Alloc_mem( size*sizeof(int), MPI_INFO_NULL, &source_disp); > > MPI_Alloc_mem( size*sizeof(int), MPI_INFO_NULL, &target_disp); > > MPI_Alloc_mem( size*sizeof(int), MPI_INFO_NULL, &flag ); > > > > memset(flag, 0, size*sizeof(int)); > > > > for(i=0;i<size;i++){ > > if(flag[i]==0){ > > tmp_target = target[i]; > > > > index = 0; > > for(j=i; j<size; j++){ > > if(flag[j]==0 && tmp_target == target[j] ){ > > source_disp[index] = sources_disp[j]; > > target_disp[index] = targets_disp[j]; > > //printf("src, target disp %d %d\n", j, disp[j]); > > index++; > > flag[j] = 1; > > } > > } > > > > MPI_Type_indexed(index, blength , source_disp, dtype, > >&source_type); > > MPI_Type_commit(&source_type); > > MPI_Type_indexed(index, blength , target_disp, dtype, > >&target_type); > > MPI_Type_commit(&target_type); > > > > > > MPI_Accumulate( data, 1, source_type, tmp_target, 0, 1, > >target_type , op, win); > > > > MPI_Type_free(&source_type); > > MPI_Type_free(&target_type); > > } > > } > > MPI_Free_mem(source_disp); > > MPI_Free_mem(target_disp); > > MPI_Free_mem(flag); > > > >} > > > >void main(){ > > int i; > > while(i<N){ > > MPI_Win_fence(MPI_MODE_NOPRECEDE, queue2_win); > > > > data_transfer(); > > > > MPI_Win_fence(MPI_MODE_NOSUCCEED, queue2_win); > > } > >} > > > >thanks, > >Ziaul > > > > > >_______________________________________________ > >users mailing list > >us...@open-mpi.org > >http://www.open-mpi.org/mailman/listinfo.cgi/users > > > -- > Brian W. Barrett > Dept. 1423: Scalable System Software > Sandia National Laboratories > > > > > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >