Dana,
would it then make sense for all outside filters to use these routines? Due to Parallel Compression's internal nature, it uses buffers allocated via H5MM_ routines to collect and scatter data, which works fine for the internal filters like deflate, since they use these as well. However, since some of the outside filters use the raw malloc/free routines, causing issues, I'm wondering if having all outside filters use the H5_ routines is the cleanest solution.. Michael, Based on the "num_writers: 4" field, the NULL "receive_requests_array" and the fact that for the same chunk, rank 0 shows "original owner: 0, new owner: 0" and rank 3 shows "original owner: 3, new_owner: 0", it seems as though everyone IS interested in the chunk the rank 0 is now working on, but now I'm more confident that at some point either the messages may have failed to send or rank 0 is having problems finding the messages. Since in the unfiltered case it won't hit this particular code path, I'm not surprised that that case succeeds. If I had to make another guess based on this, I would be inclined to think that rank 0 must be hanging on the MPI_Mprobe due to a mismatch in the "tag" field. I use the index of the chunk as the tag for the message in order to funnel specific messages to the correct rank for the correct chunk during the last part of the chunk redistribution and if rank 0 can't match the tag it of course won't find the message. Why this might be happening, I'm not entirely certain currently.
_______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5