On Mon, Aug 22, 2011 at 08:38:52AM -0700, Tom Rosmond wrote: > Yes, we are using collective I/O (mpi_file_write_at_all, > mpi_file_read_at_all). The swaping of fortran and mpi-io are just > branches in the code at strategic locations. Although the mpi-io files > are readable with fortran direct access, we don't do so from within the > application because of different data organization in the files. > > > Do you use MPI datatypes to describe either a file view or the > > application data? These noncontiguous in memory and/or noncontiguous > > in file access patterns will also trigger fcntl lock calls. You can > > use an MPI-IO hint to disable data sieving, at a potentially > > disastrous performance cost. > > Yes, we use an 'mpi_type_indexed' datatype to describe the data > organization. > > Any thoughts about the XFS vs EXT3 question?
We have machines at the lab with XFS and machines with EXT3: I can't say I have ever seen an MPI-IO problem we could trace to the specific file system. The MPI-IO library just makes a bunch of posix I/O calls under the hood: if write(2), open(2), and friends are broken for XFS or EXT3, those kinds of bugs get a lot of attention :> At this point the usual course of action is "post a small reproducing test case". Your first message said this was a big code, so perhaps that will not be so easy... ==rob -- Rob Latham Mathematics and Computer Science Division Argonne National Lab, IL USA