summaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorQuincey Koziol <koziol@hdfgroup.org>2011-12-02 13:40:37 (GMT)
committerQuincey Koziol <koziol@hdfgroup.org>2011-12-02 13:40:37 (GMT)
commit51d81cb5ef9e07c080ed08e762da5c986305392d (patch)
treee2857483cc9cd7c3fa1dd8af059112bf2daea7ee /src
parentd34c5fdb79946a97bdf81744f425ca7cfb8b9047 (diff)
downloadhdf5-51d81cb5ef9e07c080ed08e762da5c986305392d.zip
hdf5-51d81cb5ef9e07c080ed08e762da5c986305392d.tar.gz
hdf5-51d81cb5ef9e07c080ed08e762da5c986305392d.tar.bz2
[svn-r21796] Description:
Back out r21782 while I figure out what the problem is with the change. Tested on: Daily tests... :-/
Diffstat (limited to 'src')
-rw-r--r--src/H5Dmpio.c63
1 files changed, 30 insertions, 33 deletions
diff --git a/src/H5Dmpio.c b/src/H5Dmpio.c
index f2f3efa..8c78af2 100644
--- a/src/H5Dmpio.c
+++ b/src/H5Dmpio.c
@@ -172,43 +172,10 @@ H5D_mpio_opt_possible(const H5D_io_info_t *io_info, const H5S_t *file_space,
HDassert(file_space);
HDassert(type_info);
- /* Global decisions. These conditions are common to all processes and they
- * will all arrive at the same conclusions, without requiring communication.
- */
-
/* For independent I/O, get out quickly and don't try to form consensus */
if(io_info->dxpl_cache->xfer_mode == H5FD_MPIO_INDEPENDENT)
HGOTO_DONE(FALSE);
- /* Optimized MPI types flag must be set and it must be collective IO */
- /* (Don't allow parallel I/O for the MPI-posix driver, since it doesn't do real collective I/O) */
- /* (Global decision if all processes share the same environment, which is reasonable) */
- if(!(H5S_mpi_opt_types_g && io_info->dxpl_cache->xfer_mode == H5FD_MPIO_COLLECTIVE
- && !IS_H5FD_MPIPOSIX(io_info->dset->oloc.file)))
- HGOTO_DONE(FALSE);
-
- /* Dataset storage must be contiguous or chunked */
- if(!(io_info->dset->shared->layout.type == H5D_CONTIGUOUS ||
- io_info->dset->shared->layout.type == H5D_CHUNKED))
- HGOTO_DONE(FALSE);
-
- /* The handling of memory space is different for chunking and contiguous
- * storage. For contiguous storage, mem_space and file_space won't change
- * when it it is doing disk IO. For chunking storage, mem_space will
- * change for different chunks. So for chunking storage, whether we can
- * use collective IO will defer until each chunk IO is reached.
- */
-
- /* Don't allow collective operations if filters need to be applied */
- if(io_info->dset->shared->layout.type == H5D_CHUNKED &&
- io_info->dset->shared->dcpl_cache.pline.nused > 0)
- HGOTO_DONE(FALSE);
-
-
- /* Local decisions. These conditions could be different on each process,
- * and require communication to verify.
- */
-
/* Don't allow collective operations if datatype conversions need to happen */
if(!type_info->is_conv_noop) {
local_opinion = FALSE;
@@ -221,6 +188,14 @@ H5D_mpio_opt_possible(const H5D_io_info_t *io_info, const H5S_t *file_space,
goto broadcast;
} /* end if */
+ /* Optimized MPI types flag must be set and it must be collective IO */
+ /* (Don't allow parallel I/O for the MPI-posix driver, since it doesn't do real collective I/O) */
+ if(!(H5S_mpi_opt_types_g && io_info->dxpl_cache->xfer_mode == H5FD_MPIO_COLLECTIVE
+ && !IS_H5FD_MPIPOSIX(io_info->dset->oloc.file))) {
+ local_opinion = FALSE;
+ goto broadcast;
+ } /* end if */
+
/* Check whether these are both simple or scalar dataspaces */
if(!((H5S_SIMPLE == H5S_GET_EXTENT_TYPE(mem_space) || H5S_SCALAR == H5S_GET_EXTENT_TYPE(mem_space))
&& (H5S_SIMPLE == H5S_GET_EXTENT_TYPE(file_space) || H5S_SCALAR == H5S_GET_EXTENT_TYPE(file_space)))) {
@@ -235,6 +210,28 @@ H5D_mpio_opt_possible(const H5D_io_info_t *io_info, const H5S_t *file_space,
goto broadcast;
} /* end if */
+ /* Dataset storage must be contiguous or chunked */
+ if(!(io_info->dset->shared->layout.type == H5D_CONTIGUOUS ||
+ io_info->dset->shared->layout.type == H5D_CHUNKED)) {
+ local_opinion = FALSE;
+ goto broadcast;
+ } /* end if */
+
+ /* The handling of memory space is different for chunking and contiguous
+ * storage. For contiguous storage, mem_space and file_space won't change
+ * when it it is doing disk IO. For chunking storage, mem_space will
+ * change for different chunks. So for chunking storage, whether we can
+ * use collective IO will defer until each chunk IO is reached.
+ */
+
+ /* Don't allow collective operations if filters need to be applied */
+ if(io_info->dset->shared->layout.type == H5D_CHUNKED) {
+ if(io_info->dset->shared->dcpl_cache.pline.nused > 0) {
+ local_opinion = FALSE;
+ goto broadcast;
+ } /* end if */
+ } /* end if */
+
broadcast:
/* Form consensus opinion among all processes about whether to perform
* collective I/O