summaryrefslogtreecommitdiffstats
path: root/src/H5Dio.c
diff options
context:
space:
mode:
authorMuQun Yang <ymuqun@hdfgroup.org>2006-02-20 20:47:25 (GMT)
committerMuQun Yang <ymuqun@hdfgroup.org>2006-02-20 20:47:25 (GMT)
commit593f1a9a5178b0ab079f47f8f395cdba41f1a78c (patch)
treee9840a728eb87daeab09ae8ec60607d56b1ccb41 /src/H5Dio.c
parent52f2c4da502620f5981b0b52b85f033f9930b884 (diff)
downloadhdf5-593f1a9a5178b0ab079f47f8f395cdba41f1a78c.zip
hdf5-593f1a9a5178b0ab079f47f8f395cdba41f1a78c.tar.gz
hdf5-593f1a9a5178b0ab079f47f8f395cdba41f1a78c.tar.bz2
[svn-r11959] Purpose:
bug fix Description: During the process of the combination of collective chunk IO code and the original IO code, The return value of a function call hasn't been initialized. On 32-bit mode at modi4, that value isn't set to 0 by default. So a fake error value returns. Solution: Re-arrange the code so that the return value will be set properly. Platforms tested: parallel: modi4 - 32bit heping seq: heping Misc. update:
Diffstat (limited to 'src/H5Dio.c')
-rw-r--r--src/H5Dio.c30
1 files changed, 12 insertions, 18 deletions
diff --git a/src/H5Dio.c b/src/H5Dio.c
index 6b62ca9..3cdc799 100644
--- a/src/H5Dio.c
+++ b/src/H5Dio.c
@@ -988,15 +988,16 @@ H5D_contig_read(H5D_io_info_t *io_info, hsize_t nelmts,
#ifdef H5_HAVE_PARALLEL
if(io_info->dxpl_cache->xfer_mode == H5FD_MPIO_COLLECTIVE) {
if(H5D_contig_collective_io(io_info,file_space,mem_space,buf,FALSE)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't manipulate collective I/O");
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "contiguous read failed in collective mode");
}
else
#endif
{
- status = (io_info->ops.read)(io_info,
+ if((io_info->ops.read)(io_info,
(size_t)nelmts, H5T_get_size(dataset->shared->type),
file_space, mem_space,0,
- buf/*out*/);
+ buf/*out*/)<0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "contiguous read failed ");
}
#ifdef H5S_DEBUG
@@ -1005,10 +1006,6 @@ H5D_contig_read(H5D_io_info_t *io_info, hsize_t nelmts,
io_info->stats->stats[1].read_ncalls++;
#endif
- /* Check return value from optimized read */
- if (status<0) {
- HGOTO_ERROR(H5E_DATASET, H5E_READERROR, FAIL, "optimized read failed")
- } else
/* direct xfer accomplished successfully */
HGOTO_DONE(SUCCEED)
} /* end if */
@@ -1250,15 +1247,16 @@ H5D_contig_write(H5D_io_info_t *io_info, hsize_t nelmts,
#ifdef H5_HAVE_PARALLEL
if(io_info->dxpl_cache->xfer_mode == H5FD_MPIO_COLLECTIVE) {
if(H5D_contig_collective_io(io_info,file_space,mem_space,buf,TRUE)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't manipulate collective I/O");
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "contiguous write failed in collective mode");
}
else
#endif
{
- status = (io_info->ops.write)(io_info,
+ if((io_info->ops.write)(io_info,
(size_t)nelmts, H5T_get_size(dataset->shared->type),
file_space, mem_space,0,
- buf/*out*/);
+ buf/*out*/)<0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "contiguous write failed ");
}
#ifdef H5S_DEBUG
@@ -1267,10 +1265,6 @@ H5D_contig_write(H5D_io_info_t *io_info, hsize_t nelmts,
io_info->stats->stats[0].write_ncalls++;
#endif
- /* Check return value from optimized write */
- if (status<0) {
- HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "optimized write failed")
- } else
/* direct xfer accomplished successfully */
HGOTO_DONE(SUCCEED)
} /* end if */
@@ -1518,7 +1512,7 @@ H5D_chunk_read(H5D_io_info_t *io_info, hsize_t nelmts,
/* Temporarily shut down collective IO for chunking */
if(io_info->dxpl_cache->xfer_mode == H5FD_MPIO_COLLECTIVE) {
if(H5D_chunk_collective_io(io_info,&fm,buf,FALSE)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't manipulate collective I/O");
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "chunked read failed in collective mode");
}
else {/* sequential or independent read */
@@ -1543,7 +1537,7 @@ H5D_chunk_read(H5D_io_info_t *io_info, hsize_t nelmts,
/* Check return value from optimized read */
if (status<0)
- HGOTO_ERROR(H5E_DATASET, H5E_READERROR, FAIL, " optimized read failed")
+ HGOTO_ERROR(H5E_DATASET, H5E_READERROR, FAIL, " chunked read failed")
chunk_node = H5SL_next(chunk_node);
@@ -1843,7 +1837,7 @@ H5D_chunk_write(H5D_io_info_t *io_info, hsize_t nelmts,
}
if(io_info->dxpl_cache->xfer_mode == H5FD_MPIO_COLLECTIVE) {
if(H5D_chunk_collective_io(io_info,&fm,buf,TRUE)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't manipulate collective I/O");
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "chunked write failed in collective mode");
}
else {/* sequential or independent write */
@@ -1869,7 +1863,7 @@ H5D_chunk_write(H5D_io_info_t *io_info, hsize_t nelmts,
/* Check return value from optimized read */
if (status<0)
- HGOTO_ERROR(H5E_DATASET, H5E_READERROR, FAIL, " optimized read failed")
+ HGOTO_ERROR(H5E_DATASET, H5E_READERROR, FAIL, " chunked write failed")
chunk_node = H5SL_next(chunk_node);