summaryrefslogtreecommitdiffstats
path: root/src
diff options
context:
space:
mode:
authorQuincey Koziol <koziol@hdfgroup.org>2007-11-10 14:51:30 (GMT)
committerQuincey Koziol <koziol@hdfgroup.org>2007-11-10 14:51:30 (GMT)
commit9601db2b2b4ec2550b30ae7156ee53be3d6d02b2 (patch)
tree72cc45a80d99a5be7eea01b2796f7a85cc473a14 /src
parent1dc74b6177e2b5b7245a3146d13876954e8f80f9 (diff)
downloadhdf5-9601db2b2b4ec2550b30ae7156ee53be3d6d02b2.zip
hdf5-9601db2b2b4ec2550b30ae7156ee53be3d6d02b2.tar.gz
hdf5-9601db2b2b4ec2550b30ae7156ee53be3d6d02b2.tar.bz2
[svn-r14248] synced fast_append branch with trunk (bring over general dataset append record speed improvements)
Diffstat (limited to 'src')
-rw-r--r--src/H5Dint.c165
-rw-r--r--src/H5Dpkg.h1
2 files changed, 102 insertions, 64 deletions
diff --git a/src/H5Dint.c b/src/H5Dint.c
index 5584187..aebc6f2 100644
--- a/src/H5Dint.c
+++ b/src/H5Dint.c
@@ -45,6 +45,13 @@
/* Local Typedefs */
/******************/
+/* Struct for holding callback info during H5D_flush operation */
+typedef struct {
+ const H5F_t *f; /* Pointer to file being flushed */
+ hid_t dxpl_id; /* DXPL for I/O operations */
+ unsigned flags; /* Flags for flush operation */
+} H5D_flush_ud_t;
+
/********************/
/* Local Prototypes */
@@ -1484,6 +1491,13 @@ H5D_close(H5D_t *dataset)
dataset->shared->fo_count--;
if(dataset->shared->fo_count == 0) {
+ /* Update the dataspace on disk, if it's been changed */
+ if(dataset->shared->space_dirty) {
+ if(H5S_write(&(dataset->oloc), dataset->shared->space, TRUE, H5AC_dxpl_id) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace")
+ dataset->shared->space_dirty = FALSE;
+ } /* end if */
+
/* Free the data sieve buffer, if it's been allocated */
if(dataset->shared->cache.contig.sieve_buf) {
HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
@@ -2284,9 +2298,8 @@ H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id)
* Modify the dataset storage
*-------------------------------------------------------------------------
*/
- /* Save the new dataspace in the file if necessary */
- if(H5S_write(&(dset->oloc), space, TRUE, dxpl_id) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace")
+ /* Mark the dataspace as dirty, for later writing to the file */
+ dset->shared->space_dirty = TRUE;
/* Update the index values for the cached chunks for this dataset */
if(H5D_CHUNKED == dset->shared->layout.type)
@@ -2332,93 +2345,117 @@ done:
/*-------------------------------------------------------------------------
- * Function: H5D_flush
+ * Function: H5D_flush_cb
*
* Purpose: Flush any dataset information cached in memory
*
* Return: Success: Non-negative
* Failure: Negative
*
- * Programmer: Ray Lu
- * August 14, 2002
+ * Programmer: Quincey Koziol
+ * November 8, 2007
*
*-------------------------------------------------------------------------
*/
-herr_t
-H5D_flush(const H5F_t *f, hid_t dxpl_id, unsigned flags)
+static int
+H5D_flush_cb(void *_dataset, hid_t UNUSED id, void *_udata)
{
- hid_t *id_list = NULL; /* Array of dataset IDs */
- unsigned num_dsets; /* Number of datasets in file */
- herr_t ret_value = SUCCEED; /* Return value */
+ H5D_t *dataset = (H5D_t *)_dataset; /* Dataset pointer */
+ H5D_flush_ud_t *udata = (H5D_flush_ud_t *)_udata; /* User data for callback */
+ int ret_value = H5_ITER_CONT; /* Return value */
- FUNC_ENTER_NOAPI(H5D_flush, FAIL)
+ FUNC_ENTER_NOAPI_NOINIT(H5D_flush_cb)
/* Check args */
- HDassert(f);
-
- /* Retrieve the # of outstanding dataset IDs */
- num_dsets = H5F_get_obj_count(f, H5F_OBJ_DATASET);
-
- /* Check for something to do */
- if(num_dsets > 0) {
- H5D_t *dataset; /* Dataset pointer */
- unsigned u; /* Index variable */
-
- H5_CHECK_OVERFLOW(num_dsets, unsigned, size_t);
- if(NULL == (id_list = (hid_t *)H5MM_malloc((size_t)num_dsets * sizeof(hid_t))))
- HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to allocate memory for ID list")
- if(H5F_get_obj_ids(f, H5F_OBJ_DATASET, -1, id_list) != num_dsets)
- HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to get dataset ID list")
+ HDassert(dataset);
- /* Iterate through the open datasets */
- for(u = 0; u < num_dsets; u++) {
- if(NULL == (dataset = (H5D_t *)H5I_object_verify(id_list[u], H5I_DATASET)))
- HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to get dataset object")
+ /* Check for dataset in same file */
+ if(udata->f == dataset->oloc.file) {
+ /* Update the dataspace on disk, if it's been changed */
+ if(dataset->shared->space_dirty) {
+ if(H5S_write(&(dataset->oloc), dataset->shared->space, TRUE, udata->dxpl_id) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace")
+ dataset->shared->space_dirty = FALSE;
+ } /* end if */
- /* Flush the raw data buffer, if we have a dirty one */
- if(dataset->shared->cache.contig.sieve_buf && dataset->shared->cache.contig.sieve_dirty) {
- HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
+ /* Flush the raw data buffer, if we have a dirty one */
+ if(dataset->shared->cache.contig.sieve_buf && dataset->shared->cache.contig.sieve_dirty) {
+ HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
- /* Write dirty data sieve buffer to file */
- if(H5F_block_write(f, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc,
- dataset->shared->cache.contig.sieve_size, dxpl_id, dataset->shared->cache.contig.sieve_buf) < 0)
- HGOTO_ERROR(H5E_IO, H5E_WRITEERROR, FAIL, "block write failed")
+ /* Write dirty data sieve buffer to file */
+ if(H5F_block_write(udata->f, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc,
+ dataset->shared->cache.contig.sieve_size, udata->dxpl_id, dataset->shared->cache.contig.sieve_buf) < 0)
+ HGOTO_ERROR(H5E_IO, H5E_WRITEERROR, FAIL, "block write failed")
- /* Reset sieve buffer dirty flag */
- dataset->shared->cache.contig.sieve_dirty = FALSE;
- } /* end if */
+ /* Reset sieve buffer dirty flag */
+ dataset->shared->cache.contig.sieve_dirty = FALSE;
+ } /* end if */
- /* Flush cached information for each kind of dataset */
- switch(dataset->shared->layout.type) {
- case H5D_CONTIGUOUS:
- break;
+ /* Flush cached information for each kind of dataset */
+ switch(dataset->shared->layout.type) {
+ case H5D_CONTIGUOUS:
+ break;
- case H5D_CHUNKED:
- /* Flush the raw data cache */
- if(H5D_istore_flush(dataset, dxpl_id, flags & H5F_FLUSH_INVALIDATE) < 0)
- HGOTO_ERROR(H5E_CACHE, H5E_CANTFLUSH, FAIL, "unable to flush raw data cache")
- break;
+ case H5D_CHUNKED:
+ /* Flush the raw data cache */
+ if(H5D_istore_flush(dataset, udata->dxpl_id, udata->flags & H5F_FLUSH_INVALIDATE) < 0)
+ HGOTO_ERROR(H5E_CACHE, H5E_CANTFLUSH, FAIL, "unable to flush raw data cache")
+ break;
- case H5D_COMPACT:
- if(dataset->shared->layout.u.compact.dirty) {
- if(H5O_msg_write(&(dataset->oloc), H5O_LAYOUT_ID, 0, H5O_UPDATE_TIME, &(dataset->shared->layout), dxpl_id) < 0)
- HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to update layout message")
- dataset->shared->layout.u.compact.dirty = FALSE;
- } /* end if */
- break;
+ case H5D_COMPACT:
+ if(dataset->shared->layout.u.compact.dirty) {
+ if(H5O_msg_write(&(dataset->oloc), H5O_LAYOUT_ID, 0, H5O_UPDATE_TIME, &(dataset->shared->layout), udata->dxpl_id) < 0)
+ HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to update layout message")
+ dataset->shared->layout.u.compact.dirty = FALSE;
+ } /* end if */
+ break;
- default:
- HDassert("not implemented yet" && 0);
+ default:
+ HDassert("not implemented yet" && 0);
#ifdef NDEBUG
- HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
+ HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
#endif /* NDEBUG */
- } /* end switch */ /*lint !e788 All appropriate cases are covered */
- } /* end for */
+ } /* end switch */ /*lint !e788 All appropriate cases are covered */
} /* end if */
done:
- if(id_list)
- H5MM_xfree(id_list);
+ FUNC_LEAVE_NOAPI(ret_value)
+} /* end H5D_flush_cb() */
+
+
+/*-------------------------------------------------------------------------
+ * Function: H5D_flush
+ *
+ * Purpose: Flush any dataset information cached in memory
+ *
+ * Return: Success: Non-negative
+ * Failure: Negative
+ *
+ * Programmer: Ray Lu
+ * August 14, 2002
+ *
+ *-------------------------------------------------------------------------
+ */
+herr_t
+H5D_flush(const H5F_t *f, hid_t dxpl_id, unsigned flags)
+{
+ H5D_flush_ud_t udata; /* User data for callback */
+ herr_t ret_value = SUCCEED; /* Return value */
+
+ FUNC_ENTER_NOAPI(H5D_flush, FAIL)
+
+ /* Check args */
+ HDassert(f);
+
+ /* Set user data for callback */
+ udata.f = f;
+ udata.dxpl_id = dxpl_id;
+ udata.flags = flags;
+
+ /* Iterate over all the open datasets */
+ H5I_search(H5I_DATASET, H5D_flush_cb, &udata);
+
+done:
FUNC_LEAVE_NOAPI(ret_value)
} /* end H5D_flush() */
diff --git a/src/H5Dpkg.h b/src/H5Dpkg.h
index cdcc362..34a32ee 100644
--- a/src/H5Dpkg.h
+++ b/src/H5Dpkg.h
@@ -154,6 +154,7 @@ typedef struct H5D_shared_t {
hid_t type_id; /* ID for dataset's datatype */
H5T_t *type; /* datatype of this dataset */
H5S_t *space; /* dataspace of this dataset */
+ hbool_t space_dirty; /* Whether the dataspace info needs to be flushed to the file */
hid_t dcpl_id; /* dataset creation property id */
H5D_dcpl_cache_t dcpl_cache; /* Cached DCPL values */
H5D_io_ops_t io_ops; /* I/O operations */