diff options
author | Quincey Koziol <koziol@hdfgroup.org> | 2007-11-10 14:51:30 (GMT) |
---|---|---|
committer | Quincey Koziol <koziol@hdfgroup.org> | 2007-11-10 14:51:30 (GMT) |
commit | 9601db2b2b4ec2550b30ae7156ee53be3d6d02b2 (patch) | |
tree | 72cc45a80d99a5be7eea01b2796f7a85cc473a14 /src | |
parent | 1dc74b6177e2b5b7245a3146d13876954e8f80f9 (diff) | |
download | hdf5-9601db2b2b4ec2550b30ae7156ee53be3d6d02b2.zip hdf5-9601db2b2b4ec2550b30ae7156ee53be3d6d02b2.tar.gz hdf5-9601db2b2b4ec2550b30ae7156ee53be3d6d02b2.tar.bz2 |
[svn-r14248] synced fast_append branch with trunk (bring over general dataset append record speed improvements)
Diffstat (limited to 'src')
-rw-r--r-- | src/H5Dint.c | 165 | ||||
-rw-r--r-- | src/H5Dpkg.h | 1 |
2 files changed, 102 insertions, 64 deletions
diff --git a/src/H5Dint.c b/src/H5Dint.c index 5584187..aebc6f2 100644 --- a/src/H5Dint.c +++ b/src/H5Dint.c @@ -45,6 +45,13 @@ /* Local Typedefs */ /******************/ +/* Struct for holding callback info during H5D_flush operation */ +typedef struct { + const H5F_t *f; /* Pointer to file being flushed */ + hid_t dxpl_id; /* DXPL for I/O operations */ + unsigned flags; /* Flags for flush operation */ +} H5D_flush_ud_t; + /********************/ /* Local Prototypes */ @@ -1484,6 +1491,13 @@ H5D_close(H5D_t *dataset) dataset->shared->fo_count--; if(dataset->shared->fo_count == 0) { + /* Update the dataspace on disk, if it's been changed */ + if(dataset->shared->space_dirty) { + if(H5S_write(&(dataset->oloc), dataset->shared->space, TRUE, H5AC_dxpl_id) < 0) + HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace") + dataset->shared->space_dirty = FALSE; + } /* end if */ + /* Free the data sieve buffer, if it's been allocated */ if(dataset->shared->cache.contig.sieve_buf) { HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */ @@ -2284,9 +2298,8 @@ H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id) * Modify the dataset storage *------------------------------------------------------------------------- */ - /* Save the new dataspace in the file if necessary */ - if(H5S_write(&(dset->oloc), space, TRUE, dxpl_id) < 0) - HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace") + /* Mark the dataspace as dirty, for later writing to the file */ + dset->shared->space_dirty = TRUE; /* Update the index values for the cached chunks for this dataset */ if(H5D_CHUNKED == dset->shared->layout.type) @@ -2332,93 +2345,117 @@ done: /*------------------------------------------------------------------------- - * Function: H5D_flush + * Function: H5D_flush_cb * * Purpose: Flush any dataset information cached in memory * * Return: Success: Non-negative * Failure: Negative * - * Programmer: Ray Lu - * August 14, 2002 + * Programmer: Quincey Koziol + * November 8, 2007 * *------------------------------------------------------------------------- */ -herr_t -H5D_flush(const H5F_t *f, hid_t dxpl_id, unsigned flags) +static int +H5D_flush_cb(void *_dataset, hid_t UNUSED id, void *_udata) { - hid_t *id_list = NULL; /* Array of dataset IDs */ - unsigned num_dsets; /* Number of datasets in file */ - herr_t ret_value = SUCCEED; /* Return value */ + H5D_t *dataset = (H5D_t *)_dataset; /* Dataset pointer */ + H5D_flush_ud_t *udata = (H5D_flush_ud_t *)_udata; /* User data for callback */ + int ret_value = H5_ITER_CONT; /* Return value */ - FUNC_ENTER_NOAPI(H5D_flush, FAIL) + FUNC_ENTER_NOAPI_NOINIT(H5D_flush_cb) /* Check args */ - HDassert(f); - - /* Retrieve the # of outstanding dataset IDs */ - num_dsets = H5F_get_obj_count(f, H5F_OBJ_DATASET); - - /* Check for something to do */ - if(num_dsets > 0) { - H5D_t *dataset; /* Dataset pointer */ - unsigned u; /* Index variable */ - - H5_CHECK_OVERFLOW(num_dsets, unsigned, size_t); - if(NULL == (id_list = (hid_t *)H5MM_malloc((size_t)num_dsets * sizeof(hid_t)))) - HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to allocate memory for ID list") - if(H5F_get_obj_ids(f, H5F_OBJ_DATASET, -1, id_list) != num_dsets) - HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to get dataset ID list") + HDassert(dataset); - /* Iterate through the open datasets */ - for(u = 0; u < num_dsets; u++) { - if(NULL == (dataset = (H5D_t *)H5I_object_verify(id_list[u], H5I_DATASET))) - HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to get dataset object") + /* Check for dataset in same file */ + if(udata->f == dataset->oloc.file) { + /* Update the dataspace on disk, if it's been changed */ + if(dataset->shared->space_dirty) { + if(H5S_write(&(dataset->oloc), dataset->shared->space, TRUE, udata->dxpl_id) < 0) + HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace") + dataset->shared->space_dirty = FALSE; + } /* end if */ - /* Flush the raw data buffer, if we have a dirty one */ - if(dataset->shared->cache.contig.sieve_buf && dataset->shared->cache.contig.sieve_dirty) { - HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */ + /* Flush the raw data buffer, if we have a dirty one */ + if(dataset->shared->cache.contig.sieve_buf && dataset->shared->cache.contig.sieve_dirty) { + HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */ - /* Write dirty data sieve buffer to file */ - if(H5F_block_write(f, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc, - dataset->shared->cache.contig.sieve_size, dxpl_id, dataset->shared->cache.contig.sieve_buf) < 0) - HGOTO_ERROR(H5E_IO, H5E_WRITEERROR, FAIL, "block write failed") + /* Write dirty data sieve buffer to file */ + if(H5F_block_write(udata->f, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc, + dataset->shared->cache.contig.sieve_size, udata->dxpl_id, dataset->shared->cache.contig.sieve_buf) < 0) + HGOTO_ERROR(H5E_IO, H5E_WRITEERROR, FAIL, "block write failed") - /* Reset sieve buffer dirty flag */ - dataset->shared->cache.contig.sieve_dirty = FALSE; - } /* end if */ + /* Reset sieve buffer dirty flag */ + dataset->shared->cache.contig.sieve_dirty = FALSE; + } /* end if */ - /* Flush cached information for each kind of dataset */ - switch(dataset->shared->layout.type) { - case H5D_CONTIGUOUS: - break; + /* Flush cached information for each kind of dataset */ + switch(dataset->shared->layout.type) { + case H5D_CONTIGUOUS: + break; - case H5D_CHUNKED: - /* Flush the raw data cache */ - if(H5D_istore_flush(dataset, dxpl_id, flags & H5F_FLUSH_INVALIDATE) < 0) - HGOTO_ERROR(H5E_CACHE, H5E_CANTFLUSH, FAIL, "unable to flush raw data cache") - break; + case H5D_CHUNKED: + /* Flush the raw data cache */ + if(H5D_istore_flush(dataset, udata->dxpl_id, udata->flags & H5F_FLUSH_INVALIDATE) < 0) + HGOTO_ERROR(H5E_CACHE, H5E_CANTFLUSH, FAIL, "unable to flush raw data cache") + break; - case H5D_COMPACT: - if(dataset->shared->layout.u.compact.dirty) { - if(H5O_msg_write(&(dataset->oloc), H5O_LAYOUT_ID, 0, H5O_UPDATE_TIME, &(dataset->shared->layout), dxpl_id) < 0) - HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to update layout message") - dataset->shared->layout.u.compact.dirty = FALSE; - } /* end if */ - break; + case H5D_COMPACT: + if(dataset->shared->layout.u.compact.dirty) { + if(H5O_msg_write(&(dataset->oloc), H5O_LAYOUT_ID, 0, H5O_UPDATE_TIME, &(dataset->shared->layout), udata->dxpl_id) < 0) + HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to update layout message") + dataset->shared->layout.u.compact.dirty = FALSE; + } /* end if */ + break; - default: - HDassert("not implemented yet" && 0); + default: + HDassert("not implemented yet" && 0); #ifdef NDEBUG - HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout") + HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout") #endif /* NDEBUG */ - } /* end switch */ /*lint !e788 All appropriate cases are covered */ - } /* end for */ + } /* end switch */ /*lint !e788 All appropriate cases are covered */ } /* end if */ done: - if(id_list) - H5MM_xfree(id_list); + FUNC_LEAVE_NOAPI(ret_value) +} /* end H5D_flush_cb() */ + + +/*------------------------------------------------------------------------- + * Function: H5D_flush + * + * Purpose: Flush any dataset information cached in memory + * + * Return: Success: Non-negative + * Failure: Negative + * + * Programmer: Ray Lu + * August 14, 2002 + * + *------------------------------------------------------------------------- + */ +herr_t +H5D_flush(const H5F_t *f, hid_t dxpl_id, unsigned flags) +{ + H5D_flush_ud_t udata; /* User data for callback */ + herr_t ret_value = SUCCEED; /* Return value */ + + FUNC_ENTER_NOAPI(H5D_flush, FAIL) + + /* Check args */ + HDassert(f); + + /* Set user data for callback */ + udata.f = f; + udata.dxpl_id = dxpl_id; + udata.flags = flags; + + /* Iterate over all the open datasets */ + H5I_search(H5I_DATASET, H5D_flush_cb, &udata); + +done: FUNC_LEAVE_NOAPI(ret_value) } /* end H5D_flush() */ diff --git a/src/H5Dpkg.h b/src/H5Dpkg.h index cdcc362..34a32ee 100644 --- a/src/H5Dpkg.h +++ b/src/H5Dpkg.h @@ -154,6 +154,7 @@ typedef struct H5D_shared_t { hid_t type_id; /* ID for dataset's datatype */ H5T_t *type; /* datatype of this dataset */ H5S_t *space; /* dataspace of this dataset */ + hbool_t space_dirty; /* Whether the dataspace info needs to be flushed to the file */ hid_t dcpl_id; /* dataset creation property id */ H5D_dcpl_cache_t dcpl_cache; /* Cached DCPL values */ H5D_io_ops_t io_ops; /* I/O operations */ |