summaryrefslogtreecommitdiffstats
path: root/src/H5D.c
diff options
context:
space:
mode:
Diffstat (limited to 'src/H5D.c')
-rw-r--r--src/H5D.c1159
1 files changed, 497 insertions, 662 deletions
diff --git a/src/H5D.c b/src/H5D.c
index a0304a6..7f1ee27 100644
--- a/src/H5D.c
+++ b/src/H5D.c
@@ -170,25 +170,17 @@ H5D_init_interface(void)
if(NULL == (def_dcpl = H5I_object(H5P_LST_DATASET_CREATE_g)))
HGOTO_ERROR(H5E_DATASET, H5E_BADTYPE, FAIL, "can't get default dataset creation property list")
- /* Set up the default allocation time information */
- if(H5P_get(def_dcpl, H5D_CRT_ALLOC_TIME_NAME, &H5D_def_dset.alloc_time) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve space allocation time")
-
- /* Get the default external file list information */
- if(H5P_get(def_dcpl, H5D_CRT_EXT_FILE_LIST_NAME, &H5D_def_dset.efl) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve external file list")
-
/* Get the default data storage method */
if(H5P_get(def_dcpl, H5D_CRT_LAYOUT_NAME, &H5D_def_dset.layout.type) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve layout")
- /* Get the default fill value time */
- if(H5P_get(def_dcpl, H5D_CRT_FILL_TIME_NAME, &H5D_def_dset.fill_time) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve fill time")
-
- /* Get the default fill value */
- if(H5P_get(def_dcpl, H5D_CRT_FILL_VALUE_NAME, &H5D_def_dset.fill) < 0)
+ /* Get the default dataset creation properties */
+ if(H5P_get(def_dcpl, H5D_CRT_EXT_FILE_LIST_NAME, &H5D_def_dset.dcpl_cache.efl) < 0)
+ HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve external file list")
+ if(H5P_get(def_dcpl, H5D_CRT_FILL_VALUE_NAME, &H5D_def_dset.dcpl_cache.fill) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve fill value")
+ if(H5P_get(def_dcpl, H5D_CRT_DATA_PIPELINE_NAME, &H5D_def_dset.dcpl_cache.pline) < 0)
+ HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve pipeline filter")
/* Reset the "default DXPL cache" information */
HDmemset(&H5D_def_dxpl_cache, 0, sizeof(H5D_dxpl_cache_t));
@@ -224,8 +216,8 @@ H5D_term_interface(void)
FUNC_ENTER_NOAPI_NOINIT_NOFUNC(H5D_term_interface)
- if (H5_interface_initialize_g) {
- if ((n=H5I_nmembers(H5I_DATASET))>0) {
+ if(H5_interface_initialize_g) {
+ if((n=H5I_nmembers(H5I_DATASET))>0) {
/* The dataset API uses the "force" flag set to true because it
* is using the "file objects" (H5FO) API functions to track open
* objects in the file. Using the H5FO code means that dataset
@@ -256,58 +248,7 @@ H5D_term_interface(void)
}
}
FUNC_LEAVE_NOAPI(n)
-}
-
-
-/*--------------------------------------------------------------------------
- NAME
- H5D_get_dcpl_cache
- PURPOSE
- Get all the values for the DCPL cache.
- USAGE
- herr_t H5D_get_dcpl_cache(dcpl_id, cache)
- hid_t dcpl_id; IN: DCPL to query
- H5D_dcpl_cache_t *cache;IN/OUT: DCPL cache to fill with values
- RETURNS
- Non-negative on success/Negative on failure.
- DESCRIPTION
- Query all the values from a DCPL that are needed by internal routines
- within the library.
- GLOBAL VARIABLES
- COMMENTS, BUGS, ASSUMPTIONS
- EXAMPLES
- REVISION LOG
---------------------------------------------------------------------------*/
-static herr_t
-H5D_get_dcpl_cache(hid_t dcpl_id, H5D_dcpl_cache_t *cache)
-{
- H5P_genplist_t *dc_plist; /* Data transfer property list */
- herr_t ret_value = SUCCEED; /* Return value */
-
- FUNC_ENTER_NOAPI_NOINIT(H5D_get_dcpl_cache)
-
- /* Check args */
- HDassert(cache);
-
- /* Get the dataset transfer property list */
- if (NULL == (dc_plist = H5I_object(dcpl_id)))
- HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset creation property list")
-
- /* Get I/O pipeline info */
- if(H5P_get(dc_plist, H5D_CRT_DATA_PIPELINE_NAME, &cache->pline) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "Can't retrieve I/O pipeline info")
-
- /* Get fill value info */
- if(H5P_get(dc_plist, H5D_CRT_FILL_VALUE_NAME, &cache->fill) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "Can't retrieve fill value info")
-
- /* Get fill time info */
- if(H5P_get(dc_plist, H5D_CRT_FILL_TIME_NAME, &cache->fill_time) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "Can't retrieve fill time")
-
-done:
- FUNC_LEAVE_NOAPI(ret_value)
-} /* H5D_get_dcpl_cache() */
+} /* end H5D_term_interface() */
/*-------------------------------------------------------------------------
@@ -390,7 +331,7 @@ H5Dcreate(hid_t loc_id, const char *name, hid_t type_id, hid_t space_id,
if((dset_id = H5I_register(H5I_DATASET, new_dset)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTREGISTER, FAIL, "unable to register dataset")
- if(H5G_loc(dset_id, &dset_loc) <0)
+ if(H5G_loc(dset_id, &dset_loc) < 0)
HGOTO_ERROR(H5E_SYM, H5E_NOTFOUND, FAIL, "unable to get location for dataset")
/* Link the new dataset */
@@ -661,20 +602,16 @@ H5Dopen_expand(hid_t loc_id, const char *name, hid_t dapl_id)
done:
if(ret_value < 0) {
- if(dset != NULL) {
- if(H5D_close(dset) < 0)
- HDONE_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release dataset")
- } /* end if */
+ if(dset != NULL && H5D_close(dset) < 0)
+ HDONE_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release dataset")
else {
- if(loc_found) {
- if(H5G_loc_free(&dset_loc) < 0)
- HDONE_ERROR(H5E_SYM, H5E_CANTRELEASE, FAIL, "can't free location")
- }
+ if(loc_found && H5G_loc_free(&dset_loc) < 0)
+ HDONE_ERROR(H5E_SYM, H5E_CANTRELEASE, FAIL, "can't free location")
} /* end else */
} /* end if */
FUNC_LEAVE_API(ret_value)
-}
+} /* end H5Dopen_expand() */
/*-------------------------------------------------------------------------
@@ -743,21 +680,21 @@ H5Dget_space(hid_t dset_id)
H5TRACE1("i", "i", dset_id);
/* Check args */
- if (NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
- HGOTO_ERROR (H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
+ if(NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
+ HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
/* Read the data space message and return a data space object */
- if (NULL==(space=H5S_copy (dset->shared->space, FALSE)))
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, FAIL, "unable to get data space")
+ if(NULL==(space=H5S_copy (dset->shared->space, FALSE)))
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to get data space")
/* Create an atom */
- if ((ret_value=H5I_register (H5I_DATASPACE, space))<0)
- HGOTO_ERROR (H5E_ATOM, H5E_CANTREGISTER, FAIL, "unable to register data space")
+ if((ret_value=H5I_register (H5I_DATASPACE, space)) < 0)
+ HGOTO_ERROR(H5E_ATOM, H5E_CANTREGISTER, FAIL, "unable to register data space")
done:
- if(ret_value<0) {
+ if(ret_value < 0) {
if(space!=NULL) {
- if(H5S_close(space)<0)
+ if(H5S_close(space) < 0)
HDONE_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release dataspace")
} /* end if */
} /* end if */
@@ -791,11 +728,11 @@ H5Dget_space_status(hid_t dset_id, H5D_space_status_t *allocation)
/* Check arguments */
if(NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
- HGOTO_ERROR (H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
+ HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
/* Read data space address and return */
if(FAIL==(ret_value=H5D_get_space_status(dset, allocation, H5AC_ind_dxpl_id)))
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, FAIL, "unable to get space status")
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to get space status")
done:
FUNC_LEAVE_API(ret_value)
@@ -828,18 +765,18 @@ H5D_get_space_status(H5D_t *dset, H5D_space_status_t *allocation, hid_t dxpl_id)
FUNC_ENTER_NOAPI_NOINIT(H5D_get_space_status)
- assert(dset);
+ HDassert(dset);
/* Get the dataset's dataspace */
- space=dset->shared->space;
- assert(space);
+ space = dset->shared->space;
+ HDassert(space);
/* Get the total number of elements in dataset's dataspace */
- if((total_elem=H5S_GET_EXTENT_NPOINTS(space))<0)
+ if((total_elem=H5S_GET_EXTENT_NPOINTS(space)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTCOUNT, FAIL, "unable to get # of dataspace elements")
/* Get the size of the dataset's datatype */
- if((type_size=H5T_get_size(dset->shared->type))==0)
+ if((type_size = H5T_get_size(dset->shared->type)) == 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTCOUNT, FAIL, "unable to get size of datatype")
/* Compute the maximum size of the dataset in bytes */
@@ -847,23 +784,23 @@ H5D_get_space_status(H5D_t *dset, H5D_space_status_t *allocation, hid_t dxpl_id)
full_size=((hsize_t)total_elem)*type_size;
/* Difficult to error check, since the error value is 0 and 0 is a valid value... :-/ */
- space_allocated=H5D_get_storage_size(dset,dxpl_id);
+ space_allocated = H5D_get_storage_size(dset, dxpl_id);
/* Decide on how much of the space is allocated */
- if(space_allocated==0)
+ if(space_allocated == 0)
*allocation = H5D_SPACE_STATUS_NOT_ALLOCATED;
- else if(space_allocated==full_size)
+ else if(space_allocated == full_size)
*allocation = H5D_SPACE_STATUS_ALLOCATED;
else {
/* Should only happen for chunked datasets currently */
- assert(dset->shared->layout.type==H5D_CHUNKED);
+ HDassert(dset->shared->layout.type == H5D_CHUNKED);
*allocation = H5D_SPACE_STATUS_PART_ALLOCATED;
} /* end else */
done:
FUNC_LEAVE_NOAPI(ret_value)
-}
+} /* end H5D_get_space_status() */
/*-------------------------------------------------------------------------
@@ -894,29 +831,29 @@ H5Dget_type(hid_t dset_id)
H5TRACE1("i", "i", dset_id);
/* Check args */
- if (NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
- HGOTO_ERROR (H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
+ if(NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
+ HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
/* Copy the datatype and mark it read-only */
- if (NULL==(copied_type=H5T_copy (dset->shared->type, H5T_COPY_REOPEN)))
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, FAIL, "unable to copy the datatype")
+ if(NULL==(copied_type=H5T_copy (dset->shared->type, H5T_COPY_REOPEN)))
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to copy the datatype")
/* Mark any datatypes as being in memory now */
- if (H5T_set_loc(copied_type, NULL, H5T_LOC_MEMORY)<0)
+ if(H5T_set_loc(copied_type, NULL, H5T_LOC_MEMORY) < 0)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "invalid datatype location")
/* Unlock copied type */
- if (H5T_lock (copied_type, FALSE)<0)
- HGOTO_ERROR (H5E_DATATYPE, H5E_CANTINIT, FAIL, "unable to lock transient datatype")
+ if(H5T_lock (copied_type, FALSE) < 0)
+ HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "unable to lock transient datatype")
/* Create an atom */
- if ((ret_value=H5I_register (H5I_DATATYPE, copied_type))<0)
- HGOTO_ERROR (H5E_ATOM, H5E_CANTREGISTER, FAIL, "unable to register datatype")
+ if((ret_value=H5I_register (H5I_DATATYPE, copied_type)) < 0)
+ HGOTO_ERROR(H5E_ATOM, H5E_CANTREGISTER, FAIL, "unable to register datatype")
done:
- if(ret_value<0) {
+ if(ret_value < 0) {
if(copied_type!=NULL) {
- if(H5T_close (copied_type)<0)
+ if(H5T_close (copied_type) < 0)
HDONE_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release datatype")
} /* end if */
} /* end if */
@@ -944,26 +881,26 @@ done:
hid_t
H5Dget_create_plist(hid_t dset_id)
{
- H5D_t *dset = NULL;
- H5O_fill_t copied_fill={NULL,0,NULL};
- H5P_genplist_t *dcpl_plist;
- H5P_genplist_t *new_plist;
+ H5D_t *dset; /* Dataset structure */
+ H5P_genplist_t *dcpl_plist; /* Dataset's DCPL */
+ H5P_genplist_t *new_plist; /* Copy of dataset's DCPL */
+ H5O_fill_t copied_fill; /* Fill value to tweak */
hid_t new_dcpl_id = FAIL;
- hid_t ret_value = FAIL;
+ hid_t ret_value; /* Return value */
FUNC_ENTER_API(H5Dget_create_plist, FAIL)
H5TRACE1("i", "i", dset_id);
/* Check args */
- if (NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
- HGOTO_ERROR (H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
- if (NULL == (dcpl_plist = H5I_object(dset->shared->dcpl_id)))
+ if(NULL == (dset = H5I_object_verify(dset_id, H5I_DATASET)))
+ HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
+ if(NULL == (dcpl_plist = H5I_object(dset->shared->dcpl_id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "can't get property list")
/* Copy the creation property list */
if((new_dcpl_id = H5P_copy_plist(dcpl_plist)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "unable to copy the creation property list")
- if (NULL == (new_plist = H5I_object(new_dcpl_id)))
+ if(NULL == (new_plist = H5I_object(new_dcpl_id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "can't get property list")
/* Retrieve any object creation properties */
@@ -974,9 +911,9 @@ H5Dget_create_plist(hid_t dset_id)
if(H5P_get(new_plist, H5D_CRT_FILL_VALUE_NAME, &copied_fill) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get fill value")
- /* Copy the dataset type into the fill value message */
- if(copied_fill.type==NULL)
- if(NULL==(copied_fill.type=H5T_copy(dset->shared->type, H5T_COPY_TRANSIENT)))
+ /* Copy the dataset type into the fill value message, if there's actually a fill value */
+ if(copied_fill.buf != NULL && copied_fill.type == NULL)
+ if(NULL == (copied_fill.type = H5T_copy(dset->shared->type, H5T_COPY_TRANSIENT)))
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to copy dataset datatype for fill value")
/* Set back the fill value property to property list */
@@ -984,16 +921,15 @@ H5Dget_create_plist(hid_t dset_id)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "unable to set property list fill value")
/* Set the return value */
- ret_value=new_dcpl_id;
+ ret_value = new_dcpl_id;
done:
- if(ret_value<0) {
- if(new_dcpl_id>0)
+ if(ret_value < 0)
+ if(new_dcpl_id > 0)
(void)H5I_dec_ref(new_dcpl_id);
- } /* end if */
FUNC_LEAVE_API(ret_value)
-}
+} /* end H5Dget_create_plist() */
/*-------------------------------------------------------------------------
@@ -1013,25 +949,25 @@ done:
herr_t
H5Dextend(hid_t dset_id, const hsize_t *size)
{
- H5D_t *dset = NULL;
- herr_t ret_value=SUCCEED; /* Return value */
+ H5D_t *dset;
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_API(H5Dextend, FAIL)
H5TRACE2("e", "i*h", dset_id, size);
/* Check args */
- if (NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
+ if(NULL == (dset = H5I_object_verify(dset_id, H5I_DATASET)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset")
- if (!size)
- HGOTO_ERROR (H5E_ARGS, H5E_BADVALUE, FAIL, "no size specified")
+ if(!size)
+ HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "no size specified")
/* Increase size */
- if (H5D_extend (dset, size, H5AC_dxpl_id)<0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, FAIL, "unable to extend dataset")
+ if(H5D_extend(dset, size, H5AC_dxpl_id) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to extend dataset")
done:
FUNC_LEAVE_API(ret_value)
-}
+} /* end H5Dextend() */
/*-------------------------------------------------------------------------
@@ -1059,23 +995,20 @@ H5D_new(hid_t dcpl_id, hbool_t creating, hbool_t vl_type)
/* Allocate new shared dataset structure */
if(NULL == (new_dset = H5FL_MALLOC(H5D_shared_t)))
- HGOTO_ERROR (H5E_RESOURCE, H5E_NOSPACE, NULL, "memory allocation failed")
+ HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, NULL, "memory allocation failed")
+
+ /* Copy the default dataset information */
+ HDmemcpy(new_dset, &H5D_def_dset, sizeof(H5D_shared_t));
/* If we are using the default dataset creation property list, during creation
* don't bother to copy it, just increment the reference count
*/
if(!vl_type && creating && dcpl_id == H5P_DATASET_CREATE_DEFAULT) {
- /* Copy the default dataset information */
- HDmemcpy(new_dset, &H5D_def_dset, sizeof(H5D_shared_t));
-
if(H5I_inc_ref(dcpl_id) < 0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINC, NULL, "Can't increment default DCPL ID")
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINC, NULL, "Can't increment default DCPL ID")
new_dset->dcpl_id = dcpl_id;
} /* end if */
else {
- /* Reset the dataset information */
- HDmemset(new_dset, 0, sizeof(H5D_shared_t));
-
/* Get the property list */
if(NULL == (plist = H5I_object(dcpl_id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, NULL, "not a property list")
@@ -1122,16 +1055,16 @@ H5D_init_type(H5F_t *file, const H5D_t *dset, hid_t type_id, const H5T_t *type)
FUNC_ENTER_NOAPI(H5D_init_type, FAIL)
/* Sanity checking */
- assert(file);
- assert(dset);
- assert(type);
+ HDassert(file);
+ HDassert(dset);
+ HDassert(type);
/* Check whether the datatype is relocatable */
- if((relocatable=H5T_is_relocatable(type))<0)
+ if((relocatable = H5T_is_relocatable(type)) < 0)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "can't check datatype?")
/* Check whether the datatype is immutable */
- if((immutable=H5T_is_immutable(type))<0)
+ if((immutable = H5T_is_immutable(type)) < 0)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "can't check datatype?")
/* Copy the datatype if it's a custom datatype or if it'll change when it's location is changed */
@@ -1141,17 +1074,17 @@ H5D_init_type(H5F_t *file, const H5D_t *dset, hid_t type_id, const H5T_t *type)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTCOPY, FAIL, "can't copy datatype")
/* Mark any datatypes as being on disk now */
- if(H5T_set_loc(dset->shared->type, file, H5T_LOC_DISK)<0)
+ if(H5T_set_loc(dset->shared->type, file, H5T_LOC_DISK) < 0)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "invalid datatype location")
/* Get a datatype ID for the dataset's datatype */
- if((dset->shared->type_id = H5I_register(H5I_DATATYPE, dset->shared->type))<0)
+ if((dset->shared->type_id = H5I_register(H5I_DATATYPE, dset->shared->type)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTREGISTER, FAIL, "unable to register type")
} /* end if */
/* Not a custom datatype, just use it directly */
else {
- if(H5I_inc_ref(type_id)<0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINC, FAIL, "Can't increment datatype ID")
+ if(H5I_inc_ref(type_id) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINC, FAIL, "Can't increment datatype ID")
/* Use existing datatype */
dset->shared->type_id = type_id;
@@ -1164,13 +1097,9 @@ done:
/*-------------------------------------------------------------------------
- * Function: H5D_update_entry_info
- *
- * Purpose: Create and fill an H5G_entry_t object for insertion into
- * the group LOC.
+ * Function: H5D_update_oh_info
*
- * This code was originally found at the end of H5D_create()
- * but was placed here for general use.
+ * Purpose: Create and fill object header for dataset
*
* Return: Success: SUCCEED
* Failure: FAIL
@@ -1181,113 +1110,84 @@ done:
*-------------------------------------------------------------------------
*/
static herr_t
-H5D_update_entry_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
+H5D_update_oh_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
{
+ struct H5O_t *oh = NULL; /* Pointer to dataset's object header */
size_t ohdr_size = H5D_MINHDR_SIZE; /* Size of dataset's object header */
+ unsigned oh_flags = H5AC__DIRTIED_FLAG;
H5O_loc_t *oloc = NULL; /* Dataset's object location */
H5O_layout_t *layout; /* Dataset's layout information */
H5T_t *type; /* Dataset's datatype */
- H5D_alloc_time_t alloc_time; /* Dataset's allocation time */
- H5O_efl_t *efl; /* Dataset's external file list */
- H5P_genplist_t *dc_plist = NULL; /* Dataset's creation property list */
hbool_t use_latest_format; /* Flag indicating the newest file format should be used */
-
- /* fill value variables */
- H5D_fill_time_t fill_time;
H5O_fill_t *fill_prop; /* Pointer to dataset's fill value information */
- H5O_fill_new_t fill; /* Default fill value */
- H5D_fill_value_t fill_status;
-
- struct H5O_t *oh = NULL; /* Pointer to dataset's object header */
- unsigned oh_flags = H5AC__DIRTIED_FLAG;
+ H5D_fill_value_t fill_status; /* Fill value status */
+ hbool_t fill_changed = FALSE; /* Flag indicating the fill value was changed */
herr_t ret_value = SUCCEED; /* Return value */
- FUNC_ENTER_NOAPI(H5D_update_entry_info, FAIL)
+ FUNC_ENTER_NOAPI(H5D_update_oh_info, FAIL)
/* Sanity checking */
HDassert(file);
HDassert(dset);
- /* Portably initialize the fill value */
- HDmemset(&fill, 0, sizeof(H5O_fill_new_t));
- fill.alloc_time = H5D_ALLOC_TIME_LATE;
- fill.fill_time = H5D_FILL_TIME_ALLOC;
- fill.fill_defined = TRUE;
-
- /* Set some location variables, for convenience */
+ /* Set some local variables, for convenience */
oloc = &dset->oloc;
layout = &dset->shared->layout;
type = dset->shared->type;
- alloc_time = dset->shared->alloc_time;
- efl = &dset->shared->efl;
+ fill_prop = &dset->shared->dcpl_cache.fill;
/* Get the file's 'use the latest version of the format' flag */
use_latest_format = H5F_USE_LATEST_FORMAT(file);
- /* Point at dataset's copy, to cache it for later */
- fill_prop = &dset->shared->fill;
- fill_time = dset->shared->fill_time;
-
- /* Check if dataset has non-default creation property list */
- if(dset->shared->dcpl_id != H5P_DATASET_CREATE_DEFAULT) {
- /* Get new dataset's property list object */
- if (NULL == (dc_plist = H5I_object(dset->shared->dcpl_id)))
- HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "can't get dataset creation property list")
-
- /*
- * Retrieve properties of fill value and others. Copy them into new fill
- * value struct.
- */
- if(H5P_get(dc_plist, H5D_CRT_FILL_TIME_NAME, &fill_time) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve fill time")
- dset->shared->fill_time=fill_time; /* Cache this for later */
-
- /* Get the fill value information from the property list */
- if(H5P_get(dc_plist, H5D_CRT_FILL_VALUE_NAME, fill_prop) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't retrieve fill value")
- } /* end if */
-
+ /* Retrieve "defined" status of fill value */
if(H5P_is_fill_value_defined(fill_prop, &fill_status) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't tell if fill value defined")
/* Special case handling for variable-length types */
if(H5T_detect_class(type, H5T_VLEN)) {
/* If the default fill value is chosen for variable-length types, always write it */
- if(fill_time == H5D_FILL_TIME_IFSET && fill_status == H5D_FILL_VALUE_DEFAULT) {
- dset->shared->fill_time = fill_time = H5D_FILL_TIME_ALLOC;
-
+ if(fill_prop->fill_time == H5D_FILL_TIME_IFSET && fill_status == H5D_FILL_VALUE_DEFAULT) {
/* Update dataset creation property */
- HDassert(dset->shared->dcpl_id != H5P_DATASET_CREATE_DEFAULT);
- if(H5P_set(dc_plist, H5D_CRT_FILL_TIME_NAME, &fill_time) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't set fill time")
+ fill_prop->fill_time = H5D_FILL_TIME_ALLOC;
+
+ /* Note that the fill value changed */
+ fill_changed = TRUE;
} /* end if */
/* Don't allow never writing fill values with variable-length types */
- if(fill_time == H5D_FILL_TIME_NEVER)
+ if(fill_prop->fill_time == H5D_FILL_TIME_NEVER)
HGOTO_ERROR(H5E_DATASET, H5E_UNSUPPORTED, FAIL, "Dataset doesn't support VL datatype when fill value is not defined")
} /* end if */
+ /* Determine whether fill value is defined or not */
if(fill_status == H5D_FILL_VALUE_DEFAULT || fill_status == H5D_FILL_VALUE_USER_DEFINED) {
- if(H5O_msg_copy(H5O_FILL_ID, fill_prop, &fill) == NULL)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT,FAIL, "unable to copy fill value")
-
- if(fill_prop->buf && fill_prop->size > 0 && H5O_fill_convert(&fill, type, dxpl_id) < 0)
+ /* Convert fill value buffer to dataset's datatype */
+ if(fill_prop->buf && fill_prop->size > 0 && H5O_fill_convert(fill_prop, type, &fill_changed, dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to convert fill value to dataset type")
- fill.fill_defined = TRUE;
- } else if (fill_status == H5D_FILL_VALUE_UNDEFINED) {
- fill.size = -1;
- fill.type = fill.buf = NULL;
- fill.fill_defined = FALSE;
+ fill_prop->fill_defined = TRUE;
+ } else if(fill_status == H5D_FILL_VALUE_UNDEFINED) {
+ fill_prop->fill_defined = FALSE;
} else
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "unable to determine if fill value is defined")
- fill.alloc_time = alloc_time;
- fill.fill_time = fill_time;
-
/* Check for invalid fill & allocation time setting */
- if(fill.fill_defined == FALSE && fill_time == H5D_FILL_TIME_ALLOC)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT,FAIL, "unable to create dataset")
+ if(fill_prop->fill_defined == FALSE && fill_prop->fill_time == H5D_FILL_TIME_ALLOC)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "fill value writing on allocation set, but no fill value defined")
+
+ /* Check if the fill value info changed */
+ if(fill_changed) {
+ H5P_genplist_t *dc_plist; /* Dataset's creation property list */
+
+ /* Get dataset's property list object */
+ HDassert(dset->shared->dcpl_id != H5P_DATASET_CREATE_DEFAULT);
+ if(NULL == (dc_plist = H5I_object(dset->shared->dcpl_id)))
+ HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "can't get dataset creation property list")
+
+ /* Update dataset creation property */
+ if(H5P_set(dc_plist, H5D_CRT_FILL_VALUE_NAME, fill_prop) < 0)
+ HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't set fill value info")
+ } /* end if */
/* Add the dataset's raw data size to the size of the header, if the raw data will be stored as compact */
if(layout->type == H5D_COMPACT)
@@ -1302,28 +1202,15 @@ H5D_update_entry_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to protect dataset object header")
/* Write new fill value message */
- if(H5O_msg_append(file, dxpl_id, oh, H5O_FILL_NEW_ID, H5O_MSG_FLAG_CONSTANT, 0, &fill, &oh_flags) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update fill value header message")
+ if(H5O_msg_append(file, dxpl_id, oh, H5O_FILL_NEW_ID, H5O_MSG_FLAG_CONSTANT, 0, fill_prop, &oh_flags) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update new fill value header message")
/* If there is valid information for the old fill value struct, update it */
/* (only if we aren't trying to write the latest version of the file format) */
- if(fill.buf && !use_latest_format) {
- /* Clear any previous values */
- if(H5O_msg_reset(H5O_FILL_ID, fill_prop)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTFREE, FAIL, "can't release fill info")
-
- /* Copy new fill value information to old fill value struct */
- if(H5O_msg_copy(H5O_FILL_ID, &fill, fill_prop) == NULL)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT,FAIL,"unable to copy fill value")
-
+ if(fill_prop->buf && !use_latest_format) {
/* Write old fill value */
- if(fill_prop->buf && H5O_msg_append(file, dxpl_id, oh, H5O_FILL_ID, H5O_MSG_FLAG_CONSTANT, 0, fill_prop, &oh_flags) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update fill value header message")
-
- /* Update dataset creation property */
- HDassert(dset->shared->dcpl_id!=H5P_DATASET_CREATE_DEFAULT);
- if(H5P_set(dc_plist, H5D_CRT_FILL_VALUE_NAME, fill_prop) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't set fill value")
+ if(H5O_msg_append(file, dxpl_id, oh, H5O_FILL_ID, H5O_MSG_FLAG_CONSTANT, 0, fill_prop, &oh_flags) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update old fill value header message")
} /* end if */
/* Update the type and space header messages */
@@ -1333,32 +1220,32 @@ H5D_update_entry_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
/* Update the filters message, if this is a chunked dataset */
if(layout->type == H5D_CHUNKED) {
- H5O_pline_t pline; /* Chunked data I/O pipeline info */
-
- if(H5P_get(dc_plist, H5D_CRT_DATA_PIPELINE_NAME, &pline) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "Can't retrieve pipeline filter")
+ H5O_pline_t *pline; /* Dataset's I/O pipeline information */
- if(pline.nused > 0 && H5O_msg_append(file, dxpl_id, oh, H5O_PLINE_ID, H5O_MSG_FLAG_CONSTANT, 0, &pline, &oh_flags) < 0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update filter header message")
+ pline = &dset->shared->dcpl_cache.pline;
+ if(pline->nused > 0 && H5O_msg_append(file, dxpl_id, oh, H5O_PLINE_ID, H5O_MSG_FLAG_CONSTANT, 0, pline, &oh_flags) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update filter header message")
} /* end if */
/*
* Allocate storage if space allocate time is early; otherwise delay
* allocation until later.
*/
- if (alloc_time == H5D_ALLOC_TIME_EARLY)
- if (H5D_alloc_storage(file, dxpl_id, dset, H5D_ALLOC_CREATE, FALSE, FALSE) < 0)
+ if(fill_prop->alloc_time == H5D_ALLOC_TIME_EARLY)
+ if(H5D_alloc_storage(file, dxpl_id, dset, H5D_ALLOC_CREATE, FALSE, FALSE) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to initialize storage")
- /* Update external storage message */
- if(efl->nused > 0) {
+ /* Update external storage message, if it's used */
+ if(dset->shared->dcpl_cache.efl.nused > 0) {
+ H5O_efl_t *efl = &dset->shared->dcpl_cache.efl; /* Dataset's external file list */
size_t heap_size = H5HL_ALIGN(1);
size_t u;
- for (u = 0; u < efl->nused; ++u)
+ /* Determine size of heap needed to stored the file names */
+ for(u = 0; u < efl->nused; ++u)
heap_size += H5HL_ALIGN(HDstrlen(efl->slot[u].name) + 1);
- if (H5HL_create(file, dxpl_id, heap_size, &efl->heap_addr/*out*/) < 0 ||
+ if(H5HL_create(file, dxpl_id, heap_size, &efl->heap_addr/*out*/) < 0 ||
H5HL_insert(file, dxpl_id, efl->heap_addr, (size_t)1, "") == (size_t)(-1))
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to create external file list name heap")
@@ -1366,9 +1253,9 @@ H5D_update_entry_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
size_t offset = H5HL_insert(file, dxpl_id, efl->heap_addr,
HDstrlen(efl->slot[u].name) + 1, efl->slot[u].name);
- assert(0 == efl->slot[u].name_offset);
+ HDassert(0 == efl->slot[u].name_offset);
- if (offset == (size_t)(-1))
+ if(offset == (size_t)(-1))
HGOTO_ERROR(H5E_EFL, H5E_CANTINIT, FAIL, "unable to insert URL into name heap")
efl->slot[u].name_offset = offset;
@@ -1380,8 +1267,8 @@ H5D_update_entry_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
/* Create layout message */
/* (Don't make layout message constant unless allocation time is early, since space may not be allocated) */
- /* Note: this is relying on H5D_alloc_storage not calling H5O_msg_write during dataset creation */
- if(H5O_msg_append(file, dxpl_id, oh, H5O_LAYOUT_ID, ((alloc_time == H5D_ALLOC_TIME_EARLY && H5D_COMPACT != layout->type) ? H5O_MSG_FLAG_CONSTANT : 0), 0, layout, &oh_flags) < 0)
+ /* (Note: this is relying on H5D_alloc_storage not calling H5O_msg_write during dataset creation) */
+ if(H5O_msg_append(file, dxpl_id, oh, H5O_LAYOUT_ID, ((fill_prop->alloc_time == H5D_ALLOC_TIME_EARLY && H5D_COMPACT != layout->type) ? H5O_MSG_FLAG_CONSTANT : 0), 0, layout, &oh_flags) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update layout")
#ifdef H5O_ENABLE_BOGUS
@@ -1397,17 +1284,13 @@ H5D_update_entry_info(H5F_t *file, hid_t dxpl_id, H5D_t *dset)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update modification time message")
done:
- /* Release fill value information */
- if(H5O_msg_reset(H5O_FILL_ID, &fill) < 0)
- HDONE_ERROR(H5E_DATASET, H5E_CANTRELEASE, FAIL, "unable to release fill-value info")
-
/* Release pointer to object header itself */
if(oloc != NULL && oh != NULL)
if(H5O_unprotect(oloc, oh, dxpl_id, oh_flags) < 0)
HDONE_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to unprotect dataset object header")
FUNC_LEAVE_NOAPI(ret_value)
-} /* end H5D_update_entry_info() */
+} /* end H5D_update_oh_info() */
/*-------------------------------------------------------------------------
@@ -1437,12 +1320,12 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
const H5T_t *type; /* Datatype for dataset */
H5D_t *new_dset = NULL;
int i, ndims;
- unsigned u;
unsigned chunk_ndims = 0; /* Dimensionality of chunk */
- H5P_genplist_t *dc_plist=NULL; /* New Property list */
- hbool_t has_vl_type=FALSE; /* Flag to indicate a VL-type for dataset */
- hbool_t chunk_init=FALSE; /* Flag to indicate that chunk information was initialized */
+ H5P_genplist_t *dc_plist = NULL; /* New Property list */
+ hbool_t has_vl_type = FALSE; /* Flag to indicate a VL-type for dataset */
+ hbool_t chunk_init = FALSE; /* Flag to indicate that chunk information was initialized */
H5G_loc_t dset_loc; /* Dataset location */
+ unsigned u; /* Local index variable */
H5D_t *ret_value; /* Return value */
FUNC_ENTER_NOAPI(H5D_create, NULL)
@@ -1455,11 +1338,11 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
HDassert(H5I_GENPROP_LST==H5I_get_type(dxpl_id));
/* Get the dataset's datatype */
- if (NULL == (type = H5I_object(type_id)))
+ if(NULL == (type = H5I_object(type_id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, NULL, "not a datatype")
/* Check if the datatype is "sensible" for use in a dataset */
- if(H5T_is_sensible(type)!=TRUE)
+ if(H5T_is_sensible(type) != TRUE)
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, NULL, "datatype is not sensible")
/* Check if the datatype is/contains a VL-type */
@@ -1467,12 +1350,12 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
has_vl_type=TRUE;
/* Check if the dataspace has an extent set (or is NULL) */
- if( !(H5S_has_extent(space)) )
+ if(!(H5S_has_extent(space)) )
HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, NULL, "dataspace extent has not been set.")
/* Initialize the dataset object */
- if (NULL==(new_dset = H5FL_CALLOC(H5D_t)))
- HGOTO_ERROR (H5E_RESOURCE, H5E_NOSPACE, NULL, "memory allocation failed")
+ if(NULL == (new_dset = H5FL_CALLOC(H5D_t)))
+ HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, NULL, "memory allocation failed")
/* Set up & reset dataset location */
dset_loc.oloc = &(new_dset->oloc);
@@ -1480,15 +1363,15 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
H5G_loc_reset(&dset_loc);
/* Initialize the shared dataset space */
- if(NULL == (new_dset->shared = H5D_new(dcpl_id,TRUE,has_vl_type)))
- HGOTO_ERROR (H5E_RESOURCE, H5E_NOSPACE, NULL, "memory allocation failed")
+ if(NULL == (new_dset->shared = H5D_new(dcpl_id, TRUE, has_vl_type)))
+ HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, NULL, "memory allocation failed")
/* Copy datatype for dataset */
- if(H5D_init_type(file, new_dset, type_id, type)<0)
+ if(H5D_init_type(file, new_dset, type_id, type) < 0)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTCOPY, NULL, "can't copy datatype")
/* Check if the filters in the DCPL can be applied to this dataset */
- if(H5Z_can_apply(new_dset->shared->dcpl_id,new_dset->shared->type_id)<0)
+ if(H5Z_can_apply(new_dset->shared->dcpl_id,new_dset->shared->type_id) < 0)
HGOTO_ERROR(H5E_ARGS, H5E_CANTINIT, NULL, "I/O filters can't operate on this dataset")
/* Set the dataset's checked_filters flag to enable writing */
@@ -1499,45 +1382,46 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTCOPY, NULL, "can't copy dataspace")
/* Set the dataset's dataspace to 'all' selection */
- if(H5S_select_all(new_dset->shared->space,1)<0)
- HGOTO_ERROR (H5E_DATASPACE, H5E_CANTSET, NULL, "unable to set all selection")
+ if(H5S_select_all(new_dset->shared->space,1) < 0)
+ HGOTO_ERROR(H5E_DATASPACE, H5E_CANTSET, NULL, "unable to set all selection")
/* Make the "set local" filter callbacks for this dataset */
- if(H5Z_set_local(new_dset->shared->dcpl_id,new_dset->shared->type_id)<0)
+ if(H5Z_set_local(new_dset->shared->dcpl_id,new_dset->shared->type_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "unable to set local filter parameters")
/* Check if the dataset has a non-default DCPL & get important values, if so */
- if(new_dset->shared->dcpl_id!=H5P_DATASET_CREATE_DEFAULT) {
- H5D_layout_t dcpl_layout; /* Dataset's layout information */
- H5O_pline_t dcpl_pline; /* Dataset's I/O pipeline information */
- H5D_alloc_time_t alloc_time; /* Dataset's allocation time */
+ if(new_dset->shared->dcpl_id != H5P_DATASET_CREATE_DEFAULT) {
+ H5D_layout_t *layout; /* Dataset's layout information */
+ H5O_pline_t *pline; /* Dataset's I/O pipeline information */
+ H5O_fill_t *fill; /* Dataset's fill value info */
/* Get new dataset's property list object */
- if (NULL == (dc_plist = H5I_object(new_dset->shared->dcpl_id)))
+ if(NULL == (dc_plist = H5I_object(new_dset->shared->dcpl_id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, NULL, "can't get dataset creation property list")
- if(H5P_get(dc_plist, H5D_CRT_DATA_PIPELINE_NAME, &dcpl_pline) < 0)
+ /* Retrieve the properties we need */
+ pline = &new_dset->shared->dcpl_cache.pline;
+ if(H5P_get(dc_plist, H5D_CRT_DATA_PIPELINE_NAME, pline) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve pipeline filter")
- if(H5P_get(dc_plist, H5D_CRT_LAYOUT_NAME, &dcpl_layout) < 0)
+ layout = &new_dset->shared->layout.type;
+ if(H5P_get(dc_plist, H5D_CRT_LAYOUT_NAME, layout) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve layout")
- if(dcpl_pline.nused > 0 && H5D_CHUNKED != dcpl_layout)
+ if(pline->nused > 0 && H5D_CHUNKED != *layout)
HGOTO_ERROR(H5E_DATASET, H5E_BADVALUE, NULL, "filters can only be used with chunked layout")
- if(H5P_get(dc_plist, H5D_CRT_ALLOC_TIME_NAME, &alloc_time) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve space allocation time")
+ fill = &new_dset->shared->dcpl_cache.fill;
+ if(H5P_get(dc_plist, H5D_CRT_FILL_VALUE_NAME, fill) < 0)
+ HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve fill value info")
/* Check if the alloc_time is the default and error out */
- if(alloc_time==H5D_ALLOC_TIME_DEFAULT)
+ if(fill->alloc_time == H5D_ALLOC_TIME_DEFAULT)
HGOTO_ERROR(H5E_DATASET, H5E_BADVALUE, NULL, "invalid space allocation state")
/* Don't allow compact datasets to allocate space later */
- if(dcpl_layout==H5D_COMPACT && alloc_time!=H5D_ALLOC_TIME_EARLY)
- HGOTO_ERROR(H5E_DATASET, H5E_BADVALUE, NULL, "compact dataset doesn't support late space allocation")
-
- /* Set the alloc_time for the dataset, in case the default was used */
- new_dset->shared->alloc_time=alloc_time;
+ if(*layout == H5D_COMPACT && fill->alloc_time != H5D_ALLOC_TIME_EARLY)
+ HGOTO_ERROR(H5E_DATASET, H5E_BADVALUE, NULL, "compact dataset must have early space allocation")
/* If MPI VFD is used, no filter support yet. */
- if(IS_H5FD_MPI(file) && dcpl_pline.nused > 0)
+ if(IS_H5FD_MPI(file) && pline->nused > 0)
HGOTO_ERROR(H5E_DATASET, H5E_UNSUPPORTED, NULL, "Parallel I/O does not support filters yet")
/* Chunked datasets are non-default, so retrieve their info here */
@@ -1545,18 +1429,15 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve chunk dimensions")
/* Get the dataset's external file list information */
- if(H5P_get(dc_plist, H5D_CRT_EXT_FILE_LIST_NAME, &new_dset->shared->efl) < 0)
+ if(H5P_get(dc_plist, H5D_CRT_EXT_FILE_LIST_NAME, &new_dset->shared->dcpl_cache.efl) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve external file list")
-
- /* Set the dataset's data storage method */
- new_dset->shared->layout.type=dcpl_layout;
} /* end if */
/* Check if this dataset is going into a parallel file and set space allocation time */
if(IS_H5FD_MPI(file))
- new_dset->shared->alloc_time=H5D_ALLOC_TIME_EARLY;
+ new_dset->shared->dcpl_cache.fill.alloc_time = H5D_ALLOC_TIME_EARLY;
- switch (new_dset->shared->layout.type) {
+ switch(new_dset->shared->layout.type) {
case H5D_CONTIGUOUS:
{
hssize_t tmp_size; /* Temporary holder for raw data size */
@@ -1570,41 +1451,40 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
*/
new_dset->shared->layout.u.contig.addr = HADDR_UNDEF; /* Initialize to no address */
- if ((ndims=H5S_get_simple_extent_dims(new_dset->shared->space, dim, max_dim))<0)
+ if((ndims = H5S_get_simple_extent_dims(new_dset->shared->space, dim, max_dim)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "unable to initialize contiguous storage")
- for (i=1; i<ndims; i++)
- if (max_dim[i]>dim[i])
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, NULL, "only the first dimension can be extendible")
- if (new_dset->shared->efl.nused>0) {
- hsize_t max_points = H5S_get_npoints_max (new_dset->shared->space);
- hsize_t max_storage = H5O_efl_total_size (&new_dset->shared->efl);
-
- if (H5S_UNLIMITED==max_points) {
- if (H5O_EFL_UNLIMITED!=max_storage)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, NULL, "unlimited data space but finite storage")
- } else if (max_points * H5T_get_size (type) < max_points) {
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, NULL, "data space * type size overflowed")
- } else if (max_points * H5T_get_size (type) > max_storage) {
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, NULL, "data space size exceeds external storage size")
+ for(i = 1; i < ndims; i++)
+ if(max_dim[i] > dim[i])
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "only the first dimension can be extendible")
+ if(new_dset->shared->dcpl_cache.efl.nused > 0) {
+ hsize_t max_points = H5S_get_npoints_max(new_dset->shared->space);
+ hsize_t max_storage = H5O_efl_total_size(&new_dset->shared->dcpl_cache.efl);
+
+ if(H5S_UNLIMITED == max_points) {
+ if(H5O_EFL_UNLIMITED != max_storage)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "unlimited data space but finite storage")
+ } else if(max_points * H5T_get_size(type) < max_points) {
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "data space * type size overflowed")
+ } else if(max_points * H5T_get_size(type) > max_storage) {
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "data space size exceeds external storage size")
}
/* Set the I/O functions for this layout type */
- new_dset->shared->io_ops.readvv=H5D_efl_readvv;
- new_dset->shared->io_ops.writevv=H5D_efl_writevv;
+ new_dset->shared->io_ops.readvv = H5D_efl_readvv;
+ new_dset->shared->io_ops.writevv = H5D_efl_writevv;
} /* end if */
else {
- if (ndims>0 && max_dim[0]>dim[0])
- HGOTO_ERROR (H5E_DATASET, H5E_UNSUPPORTED, NULL, "extendible contiguous non-external dataset")
+ if(ndims > 0 && max_dim[0] > dim[0])
+ HGOTO_ERROR(H5E_DATASET, H5E_UNSUPPORTED, NULL, "extendible contiguous non-external dataset")
/* Set the I/O functions for this layout type */
- new_dset->shared->io_ops.readvv=H5D_contig_readvv;
- new_dset->shared->io_ops.writevv=H5D_contig_writevv;
+ new_dset->shared->io_ops.readvv = H5D_contig_readvv;
+ new_dset->shared->io_ops.writevv = H5D_contig_writevv;
} /* end else */
/* Compute the total size of a chunk */
- tmp_size = H5S_GET_EXTENT_NPOINTS(new_dset->shared->space) *
- H5T_get_size(new_dset->shared->type);
- H5_ASSIGN_OVERFLOW(new_dset->shared->layout.u.contig.size,tmp_size,hssize_t,hsize_t);
+ tmp_size = H5S_GET_EXTENT_NPOINTS(new_dset->shared->space) * H5T_get_size(new_dset->shared->type);
+ H5_ASSIGN_OVERFLOW(new_dset->shared->layout.u.contig.size, tmp_size, hssize_t, hsize_t);
/* Get the sieve buffer size for this dataset */
new_dset->shared->cache.contig.sieve_buf_size = H5F_SIEVE_BUF_SIZE(file);
@@ -1616,12 +1496,13 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
hsize_t max_dim[H5O_LAYOUT_NDIMS]; /* Maximum size of data in elements */
/* Set up layout information */
- if((ndims=H5S_GET_EXTENT_NDIMS(new_dset->shared->space))<0)
+ if((ndims = H5S_GET_EXTENT_NDIMS(new_dset->shared->space)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, NULL, "unable to get rank")
new_dset->shared->layout.u.chunk.ndims = (unsigned)ndims + 1;
- assert((unsigned)(new_dset->shared->layout.u.chunk.ndims) <= NELMTS(new_dset->shared->layout.u.chunk.dim));
+ HDassert((unsigned)(new_dset->shared->layout.u.chunk.ndims) <= NELMTS(new_dset->shared->layout.u.chunk.dim));
- new_dset->shared->layout.u.chunk.addr = HADDR_UNDEF; /* Initialize to no address */
+ /* Initialize to no address */
+ new_dset->shared->layout.u.chunk.addr = HADDR_UNDEF;
/*
* Chunked storage allows any type of data space extension, so we
@@ -1629,8 +1510,8 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
*/
if(chunk_ndims != (unsigned)ndims)
HGOTO_ERROR(H5E_DATASET, H5E_BADVALUE, NULL, "dimensionality of chunks doesn't match the data space")
- if (new_dset->shared->efl.nused>0)
- HGOTO_ERROR (H5E_DATASET, H5E_BADVALUE, NULL, "external storage not supported with chunked layout")
+ if(new_dset->shared->dcpl_cache.efl.nused > 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_BADVALUE, NULL, "external storage not supported with chunked layout")
/*
* The chunk size of a dimension with a fixed size cannot exceed
@@ -1640,26 +1521,26 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, NULL, "can't retrieve chunk size")
new_dset->shared->layout.u.chunk.dim[new_dset->shared->layout.u.chunk.ndims-1] = H5T_get_size(new_dset->shared->type);
- if (H5S_get_simple_extent_dims(new_dset->shared->space, NULL, max_dim)<0)
+ if(H5S_get_simple_extent_dims(new_dset->shared->space, NULL, max_dim) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "unable to query maximum dimensions")
- for (u=0; u<new_dset->shared->layout.u.chunk.ndims-1; u++)
+ for(u = 0; u < new_dset->shared->layout.u.chunk.ndims - 1; u++)
if(max_dim[u] != H5S_UNLIMITED && max_dim[u] < new_dset->shared->layout.u.chunk.dim[u])
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, NULL, "chunk size must be <= maximum dimension size for fixed-sized dimensions")
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "chunk size must be <= maximum dimension size for fixed-sized dimensions")
/* Compute the total size of a chunk */
- for (u=1, new_dset->shared->layout.u.chunk.size=new_dset->shared->layout.u.chunk.dim[0]; u<new_dset->shared->layout.u.chunk.ndims; u++)
+ for(u = 1, new_dset->shared->layout.u.chunk.size = new_dset->shared->layout.u.chunk.dim[0]; u < new_dset->shared->layout.u.chunk.ndims; u++)
new_dset->shared->layout.u.chunk.size *= new_dset->shared->layout.u.chunk.dim[u];
/* Set the I/O functions for this layout type */
- new_dset->shared->io_ops.readvv=H5D_istore_readvv;
- new_dset->shared->io_ops.writevv=H5D_istore_writevv;
+ new_dset->shared->io_ops.readvv = H5D_istore_readvv;
+ new_dset->shared->io_ops.writevv = H5D_istore_writevv;
/* Initialize the chunk cache for the dataset */
- if(H5D_istore_init(file,new_dset)<0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, NULL, "can't initialize chunk cache")
+ if(H5D_istore_init(file, new_dset) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "can't initialize chunk cache")
/* Indicate that the chunk information was initialized */
- chunk_init=TRUE;
+ chunk_init = TRUE;
} /* end case */
break;
@@ -1672,20 +1553,19 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
* Compact dataset is stored in dataset object header message of
* layout.
*/
- tmp_size = H5S_GET_EXTENT_NPOINTS(space) *
- H5T_get_size(new_dset->shared->type);
- H5_ASSIGN_OVERFLOW(new_dset->shared->layout.u.compact.size,tmp_size,hssize_t,size_t);
+ tmp_size = H5S_GET_EXTENT_NPOINTS(space) * H5T_get_size(new_dset->shared->type);
+ H5_ASSIGN_OVERFLOW(new_dset->shared->layout.u.compact.size, tmp_size, hssize_t, size_t);
/* Verify data size is smaller than maximum header message size
* (64KB) minus other layout message fields.
*/
- comp_data_size=H5O_MESG_MAX_SIZE-H5O_layout_meta_size(file, &(new_dset->shared->layout));
+ comp_data_size = H5O_MESG_MAX_SIZE - H5O_layout_meta_size(file, &(new_dset->shared->layout));
if(new_dset->shared->layout.u.compact.size > comp_data_size)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "compact dataset size is bigger than header message maximum size")
/* Set the I/O functions for this layout type */
- new_dset->shared->io_ops.readvv=H5D_compact_readvv;
- new_dset->shared->io_ops.writevv=H5D_compact_writevv;
+ new_dset->shared->io_ops.readvv = H5D_compact_readvv;
+ new_dset->shared->io_ops.writevv = H5D_compact_writevv;
} /* end case */
break;
@@ -1693,14 +1573,10 @@ H5D_create(H5F_t *file, hid_t type_id, const H5S_t *space,
HGOTO_ERROR(H5E_DATASET, H5E_UNSUPPORTED, NULL, "not implemented yet")
} /* end switch */ /*lint !e788 All appropriate cases are covered */
- /* Update the dataset's entry info. */
- if(H5D_update_entry_info(file, dxpl_id, new_dset) != SUCCEED)
+ /* Update the dataset's object header info. */
+ if(H5D_update_oh_info(file, dxpl_id, new_dset) != SUCCEED)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, NULL, "can't update the metadata cache")
- /* Get the dataset's DCPL cache info */
- if (H5D_get_dcpl_cache(new_dset->shared->dcpl_id,&new_dset->shared->dcpl_cache)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, NULL, "can't fill DCPL cache")
-
/* Add the dataset to the list of opened objects in the file */
if(H5FO_top_incr(new_dset->oloc.file, new_dset->oloc.addr) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINC, NULL, "can't incr object ref. count")
@@ -1782,11 +1658,11 @@ H5D_open(const H5G_loc_t *loc, hid_t dxpl_id)
/* Shallow copy (take ownership) of the object location object */
if(H5O_loc_copy(&(dataset->oloc), loc->oloc, H5_COPY_SHALLOW) < 0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTCOPY, NULL, "can't copy object location")
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, NULL, "can't copy object location")
/* Shallow copy (take ownership) of the group hier. path */
if(H5G_name_copy(&(dataset->path), loc->path, H5_COPY_SHALLOW) < 0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTCOPY, NULL, "can't copy path")
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, NULL, "can't copy path")
/* Check if dataset was already open */
if((shared_fo = H5FO_opened(dataset->oloc.file, dataset->oloc.addr)) == NULL) {
@@ -1864,27 +1740,20 @@ done:
static herr_t
H5D_open_oid(H5D_t *dataset, hid_t dxpl_id)
{
- H5O_fill_new_t fill; /* Default fill value */
- unsigned alloc_time_state; /* Allocation time state */
- H5O_fill_t *fill_prop; /* Pointer to dataset's fill value area */
- H5O_pline_t pline; /* I/O pipeline information */
H5P_genplist_t *plist; /* Property list */
- herr_t ret_value = SUCCEED; /* Return value */
+ H5O_pline_t *pline; /* I/O pipeline information */
+ H5O_fill_t *fill_prop; /* Pointer to dataset's fill value info */
+ unsigned alloc_time_state; /* Allocation time state */
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_NOAPI_NOINIT(H5D_open_oid)
/* check args */
HDassert(dataset);
- /* Portably initialize the fill value */
- HDmemset(&fill, 0, sizeof(H5O_fill_new_t));
- fill.alloc_time = H5D_ALLOC_TIME_LATE;
- fill.fill_time = H5D_FILL_TIME_ALLOC;
- fill.fill_defined = TRUE;
-
/* (Set the 'vl_type' parameter to FALSE since it doesn't matter from here) */
if(NULL == (dataset->shared = H5D_new(H5P_DATASET_CREATE_DEFAULT, FALSE, FALSE)))
- HGOTO_ERROR (H5E_RESOURCE, H5E_NOSPACE, FAIL, "memory allocation failed")
+ HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, FAIL, "memory allocation failed")
/* Open the dataset object */
if(H5O_open(&(dataset->oloc)) < 0)
@@ -1893,24 +1762,25 @@ H5D_open_oid(H5D_t *dataset, hid_t dxpl_id)
/* Get the type and space */
if(NULL == (dataset->shared->type = H5O_msg_read(&(dataset->oloc), H5O_DTYPE_ID, 0, NULL, dxpl_id)))
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to load type info from dataset header")
- /* Get a datatype ID for the dataset's datatype */
- if((dataset->shared->type_id = H5I_register(H5I_DATATYPE, dataset->shared->type))<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTREGISTER, FAIL, "unable to register type")
-
if(NULL == (dataset->shared->space = H5S_read(&(dataset->oloc), dxpl_id)))
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to load space info from dataset header")
+ /* Get a datatype ID for the dataset's datatype */
+ if((dataset->shared->type_id = H5I_register(H5I_DATATYPE, dataset->shared->type)) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTREGISTER, FAIL, "unable to register type")
+
/* Get dataset creation property list object */
if(NULL == (plist = H5I_object(dataset->shared->dcpl_id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "can't get dataset creation property list")
/* Get the optional filters message */
- if(NULL == H5O_msg_read(&(dataset->oloc), H5O_PLINE_ID, 0, &pline, dxpl_id)) {
- H5E_clear_stack(NULL);
- HDmemset(&pline, 0, sizeof(pline));
+ pline = &dataset->shared->dcpl_cache.pline;
+ if(NULL != H5O_msg_read(&(dataset->oloc), H5O_PLINE_ID, 0, pline, dxpl_id)) {
+ if(H5P_set(plist, H5D_CRT_DATA_PIPELINE_NAME, pline) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set pipeline")
} /* end if */
- if(H5P_set(plist, H5D_CRT_DATA_PIPELINE_NAME, &pline) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set pipeline")
+ else
+ H5E_clear_stack(NULL);
/*
* Get the raw data layout info. It's actually stored in two locations:
@@ -1922,23 +1792,22 @@ H5D_open_oid(H5D_t *dataset, hid_t dxpl_id)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to read data layout message")
if(H5P_set(plist, H5D_CRT_LAYOUT_NAME, &dataset->shared->layout.type) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set layout")
- switch (dataset->shared->layout.type) {
+ switch(dataset->shared->layout.type) {
case H5D_CONTIGUOUS:
/* Compute the size of the contiguous storage for versions of the
* layout message less than version 3 because versions 1 & 2 would
* truncate the dimension sizes to 32-bits of information. - QAK 5/26/04
*/
- if(dataset->shared->layout.version<3) {
+ if(dataset->shared->layout.version < 3) {
hssize_t tmp_size; /* Temporary holder for raw data size */
- tmp_size = H5S_GET_EXTENT_NPOINTS(dataset->shared->space) *
- H5T_get_size(dataset->shared->type);
- H5_ASSIGN_OVERFLOW(dataset->shared->layout.u.contig.size,tmp_size,hssize_t,hsize_t);
+ tmp_size = H5S_GET_EXTENT_NPOINTS(dataset->shared->space) * H5T_get_size(dataset->shared->type);
+ H5_ASSIGN_OVERFLOW(dataset->shared->layout.u.contig.size, tmp_size, hssize_t, hsize_t);
} /* end if */
/* Set the I/O functions for this layout type */
- dataset->shared->io_ops.readvv=H5D_contig_readvv;
- dataset->shared->io_ops.writevv=H5D_contig_writevv;
+ dataset->shared->io_ops.readvv = H5D_contig_readvv;
+ dataset->shared->io_ops.writevv = H5D_contig_writevv;
/* Get the sieve buffer size for this dataset */
dataset->shared->cache.contig.sieve_buf_size = H5F_SIEVE_BUF_SIZE(dataset->oloc.file);
@@ -1961,19 +1830,19 @@ H5D_open_oid(H5D_t *dataset, hid_t dxpl_id)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set chunk size")
/* Initialize the chunk cache for the dataset */
- if(H5D_istore_init(dataset->oloc.file, dataset)<0)
+ if(H5D_istore_init(dataset->oloc.file, dataset) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "can't initialize chunk cache")
}
/* Set the I/O functions for this layout type */
- dataset->shared->io_ops.readvv=H5D_istore_readvv;
- dataset->shared->io_ops.writevv=H5D_istore_writevv;
+ dataset->shared->io_ops.readvv = H5D_istore_readvv;
+ dataset->shared->io_ops.writevv = H5D_istore_writevv;
break;
case H5D_COMPACT:
/* Set the I/O functions for this layout type */
- dataset->shared->io_ops.readvv=H5D_compact_readvv;
- dataset->shared->io_ops.writevv=H5D_compact_writevv;
+ dataset->shared->io_ops.readvv = H5D_compact_readvv;
+ dataset->shared->io_ops.writevv = H5D_compact_writevv;
break;
default:
@@ -1981,86 +1850,65 @@ H5D_open_oid(H5D_t *dataset, hid_t dxpl_id)
} /* end switch */ /*lint !e788 All appropriate cases are covered */
/* Point at dataset's copy, to cache it for later */
- fill_prop=&dataset->shared->fill;
+ fill_prop = &dataset->shared->dcpl_cache.fill;
- /* Retrieve & release the previous fill-value settings */
- if(H5P_get(plist, H5D_CRT_FILL_VALUE_NAME, fill_prop) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't get fill value")
- if(H5O_msg_reset(H5O_FILL_ID, fill_prop)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTFREE, FAIL, "can't release fill info")
-
- /* Get the new fill value message */
- if(NULL == H5O_msg_read(&(dataset->oloc), H5O_FILL_NEW_ID, 0, &fill, dxpl_id)) {
+ /* Try to get the new fill value message from the object header */
+ if(NULL == H5O_msg_read(&(dataset->oloc), H5O_FILL_NEW_ID, 0, fill_prop, dxpl_id)) {
H5E_clear_stack(NULL);
- HDmemset(&fill, 0, sizeof(fill));
- /* Set the space allocation time appropriately, based on the type of dataset storage */
- switch (dataset->shared->layout.type) {
- case H5D_COMPACT:
- fill.alloc_time=H5D_ALLOC_TIME_EARLY;
- break;
+ /* For backward compatibility, try to retrieve the old fill value message */
+ if(NULL == H5O_msg_read(&(dataset->oloc), H5O_FILL_ID, 0, fill_prop, dxpl_id)) {
+ H5E_clear_stack(NULL);
- case H5D_CONTIGUOUS:
- fill.alloc_time=H5D_ALLOC_TIME_LATE;
- break;
+ /* Set the space allocation time appropriately, based on the type of dataset storage */
+ switch(dataset->shared->layout.type) {
+ case H5D_COMPACT:
+ fill_prop->alloc_time = H5D_ALLOC_TIME_EARLY;
+ break;
- case H5D_CHUNKED:
- fill.alloc_time=H5D_ALLOC_TIME_INCR;
- break;
+ case H5D_CONTIGUOUS:
+ fill_prop->alloc_time = H5D_ALLOC_TIME_LATE;
+ break;
- default:
- HGOTO_ERROR(H5E_DATASET, H5E_UNSUPPORTED, FAIL, "not implemented yet")
- } /* end switch */ /*lint !e788 All appropriate cases are covered */
+ case H5D_CHUNKED:
+ fill_prop->alloc_time = H5D_ALLOC_TIME_INCR;
+ break;
- /* Set the default fill time */
- fill.fill_time = H5D_FILL_TIME_IFSET;
- } /* end if */
- if(fill.fill_defined) {
- if(NULL==H5O_msg_copy(H5O_FILL_ID, &fill, fill_prop))
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "can't copy fill value")
- } else {
- /* For compatibility with v1.4. Retrieve the old fill value message.
- * If size is 0, make it -1 for undefined. */
- if(NULL == H5O_msg_read(&(dataset->oloc), H5O_FILL_ID, 0, fill_prop, dxpl_id)) {
- H5E_clear_stack(NULL);
- HDmemset(fill_prop, 0, sizeof(H5O_fill_t));
+ default:
+ HGOTO_ERROR(H5E_DATASET, H5E_UNSUPPORTED, FAIL, "not implemented yet")
+ } /* end switch */ /*lint !e788 All appropriate cases are covered */
} /* end if */
- if(fill_prop->size == 0) {
- fill_prop->type = fill_prop->buf = NULL;
+
+ /* If size is 0, make it -1 for undefined. */
+ if(fill_prop->size == 0)
fill_prop->size = (size_t)-1;
- } /* end if */
- } /* end else */
- alloc_time_state=0;
- if( (dataset->shared->layout.type==H5D_COMPACT && fill.alloc_time==H5D_ALLOC_TIME_EARLY)
- || (dataset->shared->layout.type==H5D_CONTIGUOUS && fill.alloc_time==H5D_ALLOC_TIME_LATE)
- || (dataset->shared->layout.type==H5D_CHUNKED && fill.alloc_time==H5D_ALLOC_TIME_INCR))
- alloc_time_state=1;
+ } /* end if */
+ alloc_time_state = 0;
+ if((dataset->shared->layout.type == H5D_COMPACT && fill_prop->alloc_time == H5D_ALLOC_TIME_EARLY)
+ || (dataset->shared->layout.type == H5D_CONTIGUOUS && fill_prop->alloc_time == H5D_ALLOC_TIME_LATE)
+ || (dataset->shared->layout.type == H5D_CHUNKED && fill_prop->alloc_time == H5D_ALLOC_TIME_INCR))
+ alloc_time_state = 1;
/* Set revised fill value properties */
if(H5P_set(plist, H5D_CRT_FILL_VALUE_NAME, fill_prop) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set fill value")
- dataset->shared->alloc_time=fill.alloc_time; /* Cache this for later */
- if(H5P_set(plist, H5D_CRT_ALLOC_TIME_NAME, &fill.alloc_time) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set allocation time")
if(H5P_set(plist, H5D_CRT_ALLOC_TIME_STATE_NAME, &alloc_time_state) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set allocation time state")
- dataset->shared->fill_time=fill.fill_time; /* Cache this for later */
- if(H5P_set(plist, H5D_CRT_FILL_TIME_NAME, &fill.fill_time) < 0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set fill time")
/* Get the external file list message, which might not exist. Space is
* also undefined when space allocate time is H5D_ALLOC_TIME_LATE. */
- if((dataset->shared->layout.type==H5D_CONTIGUOUS && !H5F_addr_defined(dataset->shared->layout.u.contig.addr))
- || (dataset->shared->layout.type==H5D_CHUNKED && !H5F_addr_defined(dataset->shared->layout.u.chunk.addr))) {
- HDmemset(&dataset->shared->efl,0,sizeof(H5O_efl_t));
- if(NULL != H5O_msg_read(&(dataset->oloc), H5O_EFL_ID, 0, &dataset->shared->efl, dxpl_id)) {
- if(H5P_set(plist, H5D_CRT_EXT_FILE_LIST_NAME, &dataset->shared->efl) < 0)
+ if((dataset->shared->layout.type == H5D_CONTIGUOUS && !H5F_addr_defined(dataset->shared->layout.u.contig.addr))
+ || (dataset->shared->layout.type == H5D_CHUNKED && !H5F_addr_defined(dataset->shared->layout.u.chunk.addr))) {
+ if(NULL != H5O_msg_read(&(dataset->oloc), H5O_EFL_ID, 0, &dataset->shared->dcpl_cache.efl, dxpl_id)) {
+ if(H5P_set(plist, H5D_CRT_EXT_FILE_LIST_NAME, &dataset->shared->dcpl_cache.efl) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTSET, FAIL, "can't set external file list")
/* Override the I/O functions for this layout type */
- dataset->shared->io_ops.readvv=H5D_efl_readvv;
- dataset->shared->io_ops.writevv=H5D_efl_writevv;
+ dataset->shared->io_ops.readvv = H5D_efl_readvv;
+ dataset->shared->io_ops.writevv = H5D_efl_writevv;
} /* end if */
+ else
+ H5E_clear_stack(NULL);
} /* end if */
/*
@@ -2069,22 +1917,14 @@ H5D_open_oid(H5D_t *dataset, hid_t dxpl_id)
* be fully allocated before I/O can happen.
*/
if((H5F_get_intent(dataset->oloc.file) & H5F_ACC_RDWR)
- && ((dataset->shared->layout.type==H5D_CONTIGUOUS && !H5F_addr_defined(dataset->shared->layout.u.contig.addr))
- || (dataset->shared->layout.type==H5D_CHUNKED && !H5F_addr_defined(dataset->shared->layout.u.chunk.addr)))
+ && ((dataset->shared->layout.type == H5D_CONTIGUOUS && !H5F_addr_defined(dataset->shared->layout.u.contig.addr))
+ || (dataset->shared->layout.type == H5D_CHUNKED && !H5F_addr_defined(dataset->shared->layout.u.chunk.addr)))
&& IS_H5FD_MPI(dataset->oloc.file)) {
- if(H5D_alloc_storage(dataset->oloc.file, dxpl_id, dataset,H5D_ALLOC_OPEN, TRUE, FALSE)<0)
+ if(H5D_alloc_storage(dataset->oloc.file, dxpl_id, dataset, H5D_ALLOC_OPEN, TRUE, FALSE) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to initialize file storage")
} /* end if */
- /* Get the dataset's DCPL cache info */
- if(H5D_get_dcpl_cache(dataset->shared->dcpl_id,&dataset->shared->dcpl_cache)<0)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't fill DCPL cache")
-
done:
- /* Release fill value information */
- if(H5O_msg_reset(H5O_FILL_ID, &fill) < 0)
- HDONE_ERROR(H5E_DATASET, H5E_CANTRELEASE, FAIL, "unable to release fill-value info")
-
if(ret_value < 0) {
if(H5F_addr_defined(dataset->oloc.addr)) {
if(H5O_close(&(dataset->oloc)) < 0)
@@ -2121,14 +1961,14 @@ done:
herr_t
H5D_close(H5D_t *dataset)
{
- unsigned free_failed=FALSE;
- herr_t ret_value=SUCCEED; /* Return value */
+ unsigned free_failed = FALSE;
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_NOAPI(H5D_close, FAIL)
/* check args */
- assert(dataset && dataset->oloc.file && dataset->shared);
- assert(dataset->shared->fo_count >0);
+ HDassert(dataset && dataset->oloc.file && dataset->shared);
+ HDassert(dataset->shared->fo_count >0);
/* Dump debugging info */
#ifdef H5D_ISTORE_DEBUG
@@ -2136,21 +1976,20 @@ H5D_close(H5D_t *dataset)
#endif /* H5F_ISTORE_DEBUG */
dataset->shared->fo_count--;
- if(dataset->shared->fo_count ==0)
- {
+ if(dataset->shared->fo_count == 0) {
/* Free the data sieve buffer, if it's been allocated */
if(dataset->shared->cache.contig.sieve_buf) {
- assert(dataset->shared->layout.type!=H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
+ HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
/* Flush the raw data buffer, if its dirty */
- if (dataset->shared->cache.contig.sieve_dirty) {
+ if(dataset->shared->cache.contig.sieve_dirty) {
/* Write dirty data sieve buffer to file */
- if (H5F_block_write(dataset->oloc.file, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc,
+ if(H5F_block_write(dataset->oloc.file, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc,
dataset->shared->cache.contig.sieve_size, H5AC_dxpl_id, dataset->shared->cache.contig.sieve_buf) < 0)
HGOTO_ERROR(H5E_IO, H5E_WRITEERROR, FAIL, "block write failed")
/* Reset sieve buffer dirty flag */
- dataset->shared->cache.contig.sieve_dirty=0;
+ dataset->shared->cache.contig.sieve_dirty = FALSE;
} /* end if */
dataset->shared->cache.contig.sieve_buf = H5FL_BLK_FREE (sieve_buf,dataset->shared->cache.contig.sieve_buf);
@@ -2163,7 +2002,7 @@ H5D_close(H5D_t *dataset)
case H5D_CHUNKED:
/* Flush and destroy chunks in the cache */
- if(H5D_istore_dest(dataset,H5AC_dxpl_id)<0)
+ if(H5D_istore_dest(dataset,H5AC_dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTRELEASE, FAIL, "unable to destroy chunk cache")
break;
@@ -2176,13 +2015,13 @@ H5D_close(H5D_t *dataset)
} /* end if */
/* Free the buffer for the raw data for compact datasets */
- dataset->shared->layout.u.compact.buf=H5MM_xfree(dataset->shared->layout.u.compact.buf);
+ dataset->shared->layout.u.compact.buf = H5MM_xfree(dataset->shared->layout.u.compact.buf);
break;
default:
- assert ("not implemented yet" && 0);
+ HDassert("not implemented yet" && 0);
#ifdef NDEBUG
- HGOTO_ERROR (H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
+ HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
#endif /* NDEBUG */
} /* end switch */ /*lint !e788 All appropriate cases are covered */
@@ -2190,7 +2029,7 @@ H5D_close(H5D_t *dataset)
* Release datatype, dataspace and creation property list -- there isn't
* much we can do if one of these fails, so we just continue.
*/
- free_failed=(H5I_dec_ref(dataset->shared->type_id)<0 || H5S_close(dataset->shared->space)<0 ||
+ free_failed = (H5I_dec_ref(dataset->shared->type_id) < 0 || H5S_close(dataset->shared->space) < 0 ||
H5I_dec_ref(dataset->shared->dcpl_id) < 0);
/* Remove the dataset from the list of opened objects in the file */
@@ -2205,17 +2044,16 @@ H5D_close(H5D_t *dataset)
HGOTO_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release object header")
/*
- * Free memory. Before freeing the memory set the file pointer to NULL.
- * We always check for a null file pointer in other H5D functions to be
- * sure we're not accessing an already freed dataset (see the assert()
- * above).
- */
+ * Free memory. Before freeing the memory set the file pointer to NULL.
+ * We always check for a null file pointer in other H5D functions to be
+ * sure we're not accessing an already freed dataset (see the HDassert()
+ * above).
+ */
dataset->oloc.file = NULL;
- H5FL_FREE(H5D_shared_t,dataset->shared);
+ H5FL_FREE(H5D_shared_t, dataset->shared);
} /* end if */
- else
- {
+ else {
/* Decrement the ref. count for this object in the top file */
if(H5FO_top_decr(dataset->oloc.file, dataset->oloc.addr) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTRELEASE, FAIL, "can't decrement count for object")
@@ -2226,17 +2064,20 @@ H5D_close(H5D_t *dataset)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to close")
} /* end else */
+ /* Release the dataset's path info */
if(H5G_name_free(&(dataset->path)) < 0)
- free_failed=TRUE;
+ free_failed = TRUE;
- H5FL_FREE(H5D_t,dataset);
+ /* Free the dataset's memory structure */
+ H5FL_FREE(H5D_t, dataset);
- if (free_failed)
- HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "couldn't free the type or creation property list, but the dataset was freed anyway.")
+ /* Check if anything failed in the middle... */
+ if(free_failed)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "couldn't free a component of the dataset, but the dataset was freed anyway.")
done:
FUNC_LEAVE_NOAPI(ret_value)
-}
+} /* end H5D_close() */
/*-------------------------------------------------------------------------
@@ -2252,45 +2093,43 @@ done:
*-------------------------------------------------------------------------
*/
static herr_t
-H5D_extend (H5D_t *dataset, const hsize_t *size, hid_t dxpl_id)
+H5D_extend(H5D_t *dataset, const hsize_t *size, hid_t dxpl_id)
{
- int changed; /* Flag to indicate that the dataspace was successfully extended */
- H5S_t *space = NULL; /* Dataset's dataspace */
- H5D_fill_value_t fill_status;
- H5D_fill_time_t fill_time;
- herr_t ret_value=SUCCEED; /* Return value */
+ htri_t changed; /* Flag to indicate that the dataspace was successfully extended */
+ H5S_t *space; /* Dataset's dataspace */
+ H5O_fill_t *fill; /* Dataset's fill value */
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_NOAPI(H5D_extend, FAIL)
/* Check args */
- assert (dataset);
- assert (size);
+ HDassert(dataset);
+ HDassert(size);
/* Check if the filters in the DCPL will need to encode, and if so, can they?
* Filters need encoding if fill value is defined and a fill policy is set that requires
* writing on an extend.
*/
- if(! dataset->shared->checked_filters)
- {
- if(H5P_is_fill_value_defined(&(dataset->shared->fill), &fill_status) < 0)
- HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "Couldn't retrieve fill value from dataset.")
+ fill = &dataset->shared->dcpl_cache.fill;
+ if(!dataset->shared->checked_filters) {
+ H5D_fill_value_t fill_status; /* Whether the fill value is defined */
- if(fill_status == H5D_FILL_VALUE_DEFAULT || fill_status == H5D_FILL_VALUE_USER_DEFINED)
- {
- if( H5Pget_fill_time(dataset->shared->dcpl_id, &fill_time) < 0)
- HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "Couldn't retrieve fill time from dataset.")
+ /* Retrieve the "defined" status of the fill value */
+ if(H5P_is_fill_value_defined(fill, &fill_status) < 0)
+ HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "Couldn't retrieve fill value from dataset.")
- if(fill_time == H5D_FILL_TIME_ALLOC ||
- (fill_time == H5D_FILL_TIME_IFSET && fill_status == H5D_FILL_VALUE_USER_DEFINED) )
- {
+ /* See if we can check the filter status */
+ if(fill_status == H5D_FILL_VALUE_DEFAULT || fill_status == H5D_FILL_VALUE_USER_DEFINED) {
+ if(fill->fill_time == H5D_FILL_TIME_ALLOC ||
+ (fill->fill_time == H5D_FILL_TIME_IFSET && fill_status == H5D_FILL_VALUE_USER_DEFINED)) {
/* Filters must have encoding enabled. Ensure that all filters can be applied */
- if(H5Z_can_apply(dataset->shared->dcpl_id, dataset->shared->type_id) <0)
+ if(H5Z_can_apply(dataset->shared->dcpl_id, dataset->shared->type_id) < 0)
HGOTO_ERROR(H5E_PLINE, H5E_CANAPPLY, FAIL, "can't apply filters")
dataset->shared->checked_filters = TRUE;
- }
- }
- }
+ } /* end if */
+ } /* end if */
+ } /* end if */
/*
* NOTE: Restrictions on extensions were checked when the dataset was
@@ -2299,11 +2138,12 @@ H5D_extend (H5D_t *dataset, const hsize_t *size, hid_t dxpl_id)
*/
/* Increase the size of the data space */
- space=dataset->shared->space;
- if ((changed=H5S_extend (space, size))<0)
- HGOTO_ERROR (H5E_DATASET, H5E_CANTINIT, FAIL, "unable to increase size of data space")
+ space = dataset->shared->space;
+ if((changed = H5S_extend(space, size)) < 0)
+ HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to increase size of data space")
- if (changed>0){
+ /* Updated the dataset's info if the dataspace was successfully extended */
+ if(changed) {
/* Save the new dataspace in the file if necessary */
if(H5S_write(&(dataset->oloc), space, TRUE, dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update file with new dataspace")
@@ -2314,14 +2154,14 @@ H5D_extend (H5D_t *dataset, const hsize_t *size, hid_t dxpl_id)
HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update cached chunk indices")
/* Allocate space for the new parts of the dataset, if appropriate */
- if(dataset->shared->alloc_time == H5D_ALLOC_TIME_EARLY)
+ if(fill->alloc_time == H5D_ALLOC_TIME_EARLY)
if(H5D_alloc_storage(dataset->oloc.file, dxpl_id, dataset, H5D_ALLOC_EXTEND, TRUE, FALSE) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to initialize dataset with fill value")
} /* end if */
done:
FUNC_LEAVE_NOAPI(ret_value)
-}
+} /* end H5D_extend() */
/*-------------------------------------------------------------------------
@@ -2386,16 +2226,16 @@ H5D_nameof(H5D_t *dataset)
*-------------------------------------------------------------------------
*/
H5T_t *
-H5D_typeof (const H5D_t *dset)
+H5D_typeof(const H5D_t *dset)
{
/* Use FUNC_ENTER_NOAPI_NOINIT_NOFUNC here to avoid performance issues */
FUNC_ENTER_NOAPI_NOINIT_NOFUNC(H5D_typeof)
- assert (dset);
- assert (dset->shared->type);
+ HDassert(dset);
+ HDassert(dset->shared->type);
FUNC_LEAVE_NOAPI(dset->shared->type)
-}
+} /* end H5D_typeof() */
/*-------------------------------------------------------------------------
@@ -2442,92 +2282,90 @@ H5D_alloc_storage(H5F_t *f, hid_t dxpl_id, H5D_t *dset/*in,out*/, H5D_time_alloc
hbool_t update_time, hbool_t full_overwrite)
{
struct H5O_layout_t *layout; /* The dataset's layout information */
- unsigned init_space=0; /* Flag to indicate that space should be initialized */
- unsigned addr_set=0; /* Flag to indicate that the dataset's storage address was set */
+ hbool_t must_init_space = FALSE; /* Flag to indicate that space should be initialized */
+ hbool_t addr_set = FALSE; /* Flag to indicate that the dataset's storage address was set */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_NOAPI_NOINIT(H5D_alloc_storage)
/* check args */
- assert (f);
- assert (dset);
+ HDassert(f);
+ HDassert(dset);
/* If the data is stored in external files, don't set an address for the layout
* We assume that external storage is already
* allocated by the caller, or at least will be before I/O is performed.
*/
- if(!(H5S_NULL == H5S_GET_EXTENT_TYPE(dset->shared->space) || dset->shared->efl.nused>0)) {
+ if(!(H5S_NULL == H5S_GET_EXTENT_TYPE(dset->shared->space) || dset->shared->dcpl_cache.efl.nused > 0)) {
/* Get a pointer to the dataset's layout information */
- layout=&(dset->shared->layout);
+ layout = &(dset->shared->layout);
- switch (layout->type) {
+ switch(layout->type) {
case H5D_CONTIGUOUS:
- if(layout->u.contig.addr==HADDR_UNDEF) {
+ if(!H5F_addr_defined(layout->u.contig.addr)) {
/* Reserve space in the file for the entire array */
- if (H5D_contig_create (f, dxpl_id, layout/*out*/)<0)
- HGOTO_ERROR (H5E_IO, H5E_CANTINIT, FAIL, "unable to initialize contiguous storage")
+ if(H5D_contig_create(f, dxpl_id, layout/*out*/) < 0)
+ HGOTO_ERROR(H5E_IO, H5E_CANTINIT, FAIL, "unable to initialize contiguous storage")
/* Indicate that we set the storage addr */
- addr_set=1;
+ addr_set = TRUE;
/* Indicate that we should initialize storage space */
- init_space=1;
+ must_init_space = TRUE;
} /* end if */
break;
case H5D_CHUNKED:
- if(layout->u.chunk.addr==HADDR_UNDEF) {
+ if(!H5F_addr_defined(layout->u.chunk.addr)) {
/* Create the root of the B-tree that describes chunked storage */
- if (H5D_istore_create (f, dxpl_id, layout/*out*/)<0)
- HGOTO_ERROR (H5E_IO, H5E_CANTINIT, FAIL, "unable to initialize chunked storage")
+ if(H5D_istore_create(f, dxpl_id, layout/*out*/) < 0)
+ HGOTO_ERROR(H5E_IO, H5E_CANTINIT, FAIL, "unable to initialize chunked storage")
/* Indicate that we set the storage addr */
- addr_set=1;
+ addr_set = TRUE;
/* Indicate that we should initialize storage space */
- init_space=1;
+ must_init_space = TRUE;
} /* end if */
/* If space allocation is set to 'early' and we are extending
* the dataset, indicate that space should be allocated, so the
* B-tree gets expanded. -QAK
*/
- if(dset->shared->alloc_time == H5D_ALLOC_TIME_EARLY &&
- time_alloc == H5D_ALLOC_EXTEND)
- {
- init_space=1;
- }
+ if(dset->shared->dcpl_cache.fill.alloc_time == H5D_ALLOC_TIME_EARLY
+ && time_alloc == H5D_ALLOC_EXTEND)
+ must_init_space = 1;
break;
case H5D_COMPACT:
/* Check if space is already allocated */
if(layout->u.compact.buf==NULL) {
/* Reserve space in layout header message for the entire array. */
- assert(layout->u.compact.size>0);
- if (NULL==(layout->u.compact.buf=H5MM_malloc(layout->u.compact.size)))
- HGOTO_ERROR (H5E_RESOURCE, H5E_NOSPACE, FAIL, "unable to allocate memory for compact dataset")
+ HDassert(layout->u.compact.size > 0);
+ if(NULL == (layout->u.compact.buf = H5MM_malloc(layout->u.compact.size)))
+ HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, FAIL, "unable to allocate memory for compact dataset")
if(!full_overwrite)
HDmemset(layout->u.compact.buf, 0, layout->u.compact.size);
layout->u.compact.dirty = TRUE;
/* Indicate that we set the storage addr */
- addr_set=1;
+ addr_set = TRUE;
/* Indicate that we should initialize storage space */
- init_space=1;
+ must_init_space = TRUE;
} /* end if */
break;
default:
- assert ("not implemented yet" && 0);
+ HDassert("not implemented yet" && 0);
#ifdef NDEBUG
- HGOTO_ERROR (H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
+ HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
#endif /* NDEBUG */
} /* end switch */ /*lint !e788 All appropriate cases are covered */
/* Check if we need to initialize the space */
- if(init_space) {
- if (layout->type==H5D_CHUNKED) {
+ if(must_init_space) {
+ if(layout->type == H5D_CHUNKED) {
/* If we are doing incremental allocation and the B-tree got
* created during a H5Dwrite call, don't initialize the storage
* now, wait for the actual writes to each block and let the
@@ -2537,22 +2375,21 @@ H5D_alloc_storage(H5F_t *f, hid_t dxpl_id, H5D_t *dset/*in,out*/, H5D_time_alloc
* fill values to the chunks they allocate space for. Yes,
* this is icky. -QAK
*/
- if(!(dset->shared->alloc_time==H5D_ALLOC_TIME_INCR && time_alloc==H5D_ALLOC_WRITE)) {
+ if(!(dset->shared->dcpl_cache.fill.alloc_time == H5D_ALLOC_TIME_INCR && time_alloc == H5D_ALLOC_WRITE))
if(H5D_init_storage(dset, full_overwrite, dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to initialize dataset with fill value")
- } /* end if */
} /* end if */
else {
H5D_fill_value_t fill_status; /* The fill value status */
/* Check the dataset's fill-value status */
- if (H5P_is_fill_value_defined(&(dset->shared->fill), &fill_status) < 0)
+ if(H5P_is_fill_value_defined(&dset->shared->dcpl_cache.fill, &fill_status) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't tell if fill value defined")
/* If we are filling the dataset on allocation or "if set" and
* the fill value _is_ set, do that now */
- if(dset->shared->fill_time==H5D_FILL_TIME_ALLOC ||
- (dset->shared->fill_time==H5D_FILL_TIME_IFSET && fill_status==H5D_FILL_VALUE_USER_DEFINED)) {
+ if(dset->shared->dcpl_cache.fill.fill_time == H5D_FILL_TIME_ALLOC ||
+ (dset->shared->dcpl_cache.fill.fill_time == H5D_FILL_TIME_IFSET && fill_status == H5D_FILL_VALUE_USER_DEFINED)) {
if(H5D_init_storage(dset, full_overwrite, dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to initialize dataset with fill value")
} /* end if */
@@ -2563,7 +2400,7 @@ H5D_alloc_storage(H5F_t *f, hid_t dxpl_id, H5D_t *dset/*in,out*/, H5D_time_alloc
* set the address. (this is improves forward compatibility).
*/
if(time_alloc != H5D_ALLOC_CREATE && addr_set)
- if(H5O_msg_write(&(dset->oloc), H5O_LAYOUT_ID, 0, H5O_MSG_FLAG_CONSTANT, update_time, &(dset->shared->layout), dxpl_id) < 0)
+ if(H5O_msg_write(&dset->oloc, H5O_LAYOUT_ID, 0, H5O_MSG_FLAG_CONSTANT, update_time, &dset->shared->layout, dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to update layout message")
} /* end if */
@@ -2593,25 +2430,25 @@ H5D_init_storage(H5D_t *dset, hbool_t full_overwrite, hid_t dxpl_id)
FUNC_ENTER_NOAPI_NOINIT(H5D_init_storage)
- assert(dset);
+ HDassert(dset);
switch (dset->shared->layout.type) {
case H5D_COMPACT:
{
- hssize_t snpoints; /* Number of points in space (for error checking) */
- size_t npoints; /* Number of points in space */
+ hssize_t snpoints; /* Number of points in space (for error checking) */
+ size_t npoints; /* Number of points in space */
/* Get the number of elements in the dataset's dataspace */
snpoints = H5S_GET_EXTENT_NPOINTS(dset->shared->space);
- assert(snpoints>=0);
- H5_ASSIGN_OVERFLOW(npoints,snpoints,hssize_t,size_t);
+ HDassert(snpoints >= 0);
+ H5_ASSIGN_OVERFLOW(npoints, snpoints, hssize_t, size_t);
/* If we will be immediately overwriting the values, don't bother to clear them */
if(!full_overwrite) {
/* If the fill value is defined, initialize the data buffer with it */
- if(dset->shared->fill.buf)
+ if(dset->shared->dcpl_cache.fill.buf)
/* Initialize the cached data buffer with the fill value */
- H5V_array_fill(dset->shared->layout.u.compact.buf, dset->shared->fill.buf, dset->shared->fill.size, npoints);
+ H5V_array_fill(dset->shared->layout.u.compact.buf, dset->shared->dcpl_cache.fill.buf, (size_t)dset->shared->dcpl_cache.fill.size, npoints);
else /* If the fill value is default, zero set data buf. */
HDmemset(dset->shared->layout.u.compact.buf, 0, dset->shared->layout.u.compact.size);
} /* end if */
@@ -2621,10 +2458,9 @@ H5D_init_storage(H5D_t *dset, hbool_t full_overwrite, hid_t dxpl_id)
case H5D_CONTIGUOUS:
/* Don't write default fill values to external files */
/* If we will be immediately overwriting the values, don't bother to clear them */
- if((dset->shared->efl.nused==0 || dset->shared->fill.buf) && !full_overwrite) {
- if (H5D_contig_fill(dset, dxpl_id)<0)
+ if((dset->shared->dcpl_cache.efl.nused == 0 || dset->shared->dcpl_cache.fill.buf) && !full_overwrite)
+ if(H5D_contig_fill(dset, dxpl_id) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to allocate all chunks of dataset")
- } /* end if */
break;
case H5D_CHUNKED:
@@ -2632,14 +2468,14 @@ H5D_init_storage(H5D_t *dset, hbool_t full_overwrite, hid_t dxpl_id)
* Allocate file space
* for all chunks now and initialize each chunk with the fill value.
*/
- if (H5D_istore_allocate(dset, dxpl_id, full_overwrite)<0)
+ if(H5D_istore_allocate(dset, dxpl_id, full_overwrite) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to allocate all chunks of dataset")
break;
default:
- assert ("not implemented yet" && 0);
+ HDassert("not implemented yet" && 0);
#ifdef NDEBUG
- HGOTO_ERROR (H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
+ HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
#endif /* NDEBUG */
} /* end switch */ /*lint !e788 All appropriate cases are covered */
@@ -2676,7 +2512,7 @@ H5Dget_storage_size(hid_t dset_id)
H5TRACE1("h", "i", dset_id);
/* Check args */
- if (NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
+ if(NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, 0, "not a dataset")
/* Set return value */
@@ -2762,7 +2598,7 @@ H5Dget_offset(hid_t dset_id)
H5TRACE1("a", "i", dset_id);
/* Check args */
- if (NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
+ if(NULL==(dset=H5I_object_verify(dset_id, H5I_DATASET)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, HADDR_UNDEF, "not a dataset")
/* Set return value */
@@ -2797,7 +2633,7 @@ H5D_get_offset(const H5D_t *dset)
FUNC_ENTER_NOAPI(H5D_get_offset, HADDR_UNDEF)
- assert(dset);
+ HDassert(dset);
switch(dset->shared->layout.type) {
case H5D_CHUNKED:
@@ -2819,7 +2655,7 @@ H5D_get_offset(const H5D_t *dset)
break;
default:
- assert ("not implemented yet" && 0);
+ HDassert("not implemented yet" && 0);
#ifdef NDEBUG
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, HADDR_UNDEF, "unknown dataset layout type")
#endif /* NDEBUG */
@@ -2978,7 +2814,7 @@ H5Dvlen_reclaim(hid_t type_id, hid_t space_id, hid_t plist_id, void *buf)
H5TRACE4("e", "iiix", type_id, space_id, plist_id, buf);
/* Check args */
- if (H5I_DATATYPE != H5I_get_type(type_id) || buf == NULL)
+ if(H5I_DATATYPE != H5I_get_type(type_id) || buf == NULL)
HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid argument")
if(NULL == (space = H5I_object_verify(space_id, H5I_DATASPACE)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "invalid dataspace")
@@ -3031,7 +2867,7 @@ H5D_vlen_reclaim(hid_t type_id, H5S_t *space, hid_t plist_id, void *buf)
HDassert(buf);
/* Get the allocation info */
- if(H5T_vlen_get_alloc_info(plist_id,&vl_alloc_info)<0)
+ if(H5T_vlen_get_alloc_info(plist_id,&vl_alloc_info) < 0)
HGOTO_ERROR(H5E_DATATYPE, H5E_CANTGET, FAIL, "unable to retrieve VL allocation info")
/* Call H5D_iterate with args, etc. */
@@ -3066,7 +2902,7 @@ H5D_vlen_get_buf_size_alloc(size_t size, void *info)
FUNC_ENTER_NOAPI(H5D_vlen_get_buf_size_alloc, NULL)
/* Get a temporary pointer to space for the VL data */
- if ((vlen_bufsize->vl_tbuf=H5FL_BLK_REALLOC(vlen_vl_buf,vlen_bufsize->vl_tbuf,size))!=NULL)
+ if((vlen_bufsize->vl_tbuf=H5FL_BLK_REALLOC(vlen_vl_buf,vlen_bufsize->vl_tbuf,size))!=NULL)
vlen_bufsize->size+=size;
/* Set return value */
@@ -3107,10 +2943,10 @@ static herr_t
H5D_vlen_get_buf_size(void UNUSED *elem, hid_t type_id, unsigned UNUSED ndim, const hsize_t *point, void *op_data)
{
H5D_vlen_bufsize_t *vlen_bufsize = (H5D_vlen_bufsize_t *)op_data;
- H5T_t *dt = NULL;
- herr_t ret_value = 0; /* The correct return value, if this function succeeds */
+ H5T_t *dt; /* Datatype for operation */
+ herr_t ret_value = SUCCEED; /* Return value */
- FUNC_ENTER_NOAPI(H5D_vlen_get_buf_size, FAIL)
+ FUNC_ENTER_NOAPI_NOINIT(H5D_vlen_get_buf_size)
HDassert(op_data);
HDassert(H5I_DATATYPE == H5I_get_type(type_id));
@@ -3120,20 +2956,20 @@ H5D_vlen_get_buf_size(void UNUSED *elem, hid_t type_id, unsigned UNUSED ndim, co
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a datatype")
/* Make certain there is enough fixed-length buffer available */
- if((vlen_bufsize->fl_tbuf=H5FL_BLK_REALLOC(vlen_fl_buf, vlen_bufsize->fl_tbuf, H5T_get_size(dt))) == NULL)
+ if(NULL == (vlen_bufsize->fl_tbuf = H5FL_BLK_REALLOC(vlen_fl_buf, vlen_bufsize->fl_tbuf, H5T_get_size(dt))))
HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, FAIL, "can't resize tbuf")
/* Select point to read in */
- if(H5Sselect_elements(vlen_bufsize->fspace_id, H5S_SELECT_SET, (size_t)1, (const hsize_t **)point)<0)
+ if(H5Sselect_elements(vlen_bufsize->fspace_id, H5S_SELECT_SET, (size_t)1, (const hsize_t **)point) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTCREATE, FAIL, "can't select point")
/* Read in the point (with the custom VL memory allocator) */
- if(H5Dread(vlen_bufsize->dataset_id, type_id, vlen_bufsize->mspace_id, vlen_bufsize->fspace_id, vlen_bufsize->xfer_pid, vlen_bufsize->fl_tbuf)<0)
+ if(H5Dread(vlen_bufsize->dataset_id, type_id, vlen_bufsize->mspace_id, vlen_bufsize->fspace_id, vlen_bufsize->xfer_pid, vlen_bufsize->fl_tbuf) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_READERROR, FAIL, "can't read point")
done:
FUNC_LEAVE_NOAPI(ret_value)
-} /* end H5D_vlen_get_buf_size() */
+} /* end H5D_vlen_get_buf_size() */
/*-------------------------------------------------------------------------
@@ -3187,7 +3023,7 @@ H5Dvlen_get_buf_size(hid_t dataset_id, hid_t type_id, hid_t space_id,
vlen_bufsize.dataset_id=dataset_id;
/* Get a copy of the dataspace ID */
- if((vlen_bufsize.fspace_id=H5Dget_space(dataset_id))<0)
+ if((vlen_bufsize.fspace_id=H5Dget_space(dataset_id)) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTCOPY, FAIL, "can't copy dataspace")
/* Create a scalar for the memory dataspace */
@@ -3201,19 +3037,19 @@ H5Dvlen_get_buf_size(hid_t dataset_id, hid_t type_id, hid_t space_id,
HGOTO_ERROR(H5E_RESOURCE, H5E_NOSPACE, FAIL, "no temporary buffers available")
/* Get the pointer to the dataset transfer class */
- if (NULL == (pclass = H5I_object(H5P_CLS_DATASET_XFER_g)))
+ if(NULL == (pclass = H5I_object(H5P_CLS_DATASET_XFER_g)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a property list class")
/* Change to the custom memory allocation routines for reading VL data */
- if((vlen_bufsize.xfer_pid=H5P_create_id(pclass))<0)
+ if((vlen_bufsize.xfer_pid=H5P_create_id(pclass)) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTCREATE, FAIL, "no dataset xfer plists available")
/* Get the property list struct */
- if (NULL == (plist = H5I_object(vlen_bufsize.xfer_pid)))
+ if(NULL == (plist = H5I_object(vlen_bufsize.xfer_pid)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a dataset transfer property list")
/* Set the memory manager to the special allocation routine */
- if(H5P_set_vlen_mem_manager(plist,H5D_vlen_get_buf_size_alloc,&vlen_bufsize,NULL,NULL)<0)
+ if(H5P_set_vlen_mem_manager(plist,H5D_vlen_get_buf_size_alloc,&vlen_bufsize,NULL,NULL) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTINIT, FAIL, "can't set VL data allocation routine")
/* Set the initial number of bytes required */
@@ -3223,24 +3059,24 @@ H5Dvlen_get_buf_size(hid_t dataset_id, hid_t type_id, hid_t space_id,
ret_value = H5D_iterate(&bogus, type_id, space, H5D_vlen_get_buf_size, &vlen_bufsize);
/* Get the size if we succeeded */
- if(ret_value>=0)
- *size=vlen_bufsize.size;
+ if(ret_value >= 0)
+ *size = vlen_bufsize.size;
done:
- if(vlen_bufsize.fspace_id>0) {
- if(H5I_dec_ref(vlen_bufsize.fspace_id)<0)
+ if(vlen_bufsize.fspace_id > 0) {
+ if(H5I_dec_ref(vlen_bufsize.fspace_id) < 0)
HDONE_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release dataspace")
} /* end if */
- if(vlen_bufsize.mspace_id>0) {
- if(H5I_dec_ref(vlen_bufsize.mspace_id)<0)
+ if(vlen_bufsize.mspace_id > 0) {
+ if(H5I_dec_ref(vlen_bufsize.mspace_id) < 0)
HDONE_ERROR(H5E_DATASET, H5E_CLOSEERROR, FAIL, "unable to release dataspace")
} /* end if */
- if(vlen_bufsize.fl_tbuf!=NULL)
- H5FL_BLK_FREE(vlen_fl_buf,vlen_bufsize.fl_tbuf);
- if(vlen_bufsize.vl_tbuf!=NULL)
- H5FL_BLK_FREE(vlen_vl_buf,vlen_bufsize.vl_tbuf);
- if(vlen_bufsize.xfer_pid>0) {
- if(H5I_dec_ref(vlen_bufsize.xfer_pid)<0)
+ if(vlen_bufsize.fl_tbuf != NULL)
+ H5FL_BLK_FREE(vlen_fl_buf, vlen_bufsize.fl_tbuf);
+ if(vlen_bufsize.vl_tbuf != NULL)
+ H5FL_BLK_FREE(vlen_vl_buf, vlen_bufsize.vl_tbuf);
+ if(vlen_bufsize.xfer_pid > 0) {
+ if(H5I_dec_ref(vlen_bufsize.xfer_pid) < 0)
HDONE_ERROR(H5E_DATASET, H5E_CANTDEC, FAIL, "unable to decrement ref count on property list")
} /* end if */
@@ -3249,24 +3085,23 @@ done:
/*-------------------------------------------------------------------------
- * Function: H5Dset_extent
- *
- * Purpose: Modifies the dimensions of a dataset, based on H5Dextend.
- * Can change to a lower dimension.
+ * Function: H5Dset_extent
*
- * Return: Success: SUCCEED, Failure: FAIL
+ * Purpose: Modifies the dimensions of a dataset, based on H5Dextend.
+ * Can change to a lower dimension.
*
- * Programmer: Pedro Vicente, pvn@ncsa.uiuc.edu
+ * Return: Non-negative on success, negative on failure
*
- * Date: April 9, 2002
+ * Programmer: Pedro Vicente, pvn@ncsa.uiuc.edu
+ * April 9, 2002
*
*-------------------------------------------------------------------------
*/
herr_t
H5Dset_extent(hid_t dset_id, const hsize_t *size)
{
- H5D_t *dset = NULL;
- herr_t ret_value=SUCCEED; /* Return value */
+ H5D_t *dset; /* Dataset for this operation */
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_API(H5Dset_extent, FAIL)
H5TRACE2("e", "i*h", dset_id, size);
@@ -3283,41 +3118,39 @@ H5Dset_extent(hid_t dset_id, const hsize_t *size)
done:
FUNC_LEAVE_API(ret_value)
-}
+} /* end H5Dset_extent() */
/*-------------------------------------------------------------------------
- * Function: H5D_set_extent
- *
- * Purpose: Based in H5D_extend, allows change to a lower dimension,
- * calls H5S_set_extent and H5D_istore_prune_by_extent instead
+ * Function: H5D_set_extent
*
- * Return: Success: SUCCEED, Failure: FAIL
+ * Purpose: Based on H5D_extend, allows change to a lower dimension,
+ * calls H5S_set_extent and H5D_istore_prune_by_extent instead
*
- * Programmer: Pedro Vicente, pvn@ncsa.uiuc.edu
- * Robb Matzke
+ * Return: Non-negative on success, negative on failure
*
- * Date: April 9, 2002
+ * Programmer: Pedro Vicente, pvn@ncsa.uiuc.edu
+ * April 9, 2002
*
*-------------------------------------------------------------------------
*/
static herr_t
H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id)
{
- hsize_t curr_dims[H5O_LAYOUT_NDIMS]; /* Current dimension sizes */
- int rank; /* Dataspace # of dimensions */
- herr_t ret_value = SUCCEED; /* Return value */
- H5S_t *space = NULL;
- int u;
- unsigned shrink = 0; /* Flag to indicate a dimension has shrank */
- unsigned expand = 0; /* Flag to indicate a dimension has grown */
- int changed = 0;
+ H5S_t *space; /* Dataset's dataspace */
+ int rank; /* Dataspace # of dimensions */
+ hsize_t curr_dims[H5O_LAYOUT_NDIMS]; /* Current dimension sizes */
+ hbool_t shrink = FALSE; /* Flag to indicate a dimension has shrank */
+ hbool_t expand = FALSE; /* Flag to indicate a dimension has grown */
+ htri_t changed; /* Whether the dataspace changed size */
+ unsigned u; /* Local index variable */
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_NOAPI(H5D_set_extent, FAIL)
/* Check args */
- assert(dset);
- assert(size);
+ HDassert(dset);
+ HDassert(size);
/*-------------------------------------------------------------------------
* Get the data space
@@ -3332,18 +3165,19 @@ H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id)
if((rank = H5S_get_simple_extent_dims(space, curr_dims, NULL)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get dataset dimensions")
- for(u = 0; u < rank; u++) {
+ /* Determine if we are shrinking and/or expanding any dimensions */
+ for(u = 0; u < (unsigned)rank; u++) {
if(size[u] < curr_dims[u])
- shrink = 1;
+ shrink = TRUE;
if(size[u] > curr_dims[u])
- expand = 1;
- }
+ expand = TRUE;
+ } /* end for */
/*-------------------------------------------------------------------------
* Modify the size of the data space
*-------------------------------------------------------------------------
*/
- if((changed=H5S_set_extent(space, size)) < 0)
+ if((changed = H5S_set_extent(space, size)) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to modify size of data space")
/* Don't bother updating things, unless they've changed */
@@ -3362,7 +3196,7 @@ H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id)
HGOTO_ERROR(H5E_DATASET, H5E_WRITEERROR, FAIL, "unable to update cached chunk indices")
/* Allocate space for the new parts of the dataset, if appropriate */
- if(expand && dset->shared->alloc_time==H5D_ALLOC_TIME_EARLY)
+ if(expand && dset->shared->dcpl_cache.fill.alloc_time == H5D_ALLOC_TIME_EARLY)
if(H5D_alloc_storage(dset->oloc.file, dxpl_id, dset, H5D_ALLOC_EXTEND, TRUE, FALSE) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to initialize dataset storage")
@@ -3375,14 +3209,14 @@ H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id)
if(shrink && H5D_CHUNKED == dset->shared->layout.type) {
H5D_io_info_t io_info; /* Dataset I/O info */
H5D_dxpl_cache_t _dxpl_cache; /* Data transfer property cache buffer */
- H5D_dxpl_cache_t *dxpl_cache=&_dxpl_cache; /* Data transfer property cache */
+ H5D_dxpl_cache_t *dxpl_cache = &_dxpl_cache; /* Data transfer property cache */
/* Fill the DXPL cache values for later use */
- if (H5D_get_dxpl_cache(dxpl_id,&dxpl_cache)<0)
+ if(H5D_get_dxpl_cache(dxpl_id, &dxpl_cache) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't fill dxpl cache")
/* Construct dataset I/O info */
- H5D_BUILD_IO_INFO(&io_info,dset,dxpl_cache,dxpl_id,NULL);
+ H5D_BUILD_IO_INFO(&io_info, dset, dxpl_cache, dxpl_id, NULL);
/* Remove excess chunks */
if(H5D_istore_prune_by_extent(&io_info) < 0)
@@ -3396,7 +3230,7 @@ H5D_set_extent(H5D_t *dset, const hsize_t *size, hid_t dxpl_id)
done:
FUNC_LEAVE_NOAPI(ret_value)
-}
+} /* end H5D_set_extent() */
/*-------------------------------------------------------------------------
@@ -3407,52 +3241,53 @@ done:
* Return: Success: Non-negative
* Failure: Negative
*
- *
* Programmer: Ray Lu
- *
- * Date: August 14, 2002
+ * August 14, 2002
*
*-------------------------------------------------------------------------
*/
herr_t
H5D_flush(const H5F_t *f, hid_t dxpl_id, unsigned flags)
{
- unsigned num_dsets; /* Number of datasets in file */
- hid_t *id_list=NULL; /* list of dataset IDs */
- H5D_t *dataset=NULL; /* Dataset pointer */
- unsigned u; /* Index variable */
- herr_t ret_value = SUCCEED; /* Return value */
+ hid_t *id_list = NULL; /* Array of dataset IDs */
+ unsigned num_dsets; /* Number of datasets in file */
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_NOAPI(H5D_flush, FAIL)
/* Check args */
- assert(f);
+ HDassert(f);
/* Retrieve the # of outstanding dataset IDs */
- num_dsets=H5F_get_obj_count(f, H5F_OBJ_DATASET);
+ num_dsets = H5F_get_obj_count(f, H5F_OBJ_DATASET);
/* Check for something to do */
- if(num_dsets>0) {
- H5_CHECK_OVERFLOW(num_dsets,unsigned,size_t);
- if(NULL==(id_list=H5MM_malloc((size_t)num_dsets*sizeof(hid_t))))
+ if(num_dsets > 0) {
+ H5D_t *dataset; /* Dataset pointer */
+ unsigned u; /* Index variable */
+
+ H5_CHECK_OVERFLOW(num_dsets, unsigned, size_t);
+ if(NULL == (id_list = H5MM_malloc((size_t)num_dsets * sizeof(hid_t))))
HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to allocate memory for ID list")
if(H5F_get_obj_ids(f, H5F_OBJ_DATASET, -1, id_list) != num_dsets)
HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to get dataset ID list")
+
+ /* Iterate through the open datasets */
for(u = 0; u < num_dsets; u++) {
- if(NULL==(dataset=H5I_object_verify(id_list[u], H5I_DATASET)))
+ if(NULL == (dataset = H5I_object_verify(id_list[u], H5I_DATASET)))
HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to get dataset object")
/* Flush the raw data buffer, if we have a dirty one */
- if (dataset->shared->cache.contig.sieve_buf && dataset->shared->cache.contig.sieve_dirty) {
- assert(dataset->shared->layout.type!=H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
+ if(dataset->shared->cache.contig.sieve_buf && dataset->shared->cache.contig.sieve_dirty) {
+ HDassert(dataset->shared->layout.type != H5D_COMPACT); /* We should never have a sieve buffer for compact storage */
/* Write dirty data sieve buffer to file */
- if (H5F_block_write(f, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc,
+ if(H5F_block_write(f, H5FD_MEM_DRAW, dataset->shared->cache.contig.sieve_loc,
dataset->shared->cache.contig.sieve_size, dxpl_id, dataset->shared->cache.contig.sieve_buf) < 0)
HGOTO_ERROR(H5E_IO, H5E_WRITEERROR, FAIL, "block write failed")
/* Reset sieve buffer dirty flag */
- dataset->shared->cache.contig.sieve_dirty=0;
+ dataset->shared->cache.contig.sieve_dirty = FALSE;
} /* end if */
/* Flush cached information for each kind of dataset */
@@ -3462,31 +3297,31 @@ H5D_flush(const H5F_t *f, hid_t dxpl_id, unsigned flags)
case H5D_CHUNKED:
/* Flush the raw data cache */
- if (H5D_istore_flush(dataset, dxpl_id, flags & H5F_FLUSH_INVALIDATE) < 0)
+ if(H5D_istore_flush(dataset, dxpl_id, flags & H5F_FLUSH_INVALIDATE) < 0)
HGOTO_ERROR(H5E_CACHE, H5E_CANTFLUSH, FAIL, "unable to flush raw data cache")
break;
case H5D_COMPACT:
if(dataset->shared->layout.u.compact.dirty) {
- if(H5O_msg_write(&(dataset->oloc), H5O_LAYOUT_ID, 0, 0, H5O_UPDATE_TIME, &(dataset->shared->layout), dxpl_id)<0)
+ if(H5O_msg_write(&(dataset->oloc), H5O_LAYOUT_ID, 0, 0, H5O_UPDATE_TIME, &(dataset->shared->layout), dxpl_id) < 0)
HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, FAIL, "unable to update layout message")
dataset->shared->layout.u.compact.dirty = FALSE;
} /* end if */
break;
default:
- assert ("not implemented yet" && 0);
+ HDassert("not implemented yet" && 0);
#ifdef NDEBUG
- HGOTO_ERROR (H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
+ HGOTO_ERROR(H5E_IO, H5E_UNSUPPORTED, FAIL, "unsupported storage layout")
#endif /* NDEBUG */
} /* end switch */ /*lint !e788 All appropriate cases are covered */
-
- }
+ } /* end for */
} /* end if */
done:
- if(id_list!=NULL)
+ if(id_list)
H5MM_xfree(id_list);
+
FUNC_LEAVE_NOAPI(ret_value)
} /* end H5D_flush() */
@@ -3509,8 +3344,8 @@ done:
herr_t
H5Ddebug(hid_t dset_id)
{
- H5D_t *dset = NULL;
- herr_t ret_value = SUCCEED; /* Return value */
+ H5D_t *dset; /* Dataset to debug */
+ herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_API(H5Ddebug, FAIL)
H5TRACE1("e", "i", dset_id);