diff options
author | Quincey Koziol <koziol@hdfgroup.org> | 2002-11-20 13:19:37 (GMT) |
---|---|---|
committer | Quincey Koziol <koziol@hdfgroup.org> | 2002-11-20 13:19:37 (GMT) |
commit | b5e3e19f860a8ca0ebd4b7149f99a725c9d67f9f (patch) | |
tree | 9232a64dc1cd7ec5da0d231198a71ded182984c4 /test | |
parent | ac48a23e2b72b03d60d4fdf10957ed31beb5b70d (diff) | |
download | hdf5-b5e3e19f860a8ca0ebd4b7149f99a725c9d67f9f.zip hdf5-b5e3e19f860a8ca0ebd4b7149f99a725c9d67f9f.tar.gz hdf5-b5e3e19f860a8ca0ebd4b7149f99a725c9d67f9f.tar.bz2 |
[svn-r6115] Purpose:
Code Cleanup & New Tests
Description:
tmisc.c:
Switched from using H5_HAVE_COMPRESSION flag in favor of
H5_HAVE_FILTER_DEFLATE.
dsets.c:
Switched from using H5_HAVE_COMPRESSION flag in favor of
H5_HAVE_FILTER_DEFLATE.
Refactored I/O filter tests to allow new filters to be [more] easily
added.
Added tests for shuffle & deflate+shuffle I/O filters (if the filter(s)
are enabled).
Added test for creating a new dataset with a filter that is not
available.
Added test for attempting to read a dataset created with a filter that
is not available.
Platforms tested:
Tested h5committest {arabica (fortran), eirene (fortran, C++)
modi4 (parallel, fortran)}
FreeBSD 4.7 (sleipnir)
Diffstat (limited to 'test')
-rw-r--r-- | test/dsets.c | 511 | ||||
-rw-r--r-- | test/tmisc.c | 24 |
2 files changed, 408 insertions, 127 deletions
diff --git a/test/dsets.c b/test/dsets.c index 672bd17..b9d900c 100644 --- a/test/dsets.c +++ b/test/dsets.c @@ -9,6 +9,7 @@ */ #include "h5test.h" +#include "H5Zprivate.h" /* Filter functions */ const char *FILENAME[] = { "dataset", @@ -16,15 +17,20 @@ const char *FILENAME[] = { NULL }; +#define FILE_DEFLATE_NAME "deflate.h5" + #define DSET_DEFAULT_NAME "default" #define DSET_CHUNKED_NAME "chunked" #define DSET_COMPACT_NAME "compact" #define DSET_SIMPLE_IO_NAME "simple_io" #define DSET_COMPACT_IO_NAME "compact_io" #define DSET_TCONV_NAME "tconv" -#define DSET_COMPRESS_NAME "compressed" -#define DSET_ONEBYTE_SHUF_NAME "onebyte_shuffle" +#define DSET_DEFLATE_NAME "deflate" +#define DSET_SHUFFLE_NAME "shuffle" +#define DSET_SHUFFLE_DEFLATE_NAME "shuffle+deflate" #define DSET_BOGUS_NAME "bogus" +#define DSET_MISSING_NAME "missing" +#define DSET_ONEBYTE_SHUF_NAME "onebyte_shuffle" #define H5Z_BOGUS 305 @@ -489,7 +495,7 @@ bogus(unsigned int UNUSED flags, size_t UNUSED cd_nelmts, /*------------------------------------------------------------------------- - * Function: test_compression + * Function: test_compression_internal * * Purpose: Tests dataset compression. If compression is requested when * it hasn't been compiled into the library (such as when @@ -497,63 +503,46 @@ bogus(unsigned int UNUSED flags, size_t UNUSED cd_nelmts, * the file uncompressed but no errors are returned. * * Return: Success: 0 - * * Failure: -1 * * Programmer: Robb Matzke * Wednesday, April 15, 1998 * * Modifications: + * Moved out of main test_compression routine + * Quincey Koziol, November 14, 2002 * *------------------------------------------------------------------------- */ static herr_t -test_compression(hid_t file) +test_compression_internal(hid_t fid, const char *name, hid_t dcpl, hsize_t *dset_size) { - hid_t dataset, space, xfer, dc; - const hsize_t size[2] = {100, 200}; - const hsize_t chunk_size[2] = {2, 25}; - const hssize_t hs_offset[2] = {7, 30}; - const hsize_t hs_size[2] = {4, 50}; -#ifndef H5_HAVE_COMPRESSION - const char *not_supported; -#endif - - hsize_t i, j, n; - void *tconv_buf = NULL; + hid_t dataset; /* Dataset ID */ + hid_t dxpl; /* Dataset xfer property list ID */ + hid_t sid; /* Dataspace ID */ + const hsize_t size[2] = {100, 200}; /* Dataspace dimensions */ + const hssize_t hs_offset[2] = {7, 30}; /* Hyperslab offset */ + const hsize_t hs_size[2] = {4, 50}; /* Hyperslab size */ + void *tconv_buf = NULL; /* Temporary conversion buffer */ + hsize_t i, j, n; /* Local index variables */ -#ifndef H5_HAVE_COMPRESSION - not_supported = " Deflate compression is not supported.\n" - " The zlib was not found when hdf5 was configured."; -#endif - - TESTING("compression (setup)"); - /* Create the data space */ - if ((space = H5Screate_simple(2, size, NULL))<0) goto error; + if ((sid = H5Screate_simple(2, size, NULL))<0) goto error; /* * Create a small conversion buffer to test strip mining. We * might as well test all we can! */ - if ((xfer = H5Pcreate (H5P_DATASET_XFER))<0) goto error; + if ((dxpl = H5Pcreate (H5P_DATASET_XFER))<0) goto error; tconv_buf = malloc (1000); - if (H5Pset_buffer (xfer, 1000, tconv_buf, NULL)<0) goto error; - - /* Use chunked storage with compression */ - if((dc = H5Pcreate(H5P_DATASET_CREATE))<0) goto error; - if (H5Pset_chunk (dc, 2, chunk_size)<0) goto error; - if (H5Pset_deflate (dc, 6)<0) goto error; + if (H5Pset_buffer (dxpl, 1000, tconv_buf, NULL)<0) goto error; + TESTING("compression (setup)"); + /* Create the dataset */ - if ((dataset = H5Dcreate(file, DSET_COMPRESS_NAME, H5T_NATIVE_INT, space, - dc))<0) goto error; -#ifdef H5_HAVE_COMPRESSION + if ((dataset = H5Dcreate(fid, name, H5T_NATIVE_INT, sid, + dcpl))<0) goto error; PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif /*---------------------------------------------------------------------- * STEP 1: Read uninitialized data. It should be zero. @@ -561,7 +550,7 @@ test_compression(hid_t file) */ TESTING("compression (uninitialized read)"); - if (H5Dread (dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, check)<0) + if (H5Dread (dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check)<0) goto error; for (i=0; i<size[0]; i++) { @@ -575,12 +564,7 @@ test_compression(hid_t file) } } } -#ifdef H5_HAVE_COMPRESSION PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif /*---------------------------------------------------------------------- * STEP 2: Test compression by setting up a chunked dataset and writing @@ -595,14 +579,9 @@ test_compression(hid_t file) } } - if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, points)<0) + if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, points)<0) goto error; -#ifdef H5_HAVE_COMPRESSION PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif /*---------------------------------------------------------------------- * STEP 3: Try to read the data we just wrote. @@ -611,7 +590,7 @@ test_compression(hid_t file) TESTING("compression (read)"); /* Read the dataset back */ - if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, check)<0) + if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check)<0) goto error; /* Check that the values read are the same as the values written */ @@ -626,12 +605,7 @@ test_compression(hid_t file) } } } -#ifdef H5_HAVE_COMPRESSION PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif /*---------------------------------------------------------------------- * STEP 4: Write new data over the top of the old data. The new data is @@ -647,11 +621,11 @@ test_compression(hid_t file) points[i][j] = rand (); } } - if (H5Dwrite (dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, points)<0) + if (H5Dwrite (dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, points)<0) goto error; /* Read the dataset back and check it */ - if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, check)<0) + if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check)<0) goto error; /* Check that the values read are the same as the values written */ @@ -666,12 +640,7 @@ test_compression(hid_t file) } } } -#ifdef H5_HAVE_COMPRESSION PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif /*---------------------------------------------------------------------- * STEP 5: Close the dataset and then open it and read it again. This @@ -682,8 +651,8 @@ test_compression(hid_t file) TESTING("compression (re-open)"); if (H5Dclose (dataset)<0) goto error; - if ((dataset = H5Dopen (file, DSET_COMPRESS_NAME))<0) goto error; - if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, check)<0) + if ((dataset = H5Dopen (fid, name))<0) goto error; + if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check)<0) goto error; /* Check that the values read are the same as the values written */ @@ -698,12 +667,7 @@ test_compression(hid_t file) } } } -#ifdef H5_HAVE_COMPRESSION PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif /*---------------------------------------------------------------------- @@ -719,11 +683,11 @@ test_compression(hid_t file) points[hs_offset[0]+i][hs_offset[1]+j] = rand (); } } - if (H5Sselect_hyperslab(space, H5S_SELECT_SET, hs_offset, NULL, hs_size, + if (H5Sselect_hyperslab(sid, H5S_SELECT_SET, hs_offset, NULL, hs_size, NULL)<0) goto error; - if (H5Dwrite (dataset, H5T_NATIVE_INT, space, space, xfer, points)<0) + if (H5Dwrite (dataset, H5T_NATIVE_INT, sid, sid, dxpl, points)<0) goto error; - if (H5Dread (dataset, H5T_NATIVE_INT, space, space, xfer, check)<0) + if (H5Dread (dataset, H5T_NATIVE_INT, sid, sid, dxpl, check)<0) goto error; /* Check that the values read are the same as the values written */ @@ -744,59 +708,375 @@ test_compression(hid_t file) } } } -#ifdef H5_HAVE_COMPRESSION PASSED(); -#else - SKIPPED(); - puts(not_supported); -#endif - /*---------------------------------------------------------------------- - * STEP 7: Register an application-defined compression method and use it - * to write and then read the dataset. - *---------------------------------------------------------------------- - */ - TESTING("compression (app-defined method)"); + /* Get the storage size of the dataset */ + if((*dset_size=H5Dget_storage_size(dataset))==0) goto error; - if (H5Zregister (H5Z_BOGUS, "bogus", bogus)<0) goto error; - if (H5Pset_filter (dc, H5Z_BOGUS, 0, 0, NULL)<0) goto error; + /* Clean up objects used for this test */ if (H5Dclose (dataset)<0) goto error; - if (H5Sclose (space)<0) goto error; - if ((space = H5Screate_simple (2, size, NULL))<0) goto error; - if ((dataset=H5Dcreate (file, DSET_BOGUS_NAME, H5T_NATIVE_INT, space, - dc))<0) goto error; + if (H5Sclose (sid)<0) goto error; + if (H5Pclose (dxpl)<0) goto error; + free (tconv_buf); - if (H5Dwrite (dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, points)<0) - goto error; - if (H5Dread (dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, check)<0) - goto error; + return(0); + +error: + return -1; +} + + +/*------------------------------------------------------------------------- + * Function: test_compression + * + * Purpose: Tests dataset compression. + * + * Return: Success: 0 + * Failure: -1 + * + * Programmer: Robb Matzke + * Wednesday, April 15, 1998 + * + * Modifications: + * Moved guts of compression testing out of main routine. + * Also added tests for shuffle filter. + * Quincey Koziol, November 14, 2002 + * + *------------------------------------------------------------------------- + */ +static herr_t +test_compression(hid_t file) +{ + hid_t dc; /* Dataset creation property list ID */ + const hsize_t chunk_size[2] = {2, 25}; /* Chunk dimensions */ + hsize_t null_size; /* Size of dataset with null filter */ +#ifdef H5_HAVE_FILTER_DEFLATE + hsize_t deflate_size; /* Size of dataset with deflate filter */ +#endif /* H5_HAVE_FILTER_DEFLATE */ +#ifdef H5_HAVE_FILTER_SHUFFLE + hsize_t shuffle_size; /* Size of dataset with shuffle filter */ +#endif /* H5_HAVE_FILTER_SHUFFLE */ +#if defined H5_HAVE_FILTER_DEFLATE && defined H5_HAVE_FILTER_SHUFFLE + hsize_t shuff_def_size; /* Size of dataset with shuffle+deflate filter */ +#endif /* H5_HAVE_FILTER_DEFLATE && H5_HAVE_FILTER_SHUFFLE */ - for (i=0; i<size[0]; i++) { - for (j=0; j<size[1]; j++) { + /* Test null I/O filter (by itself) */ + puts("Testing 'null' filter"); + if((dc = H5Pcreate(H5P_DATASET_CREATE))<0) goto error; + if (H5Pset_chunk (dc, 2, chunk_size)<0) goto error; + if (H5Zregister (H5Z_BOGUS, "bogus", bogus)<0) goto error; + if (H5Pset_filter (dc, H5Z_BOGUS, 0, 0, NULL)<0) goto error; + + if(test_compression_internal(file,DSET_BOGUS_NAME,dc,&null_size)<0) goto error; + + /* Clean up objects used for this test */ + if (H5Pclose (dc)<0) goto error; + +#ifdef H5_HAVE_FILTER_DEFLATE + /* Test deflate I/O filter (by itself) */ + puts("Testing deflate filter"); + if((dc = H5Pcreate(H5P_DATASET_CREATE))<0) goto error; + if (H5Pset_chunk (dc, 2, chunk_size)<0) goto error; + if (H5Pset_deflate (dc, 6)<0) goto error; + + if(test_compression_internal(file,DSET_DEFLATE_NAME,dc,&deflate_size)<0) goto error; + if(deflate_size>=null_size) { + H5_FAILED(); + puts(" Deflated size greater than uncompressed size."); + goto error; + } /* end if */ + + /* Clean up objects used for this test */ + if (H5Pclose (dc)<0) goto error; +#else /* H5_HAVE_FILTER_DEFLATE */ + TESTING("deflate filter"); + SKIPPED(); + puts("Deflate filter not enabled"); +#endif /* H5_HAVE_FILTER_DEFLATE */ + +#ifdef H5_HAVE_FILTER_SHUFFLE + /* Test shuffle I/O filter (by itself) */ + puts("Testing shuffle filter"); + if((dc = H5Pcreate(H5P_DATASET_CREATE))<0) goto error; + if (H5Pset_chunk (dc, 2, chunk_size)<0) goto error; + if (H5Pset_shuffle (dc, sizeof(int))<0) goto error; + + if(test_compression_internal(file,DSET_SHUFFLE_NAME,dc,&shuffle_size)<0) goto error; + if(shuffle_size!=null_size) { + H5_FAILED(); + puts(" Shuffled size not the same as uncompressed size."); + goto error; + } /* end if */ + + /* Clean up objects used for this test */ + if (H5Pclose (dc)<0) goto error; +#else /* H5_HAVE_FILTER_SHUFFLE */ + TESTING("shuffle filter"); + SKIPPED(); + puts("Shuffle filter not enabled"); +#endif /* H5_HAVE_FILTER_SHUFFLE */ + +#if defined H5_HAVE_FILTER_DEFLATE && defined H5_HAVE_FILTER_SHUFFLE + /* Test combination of deflate & shuffle I/O filters */ + puts("Testing shuffle+deflate filters"); + if((dc = H5Pcreate(H5P_DATASET_CREATE))<0) goto error; + if (H5Pset_chunk (dc, 2, chunk_size)<0) goto error; + if (H5Pset_shuffle (dc, sizeof(int))<0) goto error; + if (H5Pset_deflate (dc, 6)<0) goto error; + + if(test_compression_internal(file,DSET_SHUFFLE_DEFLATE_NAME,dc,&shuff_def_size)<0) goto error; + if(shuff_def_size>=deflate_size) { + H5_FAILED(); + puts(" Shuffle+deflate size greater than plain deflated size."); + goto error; + } /* end if */ + + /* Clean up objects used for this test */ + if (H5Pclose (dc)<0) goto error; +#else /* H5_HAVE_FILTER_DEFLATE && H5_HAVE_FILTER_SHUFFLE */ + TESTING("shuffle+deflate filters"); + SKIPPED(); + puts("Deflate or shuffle filter not enabled"); +#endif /* H5_HAVE_FILTER_DEFLATE && H5_HAVE_FILTER_SHUFFLE */ + + return 0; + +error: + return -1; +} + + +/*------------------------------------------------------------------------- + * Function: test_missing_filter + * + * Purpose: Tests library behavior when filter is missing + * + * Return: Success: 0 + * Failure: -1 + * + * Programmer: Quincey Koziol + * Thursday, November 14, 2002 + * + *------------------------------------------------------------------------- + */ +static herr_t +test_missing_filter(hid_t file) +{ + hid_t fid; /* File ID */ + hid_t dsid; /* Dataset ID */ + hid_t sid; /* Dataspace ID */ + hid_t dcpl; /* Dataspace creation property list ID */ + const hsize_t dims[2] = {100, 200}; /* Dataspace dimensions */ + const hsize_t chunk_dims[2] = {2, 25}; /* Chunk dimensions */ + hsize_t dset_size; /* Dataset size */ + hsize_t i,j; /* Local index variables */ + herr_t ret; /* Generic return value */ + char testfile[512]=""; /* Buffer to hold name of existing test file */ + char *srcdir = HDgetenv("srcdir"); /* The source directory, if we are using the --srcdir configure option */ + + TESTING("dataset access with missing filter"); + + /* Unregister the deflate filter */ +#ifdef H5_HAVE_FILTER_DEFLATE + /* Verify deflate filter is registered currently */ + if(H5Zfilter_avail(H5Z_FILTER_DEFLATE)!=TRUE) { + H5_FAILED(); + printf(" Line %d: Deflate filter not available\n",__LINE__); + goto error; + } /* end if */ + + /* Unregister deflate filter (use internal function) */ + if (H5Z_unregister(H5Z_FILTER_DEFLATE)<0) { + H5_FAILED(); + printf(" Line %d: Can't unregister deflate filter\n",__LINE__); + goto error; + } /* end if */ +#endif /* H5_HAVE_FILTER_DEFLATE */ + /* Verify deflate filter is not registered currently */ + if(H5Zfilter_avail(H5Z_FILTER_DEFLATE)!=FALSE) { + H5_FAILED(); + printf(" Line %d: Deflate filter available\n",__LINE__); + goto error; + } /* end if */ + + /* Create dcpl with deflate filter */ + if((dcpl = H5Pcreate(H5P_DATASET_CREATE))<0) { + H5_FAILED(); + printf(" Line %d: Can't create dcpl\n",__LINE__); + goto error; + } /* end if */ + if(H5Pset_chunk(dcpl, 2, chunk_dims)<0) { + H5_FAILED(); + printf(" Line %d: Can't set chunk sizes\n",__LINE__); + goto error; + } /* end if */ + if(H5Pset_deflate(dcpl, 9)<0) { + H5_FAILED(); + printf(" Line %d: Can't set deflate filter\n",__LINE__); + goto error; + } /* end if */ + + /* Create the data space */ + if ((sid = H5Screate_simple(2, dims, NULL))<0) { + H5_FAILED(); + printf(" Line %d: Can't open dataspace\n",__LINE__); + goto error; + } /* end if */ + + /* Create new dataset */ + if ((dsid = H5Dcreate(file, DSET_MISSING_NAME, H5T_NATIVE_INT, sid, dcpl))<0) { + H5_FAILED(); + printf(" Line %d: Can't create dataset\n",__LINE__); + goto error; + } /* end if */ + + /* Write data */ + if (H5Dwrite(dsid, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, points)<0) { + H5_FAILED(); + printf(" Line %d: Error writing dataset data\n",__LINE__); + goto error; + } /* end if */ + + /* Flush the file (to clear the cache) */ + if (H5Fflush(file, H5F_SCOPE_GLOBAL)<0) { + H5_FAILED(); + printf(" Line %d: Error flushing file\n",__LINE__); + goto error; + } /* end if */ + + /* Query the dataset's size on disk */ + if((dset_size=H5Dget_storage_size(dsid))==0) { + H5_FAILED(); + printf(" Line %d: Error querying dataset size\n",__LINE__); + goto error; + } /* end if */ + + /* Verify that the size indicates data is uncompressed */ + /* (i.e. the deflation filter we asked for was silently ignored) */ + if((H5Tget_size(H5T_NATIVE_INT)*100*200)!=dset_size) { + H5_FAILED(); + printf(" Line %d: Incorrect dataset size: %lu\n",__LINE__,(unsigned long)dset_size); + goto error; + } /* end if */ + + /* Read data */ + if (H5Dread(dsid, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, check)<0) { + H5_FAILED(); + printf(" Line %d: Error reading dataset data\n",__LINE__); + goto error; + } /* end if */ + + /* Compare data */ + /* Check that the values read are the same as the values written */ + for (i=0; i<dims[0]; i++) { + for (j=0; j<dims[1]; j++) { if (points[i][j] != check[i][j]) { H5_FAILED(); - printf(" Read different values than written.\n"); - printf(" At index %lu,%lu\n", - (unsigned long)i, (unsigned long)j); + printf(" Line %d: Read different values than written.\n",__LINE__); + printf(" At index %lu,%lu\n", (unsigned long)(i), (unsigned long)(j)); + printf(" At original: %d\n",points[i][j]); + printf(" At returned: %d\n",check[i][j]); goto error; - } - } + } /* end if */ + } /* end for */ + } /* end for */ + + /* Close dataset */ + if(H5Dclose(dsid)<0) { + H5_FAILED(); + printf(" Line %d: Can't close dataset\n",__LINE__); + goto error; + } /* end if */ + + /* Close dataspace */ + if(H5Sclose(sid)<0) { + H5_FAILED(); + printf(" Line %d: Can't close dataspace\n",__LINE__); + goto error; + } /* end if */ + + /* Close dataset creation property list */ + if(H5Pclose(dcpl)<0) { + H5_FAILED(); + printf(" Line %d: Can't close dcpl\n",__LINE__); + goto error; + } /* end if */ + + + /* Try reading existing dataset with deflate filter */ + + /* Compose the name of the file to open, using the srcdir, if appropriate */ + if (srcdir && ((HDstrlen(srcdir) + HDstrlen(FILE_DEFLATE_NAME) + 1) < sizeof(testfile))){ + HDstrcpy(testfile, srcdir); + HDstrcat(testfile, "/"); } + HDstrcat(testfile, FILE_DEFLATE_NAME); + + /* Open existing file */ + if((fid=H5Fopen(testfile, H5F_ACC_RDONLY, H5P_DEFAULT))<0) { + H5_FAILED(); + printf(" Line %d: Can't open existing deflated file\n",__LINE__); + goto error; + } /* end if */ + + /* Open dataset */ + if ((dsid = H5Dopen(fid, "Dataset1"))<0) { + H5_FAILED(); + printf(" Line %d: Can't open dataset\n",__LINE__); + goto error; + } /* end if */ + + /* Read data (should fail, since deflate filter is missing) */ + H5E_BEGIN_TRY { + ret=H5Dread(dsid, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, check); + } H5E_END_TRY; + if (ret>=0) { + H5_FAILED(); + printf(" Line %d: Error reading dataset data\n",__LINE__); + goto error; + } /* end if */ + + /* Close dataset */ + if(H5Dclose(dsid)<0) { + H5_FAILED(); + printf(" Line %d: Can't close dataset\n",__LINE__); + goto error; + } /* end if */ + + /* Close existing file */ + if(H5Fclose(fid)<0) { + H5_FAILED(); + printf(" Line %d: Can't close file\n",__LINE__); + goto error; + } /* end if */ + + /* Re-register the deflate filter */ + /* Verify deflate filter is not registered currently */ + if(H5Zfilter_avail(H5Z_FILTER_DEFLATE)!=FALSE) { + H5_FAILED(); + printf(" Line %d: Deflate filter available\n",__LINE__); + goto error; + } /* end if */ +#ifdef H5_HAVE_FILTER_DEFLATE + /* Register deflate filter (use internal function) */ + if(H5Z_register(H5Z_FILTER_DEFLATE, "deflate", H5Z_filter_deflate)<0) { + H5_FAILED(); + printf(" Line %d: Can't unregister deflate filter\n",__LINE__); + goto error; + } /* end if */ + + /* Verify deflate filter is registered currently */ + if(H5Zfilter_avail(H5Z_FILTER_DEFLATE)!=TRUE) { + H5_FAILED(); + printf(" Line %d: Deflate filter not available\n",__LINE__); + goto error; + } /* end if */ +#endif /* H5_HAVE_FILTER_DEFLATE */ + PASSED(); - - - - /*---------------------------------------------------------------------- - * Cleanup - *---------------------------------------------------------------------- - */ - if (H5Pclose (xfer)<0) goto error; - if (H5Pclose (dc)<0) goto error; - if (H5Dclose(dataset)<0) goto error; - free (tconv_buf); return 0; - error: +error: return -1; } @@ -805,8 +1085,8 @@ test_compression(hid_t file) * Function: test_onebyte_shuffle * * Purpose: Tests the 8-bit array with shuffling algorithm. - The shuffled array should be the same result as - that before the shuffling. + * The shuffled array should be the same result as + * that before the shuffling. * * Return: Success: 0 * @@ -1145,6 +1425,7 @@ main(void) nerrors += test_compact_io(fapl)<0 ?1:0; nerrors += test_tconv(file)<0 ?1:0; nerrors += test_compression(file)<0 ?1:0; + nerrors += test_missing_filter(file)<0 ?1:0; nerrors += test_onebyte_shuffle(file)<0 ?1:0; nerrors += test_multiopen (file)<0 ?1:0; nerrors += test_types(file)<0 ?1:0; diff --git a/test/tmisc.c b/test/tmisc.c index da31f43..6d86983 100644 --- a/test/tmisc.c +++ b/test/tmisc.c @@ -1343,10 +1343,10 @@ test_misc8(void) CHECK(ret, FAIL, "H5Pset_alloc_time"); /* Use compression as well as chunking for these datasets */ -#ifdef H5_HAVE_COMPRESSION +#ifdef H5_HAVE_FILTER_DEFLATE ret = H5Pset_deflate(dcpl,9); CHECK(ret, FAIL, "H5Pset_deflate"); -#endif /* end H5_HAVE_COMPRESSION */ +#endif /* end H5_HAVE_FILTER_DEFLATE */ /* Create a chunked dataset, with space allocation early */ did = H5Dcreate(fid, MISC8_DSETNAME8, H5T_NATIVE_INT, sid, dcpl); @@ -1359,7 +1359,7 @@ test_misc8(void) /* Check the storage size after data is written */ storage_size=H5Dget_storage_size(did); CHECK(storage_size, 0, "H5Dget_storage_size"); -#ifdef H5_HAVE_COMPRESSION +#ifdef H5_HAVE_FILTER_DEFLATE if(storage_size>=(MISC8_DIM0*MISC8_DIM1*H5Tget_size(H5T_NATIVE_INT))) { num_errs++; printf("Error on line %d: data wasn't compressed! storage_size=%u\n",__LINE__,(unsigned)storage_size); @@ -1369,7 +1369,7 @@ test_misc8(void) num_errs++; printf("Error on line %d: wrong storage size! storage_size=%u\n",__LINE__,(unsigned)storage_size); } -#endif /* H5_HAVE_COMPRESSION */ +#endif /* H5_HAVE_FILTER_DEFLATE */ /* Close dataset ID */ ret = H5Dclose(did); @@ -1395,7 +1395,7 @@ test_misc8(void) /* Check the storage size after only four chunks are written */ storage_size=H5Dget_storage_size(did); CHECK(storage_size, 0, "H5Dget_storage_size"); -#ifdef H5_HAVE_COMPRESSION +#ifdef H5_HAVE_FILTER_DEFLATE if(storage_size>=(MISC8_DIM0*MISC8_DIM1*H5Tget_size(H5T_NATIVE_INT))) { num_errs++; printf("Error on line %d: data wasn't compressed! storage_size=%u\n",__LINE__,(unsigned)storage_size); @@ -1405,7 +1405,7 @@ test_misc8(void) num_errs++; printf("Error on line %d: wrong storage size! storage_size=%u\n",__LINE__,(unsigned)storage_size); } -#endif /* H5_HAVE_COMPRESSION */ +#endif /* H5_HAVE_FILTER_DEFLATE */ /* Write entire dataset */ ret = H5Dwrite(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, wdata); @@ -1430,7 +1430,7 @@ test_misc8(void) /* Check the storage size after data is written */ storage_size=H5Dget_storage_size(did); CHECK(storage_size, 0, "H5Dget_storage_size"); -#ifdef H5_HAVE_COMPRESSION +#ifdef H5_HAVE_FILTER_DEFLATE if(storage_size>=(MISC8_DIM0*MISC8_DIM1*H5Tget_size(H5T_NATIVE_INT))) { num_errs++; printf("Error on line %d: data wasn't compressed! storage_size=%u\n",__LINE__,(unsigned)storage_size); @@ -1440,7 +1440,7 @@ test_misc8(void) num_errs++; printf("Error on line %d: wrong storage size! storage_size=%u\n",__LINE__,(unsigned)storage_size); } -#endif /*H5_HAVE_COMPRESSION*/ +#endif /*H5_HAVE_FILTER_DEFLATE*/ /* Close dataset ID */ ret = H5Dclose(did); @@ -1465,7 +1465,7 @@ test_misc8(void) /* Check the storage size after only four chunks are written */ storage_size=H5Dget_storage_size(did); CHECK(storage_size, 0, "H5Dget_storage_size"); -#ifdef H5_HAVE_COMPRESSION +#ifdef H5_HAVE_FILTER_DEFLATE if(storage_size>=(4*MISC8_CHUNK_DIM0*MISC8_CHUNK_DIM1*H5Tget_size(H5T_NATIVE_INT))) { num_errs++; printf("Error on line %d: data wasn't compressed! storage_size=%u\n",__LINE__,(unsigned)storage_size); @@ -1475,7 +1475,7 @@ test_misc8(void) num_errs++; printf("Error on line %d: wrong storage size! storage_size=%u\n",__LINE__,(unsigned)storage_size); } -#endif /* H5_HAVE_COMPRESSION */ +#endif /* H5_HAVE_FILTER_DEFLATE */ /* Write entire dataset */ ret = H5Dwrite(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, wdata); @@ -1500,7 +1500,7 @@ test_misc8(void) /* Check the storage size after data is written */ storage_size=H5Dget_storage_size(did); CHECK(storage_size, 0, "H5Dget_storage_size"); -#ifdef H5_HAVE_COMPRESSION +#ifdef H5_HAVE_FILTER_DEFLATE if(storage_size>=(MISC8_DIM0*MISC8_DIM1*H5Tget_size(H5T_NATIVE_INT))) { num_errs++; printf("Error on line %d: data wasn't compressed! storage_size=%u\n",__LINE__,(unsigned)storage_size); @@ -1510,7 +1510,7 @@ test_misc8(void) num_errs++; printf("Error on line %d: wrong storage size! storage_size=%u\n",__LINE__,(unsigned)storage_size); } -#endif /*H5_HAVE_COMPRESSION*/ +#endif /*H5_HAVE_FILTER_DEFLATE*/ /* Close dataset ID */ ret = H5Dclose(did); |