diff options
author | Raymond Lu <songyulu@hdfgroup.org> | 2011-01-21 19:04:14 (GMT) |
---|---|---|
committer | Raymond Lu <songyulu@hdfgroup.org> | 2011-01-21 19:04:14 (GMT) |
commit | 1b6c627c30c324f7ca837a4be00b97e9837383f7 (patch) | |
tree | 4235986cc12162f056bf5806d33b3d2e3ab2707d /test | |
parent | 64651d6954fd9f7010b978df0a1d25fdf376e8b8 (diff) | |
download | hdf5-1b6c627c30c324f7ca837a4be00b97e9837383f7.zip hdf5-1b6c627c30c324f7ca837a4be00b97e9837383f7.tar.gz hdf5-1b6c627c30c324f7ca837a4be00b97e9837383f7.tar.bz2 |
[svn-r19972] I added a test case for dataset with scale-offset filter into cross_read.c and updated the data files from BE, LE, and VMS.
Tested on jam and linew.
Diffstat (limited to 'test')
-rw-r--r-- | test/be_data.h5 | bin | 2288 -> 6808 bytes | |||
-rwxr-xr-x | test/cross_read.c | 118 | ||||
-rwxr-xr-x | test/gen_cross.c | 166 | ||||
-rw-r--r-- | test/le_data.h5 | bin | 2288 -> 6808 bytes | |||
-rw-r--r-- | test/vms_data.h5 | bin | 2288 -> 6808 bytes |
5 files changed, 252 insertions, 32 deletions
diff --git a/test/be_data.h5 b/test/be_data.h5 Binary files differindex f906545..1b02828 100644 --- a/test/be_data.h5 +++ b/test/be_data.h5 diff --git a/test/cross_read.c b/test/cross_read.c index 6b2badf..6d6f26c 100755 --- a/test/cross_read.c +++ b/test/cross_read.c @@ -31,10 +31,11 @@ const char *FILENAME[] = { NULL }; -#define DATASETNAME "Array" -#define NX 5 /* output buffer dimensions */ -#define NY 6 -#define RANK 2 +#define DATASETNAME "Array" +#define DATASETNAME2 "Scale_offset_data" +#define NX 6 +#define NY 6 +#define RANK 2 static int read_data(char *fname) { @@ -46,6 +47,22 @@ static int read_data(char *fname) double data_out[NX][NY]; /* output buffer */ int i, j; unsigned nerrors = 0; + const char *not_supported= " Scaleoffset filter is not enabled."; + /*const char *not_fixed= " Scaleoffset filter bug (2131) is not fixed yet.";*/ + + /* + * Open the file. + */ + if((file = H5Fopen(pathname, H5F_ACC_RDONLY, H5P_DEFAULT)) < 0) + TEST_ERROR; + + TESTING(" regular dataset"); + + /* + * Open the regular dataset. + */ + if((dataset = H5Dopen2(file, DATASETNAME, H5P_DEFAULT)) < 0) + TEST_ERROR; /* * Data and output buffer initialization. @@ -62,17 +79,74 @@ static int read_data(char *fname) * 2 3 4 5 6 7 * 3 4 5 6 7 8 * 4 5 6 7 8 9 + * 5 6 7 8 9 10 */ /* - * Open the file and the dataset. + * Get datatype and dataspace handles and then query + * dataset class, order, size, rank and dimensions. */ - if((file = H5Fopen(pathname, H5F_ACC_RDONLY, H5P_DEFAULT)) < 0) + if((dt = H5Dget_type(dataset)) < 0) /* datatype handle */ TEST_ERROR; - if((dataset = H5Dopen2(file, DATASETNAME, H5P_DEFAULT)) < 0) + if((datatype = H5Tget_native_type(dt, H5T_DIR_DEFAULT)) < 0) + TEST_ERROR; + + /* + * Read data from hyperslab in the file into the hyperslab in + * memory and display. + */ + if(H5Dread(dataset, datatype, H5S_ALL, H5S_ALL, H5P_DEFAULT, data_out) < 0) + TEST_ERROR; + + /* Check results */ + for (j=0; j<NX; j++) { + for (i=0; i<NY; i++) { + /* if (data_out[j][i] != data_in[j][i]) { */ + if (!DBL_ABS_EQUAL(data_out[j][i], data_in[j][i])) { + if (!nerrors++) { + H5_FAILED(); + printf("element [%d][%d] is %g but should have been %g\n", + j, i, data_out[j][i], data_in[j][i]); + } + } + } + } + + /* + * Close/release resources. + */ + H5Tclose(dt); + H5Tclose(datatype); + H5Dclose(dataset); + + /* Failure */ + if (nerrors) { + printf("total of %d errors out of %d elements\n", nerrors, NX*NY); + return 1; + } + + PASSED(); + + TESTING(" dataset with scale-offset filter"); + +#ifdef H5_HAVE_FILTER_SCALEOFFSET + /* + * Open the dataset with scale-offset filter. + */ + if((dataset = H5Dopen2(file, DATASETNAME2, H5P_DEFAULT)) < 0) TEST_ERROR; /* + * Data and output buffer initialization. + */ + for (j = 0; j < NX; j++) { + for (i = 0; i < NY; i++) { + data_in[j][i] = ((double)(i + j + 1))/3; + data_out[j][i] = 0; + } + } + + /* * Get datatype and dataspace handles and then query * dataset class, order, size, rank and dimensions. */ @@ -91,7 +165,8 @@ static int read_data(char *fname) /* Check results */ for (j=0; j<NX; j++) { for (i=0; i<NY; i++) { - if (data_out[j][i] != data_in[j][i]) { + /* if (data_out[j][i] != data_in[j][i]) { */ + if (!DBL_REL_EQUAL(data_out[j][i], data_in[j][i], 0.001)) { if (!nerrors++) { H5_FAILED(); printf("element [%d][%d] is %g but should have been %g\n", @@ -107,7 +182,6 @@ static int read_data(char *fname) H5Tclose(dt); H5Tclose(datatype); H5Dclose(dataset); - H5Fclose(file); /* Failure */ if (nerrors) { @@ -116,6 +190,12 @@ static int read_data(char *fname) } PASSED(); +#else /*H5_HAVE_FILTER_SCALEOFFSET*/ + SKIPPED(); + puts(not_supported); +#endif /*H5_HAVE_FILTER_SCALEOFFSET*/ + + H5Fclose(file); return 0; error: @@ -125,6 +205,20 @@ error: return 1; } + +/*------------------------------------------------------------------------- + * Function: main + * + * Purpose: Tests the basic features of Virtual File Drivers + * + * Return: Success: exit(0) + * Failure: exit(1) + * + * Programmer: Raymond Lu + * Tuesday, Sept 24, 2002 + * + *------------------------------------------------------------------------- + */ int main(void) { char filename[1024]; @@ -132,15 +226,15 @@ int main(void) h5_reset(); - TESTING("reading data created on OpenVMS"); + puts("Testing reading data created on OpenVMS"); h5_fixname(FILENAME[0], H5P_DEFAULT, filename, sizeof filename); nerrors += read_data(filename); - TESTING("reading data created on Linux"); + puts("Testing reading data created on Linux"); h5_fixname(FILENAME[1], H5P_DEFAULT, filename, sizeof filename); nerrors += read_data(filename); - TESTING("reading data created on Solaris"); + puts("Testing reading data created on Solaris"); h5_fixname(FILENAME[2], H5P_DEFAULT, filename, sizeof filename); nerrors += read_data(filename); diff --git a/test/gen_cross.c b/test/gen_cross.c index 46afc35..2505dbb 100755 --- a/test/gen_cross.c +++ b/test/gen_cross.c @@ -27,17 +27,37 @@ #include "h5test.h" #define H5FILE_NAME "data.h5" -#define DATASETNAME "Array" -#define NX 5 /* dataset dimensions */ -#define NY 6 -#define RANK 2 +#define DATASETNAME "Array" +#define DATASETNAME2 "Scale_offset_data" +#define NX 6 +#define NY 6 +#define RANK 2 +#define CHUNK0 3 +#define CHUNK1 3 +int create_normal_dset(hid_t fid, hid_t sid, hid_t tid); +int create_scale_offset_dset(hid_t fid, hid_t sid, hid_t tid); + + +/*------------------------------------------------------------------------- + * Function: create_normal_dset + * + * Purpose: Create a regular dataset of DOUBLE datatype. + * + * Return: Success: 0 + * Failure: -1 + * + * Programmer: Raymond Lu + * Some time ago + * + * Modifications: + * + *------------------------------------------------------------------------- + */ int -main (void) +create_normal_dset(hid_t fid, hid_t sid, hid_t tid) { - hid_t file, dataset; /* file and dataset handles */ - hid_t datatype, dataspace; /* handles */ - hsize_t dimsf[2]; /* dataset dimensions */ + hid_t dataset; /* file and dataset handles */ herr_t status; float data[NX][NY]; /* data to write */ int i, j; @@ -55,7 +75,121 @@ main (void) * 2 3 4 5 6 7 * 3 4 5 6 7 8 * 4 5 6 7 8 9 + * 5 6 7 8 9 10 + */ + + /* + * Create a new dataset within the file using defined dataspace and + * datatype and default dataset creation properties. + */ + dataset = H5Dcreate2(fid, DATASETNAME, tid, sid, + H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); + + /* + * Write the data to the dataset using default transfer properties. + */ + status = H5Dwrite(dataset, H5T_NATIVE_FLOAT, H5S_ALL, H5S_ALL, + H5P_DEFAULT, data); + + /* + * Close/release resources. */ + H5Dclose(dataset); + + return 0; +} + + +/*------------------------------------------------------------------------- + * Function: create_scale_offset_dset + * + * Purpose: Create a dataset of DOUBLE datatype with scale-offset filter + * + * Return: Success: 0 + * Failure: -1 + * + * Programmer: Raymond Lu + * 21 January 2011 + * + * Modifications: + * + *------------------------------------------------------------------------- + */ +int +create_scale_offset_dset(hid_t fid, hid_t sid, hid_t tid) +{ +#ifdef H5_HAVE_FILTER_SCALEOFFSET + hid_t dataset; /* file and dataset handles */ + hid_t dcpl; + herr_t status; + float data[NX][NY]; /* data to write */ + hsize_t chunk[RANK] = {CHUNK0, CHUNK1}; + int i, j; + + /* + * Data and output buffer initialization. + */ + for (j = 0; j < NX; j++) { + for (i = 0; i < NY; i++) + data[j][i] = ((float)(i + j + 1))/3; + } + + /* + * Create the dataset creation property list, add the Scale-Offset + * filter and set the chunk size. + */ + dcpl = H5Pcreate (H5P_DATASET_CREATE); + status = H5Pset_scaleoffset (dcpl, H5Z_SO_FLOAT_DSCALE, 3); + status = H5Pset_chunk (dcpl, RANK, chunk); + + /* + * Create a new dataset within the file using defined dataspace and + * datatype and default dataset creation properties. + */ + dataset = H5Dcreate2(fid, DATASETNAME2, tid, sid, + H5P_DEFAULT, dcpl, H5P_DEFAULT); + + /* + * Write the data to the dataset using default transfer properties. + */ + status = H5Dwrite(dataset, H5T_NATIVE_FLOAT, H5S_ALL, H5S_ALL, + H5P_DEFAULT, data); + + /* + * Close/release resources. + */ + H5Pclose(dcpl); + H5Dclose(dataset); + +#else /* H5_HAVE_FILTER_SCALEOFFSET */ + const char *not_supported= "Scaleoffset filter is not enabled. Can't create the dataset."; + + puts(not_supported); +#endif /* H5_HAVE_FILTER_SCALEOFFSET */ + + return 0; +} + + +/*------------------------------------------------------------------------- + * Function: main + * + * Purpose: Create a file for cross_read.c test. + * + * Return: Success: exit(0) + * Failure: exit(1) + * + * Programmer: Raymond Lu + * Some time ago + * + *------------------------------------------------------------------------- + */ +int +main (void) +{ + hid_t file; /* file and dataset handles */ + hid_t dataspace, datatype; + hsize_t dimsf[RANK]; /* * Create a new file using H5F_ACC_TRUNC access, @@ -78,25 +212,17 @@ main (void) */ datatype = H5Tcopy(H5T_NATIVE_DOUBLE); - /* - * Create a new dataset within the file using defined dataspace and - * datatype and default dataset creation properties. - */ - dataset = H5Dcreate2(file, DATASETNAME, datatype, dataspace, - H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); + /* Create a regular dataset */ + create_normal_dset(file, dataspace, datatype); - /* - * Write the data to the dataset using default transfer properties. - */ - status = H5Dwrite(dataset, H5T_NATIVE_FLOAT, H5S_ALL, H5S_ALL, - H5P_DEFAULT, data); + /* Create a dataset with scale-offset filter */ + create_scale_offset_dset(file, dataspace, datatype); /* * Close/release resources. */ H5Sclose(dataspace); H5Tclose(datatype); - H5Dclose(dataset); H5Fclose(file); return 0; diff --git a/test/le_data.h5 b/test/le_data.h5 Binary files differindex c944c9f..5276989 100644 --- a/test/le_data.h5 +++ b/test/le_data.h5 diff --git a/test/vms_data.h5 b/test/vms_data.h5 Binary files differindex 9c243ff..f3c4be5 100644 --- a/test/vms_data.h5 +++ b/test/vms_data.h5 |