summaryrefslogtreecommitdiffstats
path: root/c++
diff options
context:
space:
mode:
authorQuincey Koziol <koziol@hdfgroup.org>2002-11-20 13:03:55 (GMT)
committerQuincey Koziol <koziol@hdfgroup.org>2002-11-20 13:03:55 (GMT)
commitbe76b331a17b6267320f2eb850fe7460b7c37a8f (patch)
tree97d259dca2b3866350ddded822ab5802ea504559 /c++
parentb1ff513e6acaeb6dc4d3296e99955fcf394ac86a (diff)
downloadhdf5-be76b331a17b6267320f2eb850fe7460b7c37a8f.zip
hdf5-be76b331a17b6267320f2eb850fe7460b7c37a8f.tar.gz
hdf5-be76b331a17b6267320f2eb850fe7460b7c37a8f.tar.bz2
[svn-r6111] Purpose:
Code cleanup/new feature Description: Switched from using H5_HAVE_COMPRESSION flag in favor of H5_HAVE_FILTER_DEFLATE. Platforms tested: Tested h5committest {arabica (fortran), eirene (fortran, C++) modi4 (parallel, fortran)} FreeBSD 4.7 (sleipnir)
Diffstat (limited to 'c++')
-rw-r--r--c++/test/dsets.cpp50
1 files changed, 12 insertions, 38 deletions
diff --git a/c++/test/dsets.cpp b/c++/test/dsets.cpp
index a5b66b5..a99a4ae 100644
--- a/c++/test/dsets.cpp
+++ b/c++/test/dsets.cpp
@@ -402,11 +402,7 @@ static herr_t
test_compression(H5File& file)
{
const char *not_supported;
- not_supported = " Deflate compression is not supported.\n"
- " The zlib was not found when hdf5 was configured.";
-
- TESTING("compression (setup)");
-
+ not_supported = " Deflate compression is not enabled.";
int points[100][200];
int check[100][200];
hsize_t i, j, n;
@@ -441,16 +437,16 @@ test_compression(H5File& file)
dscreatplist.setChunk (2, chunk_size);
dscreatplist.setDeflate (6);
+ DataSet* dataset;
+
+#ifdef H5_HAVE_FILTER_DEFLATE
+ TESTING("compression (setup)");
+
/* Create the dataset */
- DataSet* dataset = new DataSet (file.createDataSet
+ dataset = new DataSet (file.createDataSet
(DSET_COMPRESS_NAME, PredType::NATIVE_INT, space1, dscreatplist));
-#ifdef H5_HAVE_COMPRESSION
PASSED();
-#else
- SKIPPED();
- cout << not_supported << endl;
-#endif
/*----------------------------------------------------------------------
* STEP 1: Read uninitialized data. It should be zero.
@@ -471,12 +467,7 @@ test_compression(H5File& file)
}
}
}
-#ifdef H5_HAVE_COMPRESSION
PASSED();
-#else
- SKIPPED();
- cout << not_supported << endl;
-#endif
/*----------------------------------------------------------------------
* STEP 2: Test compression by setting up a chunked dataset and writing
@@ -496,12 +487,7 @@ test_compression(H5File& file)
//if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, xfer, points)<0) goto error;
dataset->write ((void*) points, PredType::NATIVE_INT, DataSpace::ALL, DataSpace::ALL, xfer);
-#ifdef H5_HAVE_COMPRESSION
PASSED();
-#else
- SKIPPED();
- cout << not_supported << endl;
-#endif
/*----------------------------------------------------------------------
* STEP 3: Try to read the data we just wrote.
@@ -520,12 +506,7 @@ test_compression(H5File& file)
if (status == -1) goto error;
}
-#ifdef H5_HAVE_COMPRESSION
PASSED();
-#else
- SKIPPED();
- cout << not_supported << endl;
-#endif
/*----------------------------------------------------------------------
* STEP 4: Write new data over the top of the old data. The new data is
@@ -556,12 +537,7 @@ test_compression(H5File& file)
if (status == -1) goto error;
}
-#ifdef H5_HAVE_COMPRESSION
PASSED();
-#else
- SKIPPED();
- cout << not_supported << endl;
-#endif
/*----------------------------------------------------------------------
* STEP 5: Close the dataset and then open it and read it again. This
@@ -585,12 +561,7 @@ test_compression(H5File& file)
if (status == -1) goto error;
}
-#ifdef H5_HAVE_COMPRESSION
PASSED();
-#else
- SKIPPED();
- cout << not_supported << endl;
-#endif
/*----------------------------------------------------------------------
@@ -628,9 +599,13 @@ test_compression(H5File& file)
}
}
}
-#ifdef H5_HAVE_COMPRESSION
+
+ delete dataset;
+
PASSED();
+
#else
+ TESTING("deflate filter");
SKIPPED();
cout << not_supported << endl;
#endif
@@ -646,7 +621,6 @@ test_compression(H5File& file)
if (H5Zregister (H5Z_BOGUS, DSET_BOGUS_NAME, bogus)<0) goto error;
if (H5Pset_filter (dscreatplist.getId(), H5Z_BOGUS, 0, 0, NULL)<0) goto error;
dscreatplist.setFilter (H5Z_BOGUS, 0, 0, NULL);
- delete dataset;
DataSpace space2 (2, size, NULL);
dataset = new DataSet (file.createDataSet (DSET_BOGUS_NAME, PredType::NATIVE_INT, space2, dscreatplist));