summaryrefslogtreecommitdiffstats
path: root/java/test/TestH5PL.java
diff options
context:
space:
mode:
authorvchoi-hdfgroup <55293060+vchoi-hdfgroup@users.noreply.github.com>2022-04-21 17:46:54 (GMT)
committerGitHub <noreply@github.com>2022-04-21 17:46:54 (GMT)
commitff8f512871b72e80c6d602d1539ae70072e18abb (patch)
tree491c1f08a68d92e400d939d96fefdc7bfef57ddc /java/test/TestH5PL.java
parentff7fe9e9af404591d3d50f3b33d190679529bb10 (diff)
parent2f86cba3ec7b9f46b84e8cd239a1fee242c15049 (diff)
downloadhdf5-ff8f512871b72e80c6d602d1539ae70072e18abb.zip
hdf5-ff8f512871b72e80c6d602d1539ae70072e18abb.tar.gz
hdf5-ff8f512871b72e80c6d602d1539ae70072e18abb.tar.bz2
Merge pull request #62 from vchoi-hdfgroup/feature/vfd_swmr
Feature/vfd swmr
Diffstat (limited to 'java/test/TestH5PL.java')
-rw-r--r--java/test/TestH5PL.java142
1 files changed, 87 insertions, 55 deletions
diff --git a/java/test/TestH5PL.java b/java/test/TestH5PL.java
index d44cc0b..f6440cd 100644
--- a/java/test/TestH5PL.java
+++ b/java/test/TestH5PL.java
@@ -14,6 +14,7 @@ package test;
import static org.junit.Assert.assertTrue;
import static org.junit.Assert.fail;
+
import hdf.hdf5lib.H5;
import hdf.hdf5lib.HDF5Constants;
import hdf.hdf5lib.exceptions.HDF5LibraryException;
@@ -26,39 +27,45 @@ import org.junit.Test;
import org.junit.rules.TestName;
public class TestH5PL {
- @Rule public TestName testname = new TestName();
- private static String FILENAME = "h5_dlopenChunk.h5";
- private static String DATASETNAME = "DS1";
- private static final int DIM_X = 6;
- private static final int DIM_Y = 8;
- private static final int CHUNK_X = 4;
- private static final int CHUNK_Y = 4;
- private static final int RANK = 2;
- private static final int NDIMS = 2;
+ @Rule
+ public TestName testname = new TestName();
+ private static String FILENAME = "h5_dlopenChunk.h5";
+ private static String DATASETNAME = "DS1";
+ private static final int DIM_X = 6;
+ private static final int DIM_Y = 8;
+ private static final int CHUNK_X = 4;
+ private static final int CHUNK_Y = 4;
+ private static final int RANK = 2;
+ private static final int NDIMS = 2;
private static final int H5Z_FILTER_DYNLIB4 = 260;
@Before
- public void checkOpenIDs() {
- assertTrue("H5 open ids is 0",H5.getOpenIDCount()==0);
+ public void checkOpenIDs()
+ {
+ assertTrue("H5 open ids is 0", H5.getOpenIDCount() == 0);
System.out.print(testname.getMethodName());
}
@After
- public void nextTestName() {
+ public void nextTestName()
+ {
System.out.println();
}
@Test
- public void TestH5PLplugins() {
+ public void TestH5PLplugins()
+ {
try {
int plugin_flags = H5.H5PLget_loading_state();
- assertTrue("H5.H5PLget_loading_state: "+plugin_flags, plugin_flags == HDF5Constants.H5PL_ALL_PLUGIN);
+ assertTrue("H5.H5PLget_loading_state: " + plugin_flags,
+ plugin_flags == HDF5Constants.H5PL_ALL_PLUGIN);
int new_setting = plugin_flags & ~HDF5Constants.H5PL_FILTER_PLUGIN;
- H5.H5PLset_loading_state (new_setting);
+ H5.H5PLset_loading_state(new_setting);
int changed_flags = H5.H5PLget_loading_state();
- assertTrue("H5.H5PLget_loading_state: "+changed_flags, changed_flags == new_setting);
- H5.H5PLset_loading_state (plugin_flags);
+ assertTrue("H5.H5PLget_loading_state: " + changed_flags, changed_flags == new_setting);
+ H5.H5PLset_loading_state(plugin_flags);
changed_flags = H5.H5PLget_loading_state();
- assertTrue("H5.H5PLget_loading_state: "+changed_flags, changed_flags == HDF5Constants.H5PL_ALL_PLUGIN);
+ assertTrue("H5.H5PLget_loading_state: " + changed_flags,
+ changed_flags == HDF5Constants.H5PL_ALL_PLUGIN);
}
catch (Throwable err) {
err.printStackTrace();
@@ -67,15 +74,16 @@ public class TestH5PL {
}
@Test
- public void TestH5PLpaths() {
+ public void TestH5PLpaths()
+ {
try {
// Get the original number of paths
int nStartPaths = H5.H5PLsize();
- int nPaths; /* # paths from H5PLSize() */
- int nTruePaths = nStartPaths; /* What the # paths should be */
- int index; /* Path table index */
- String path; /* Path from H5PLget() */
+ int nPaths; /* # paths from H5PLSize() */
+ int nTruePaths = nStartPaths; /* What the # paths should be */
+ int index; /* Path table index */
+ String path; /* Path from H5PLget() */
// APPEND a path and ensure it was added correctly
String pathAppend = "path_append";
@@ -86,8 +94,9 @@ public class TestH5PL {
assertTrue("# paths should be " + nTruePaths + " but was " + nPaths, nTruePaths == nPaths);
index = nTruePaths - 1;
- path = H5.H5PLget(index);
- assertTrue("Path should be " + pathAppend + " but was " + path, path.compareToIgnoreCase(pathAppend) == 0);
+ path = H5.H5PLget(index);
+ assertTrue("Path should be " + pathAppend + " but was " + path,
+ path.compareToIgnoreCase(pathAppend) == 0);
// PREPEND a path and ensure it was added correctly
String pathPrepend = "path_prepend";
@@ -98,13 +107,14 @@ public class TestH5PL {
assertTrue("# paths should be " + nTruePaths + " but was " + nPaths, nTruePaths == nPaths);
index = 0;
- path = H5.H5PLget(index);
- assertTrue("Path should be " + pathPrepend + " but was " + path, path.compareToIgnoreCase(pathPrepend) == 0);
+ path = H5.H5PLget(index);
+ assertTrue("Path should be " + pathPrepend + " but was " + path,
+ path.compareToIgnoreCase(pathPrepend) == 0);
// INSERT a path and ensure it was added correctly
// Inserting at the index == # of start paths ensures we're in the middle
String pathInsert = "path_insert";
- index = nStartPaths;
+ index = nStartPaths;
H5.H5PLinsert(pathInsert, index);
nPaths = H5.H5PLsize();
@@ -112,22 +122,24 @@ public class TestH5PL {
assertTrue("# paths should be " + nTruePaths + " but was " + nPaths, nTruePaths == nPaths);
path = H5.H5PLget(index);
- assertTrue("Path should be " + pathInsert + " but was " + path, path.compareToIgnoreCase(pathInsert) == 0);
+ assertTrue("Path should be " + pathInsert + " but was " + path,
+ path.compareToIgnoreCase(pathInsert) == 0);
// REPLACE the path we just added and ensure it updated correctly
String pathReplace = "path_replace";
- index = nStartPaths;
+ index = nStartPaths;
H5.H5PLreplace(pathReplace, index);
nPaths = H5.H5PLsize();
assertTrue("# paths should be " + nTruePaths + " but was " + nPaths, nTruePaths == nPaths);
path = H5.H5PLget(index);
- assertTrue("Path should be " + pathReplace + " but was " + path, path.compareToIgnoreCase(pathReplace) == 0);
+ assertTrue("Path should be " + pathReplace + " but was " + path,
+ path.compareToIgnoreCase(pathReplace) == 0);
// REMOVE the path we just replaced and check that the table was compacted
// The (index+1) path should move down to fill the space when the path is removed.
- index = nStartPaths;
+ index = nStartPaths;
String pathRemove = H5.H5PLget(index + 1);
H5.H5PLremove(index);
@@ -136,7 +148,8 @@ public class TestH5PL {
assertTrue("# paths should be " + nTruePaths + " but was " + nPaths, nTruePaths == nPaths);
path = H5.H5PLget(index);
- assertTrue("Path should be " + pathRemove + " but was " + path, path.compareToIgnoreCase(pathRemove) == 0);
+ assertTrue("Path should be " + pathRemove + " but was " + path,
+ path.compareToIgnoreCase(pathRemove) == 0);
}
catch (Throwable err) {
err.printStackTrace();
@@ -145,22 +158,23 @@ public class TestH5PL {
}
@Ignore
- public void TestH5PLdlopen() {
- long file_id = HDF5Constants.H5I_INVALID_HID;
+ public void TestH5PLdlopen()
+ {
+ long file_id = HDF5Constants.H5I_INVALID_HID;
long filespace_id = HDF5Constants.H5I_INVALID_HID;
- long dataset_id = HDF5Constants.H5I_INVALID_HID;
- long fapl_id = HDF5Constants.H5I_INVALID_HID;
- long dcpl_id = HDF5Constants.H5I_INVALID_HID;
+ long dataset_id = HDF5Constants.H5I_INVALID_HID;
+ long fapl_id = HDF5Constants.H5I_INVALID_HID;
+ long dcpl_id = HDF5Constants.H5I_INVALID_HID;
try {
- int[] cd_values = {9, 0, 0, 0};
- int[] libversion = {0, 0, 0};
- long[] dims = { DIM_X, DIM_Y };
- long[] chunk_dims = { CHUNK_X, CHUNK_Y };
- int[][] dset_data = new int[DIM_X][DIM_Y];
- int[] mdc_nelmts = {0};
+ int[] cd_values = {9, 0, 0, 0};
+ int[] libversion = {0, 0, 0};
+ long[] dims = {DIM_X, DIM_Y};
+ long[] chunk_dims = {CHUNK_X, CHUNK_Y};
+ int[][] dset_data = new int[DIM_X][DIM_Y];
+ int[] mdc_nelmts = {0};
long[] rdcc_nelmts = {0};
long[] rdcc_nbytes = {0};
- double[] rdcc_w0 = {0};
+ double[] rdcc_w0 = {0};
// Initialize data to "1", to make it easier to see the selections.
for (int indx = 0; indx < DIM_X; indx++)
@@ -170,7 +184,7 @@ public class TestH5PL {
// Create a new file using default properties.
try {
file_id = H5.H5Fcreate(FILENAME, HDF5Constants.H5F_ACC_TRUNC, HDF5Constants.H5P_DEFAULT,
- HDF5Constants.H5P_DEFAULT);
+ HDF5Constants.H5P_DEFAULT);
}
catch (Exception e) {
e.printStackTrace();
@@ -212,7 +226,8 @@ public class TestH5PL {
cd_values[2] = libversion[1];
cd_values[3] = libversion[2];
if (dcpl_id >= 0)
- H5.H5Pset_filter(dcpl_id, H5Z_FILTER_DYNLIB4, HDF5Constants.H5Z_FLAG_MANDATORY, 4, cd_values);
+ H5.H5Pset_filter(dcpl_id, H5Z_FILTER_DYNLIB4, HDF5Constants.H5Z_FLAG_MANDATORY, 4,
+ cd_values);
}
catch (Exception e) {
e.printStackTrace();
@@ -222,8 +237,9 @@ public class TestH5PL {
// Create the chunked dataset.
try {
if ((file_id >= 0) && (filespace_id >= 0) && (dcpl_id >= 0))
- dataset_id = H5.H5Dcreate(file_id, DATASETNAME, HDF5Constants.H5T_NATIVE_INT, filespace_id,
- HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT);
+ dataset_id =
+ H5.H5Dcreate(file_id, DATASETNAME, HDF5Constants.H5T_NATIVE_INT, filespace_id,
+ HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT);
}
catch (Exception e) {
e.printStackTrace();
@@ -232,8 +248,8 @@ public class TestH5PL {
try {
if (dataset_id >= 0)
- H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, HDF5Constants.H5S_ALL,
- HDF5Constants.H5S_ALL, dset_data);
+ H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL,
+ HDF5Constants.H5S_ALL, HDF5Constants.H5S_ALL, dset_data);
}
catch (Exception e) {
e.printStackTrace();
@@ -247,13 +263,29 @@ public class TestH5PL {
finally {
// End access to the dataset and release resources used by it.
if (dcpl_id >= 0)
- try {H5.H5Pclose_class(dcpl_id);} catch (Throwable err) {}
+ try {
+ H5.H5Pclose_class(dcpl_id);
+ }
+ catch (Throwable err) {
+ }
if (dataset_id >= 0)
- try {H5.H5Dclose(dataset_id);} catch (Throwable err) {}
+ try {
+ H5.H5Dclose(dataset_id);
+ }
+ catch (Throwable err) {
+ }
if (filespace_id >= 0)
- try {H5.H5Sclose(filespace_id);} catch (Throwable err) {}
+ try {
+ H5.H5Sclose(filespace_id);
+ }
+ catch (Throwable err) {
+ }
if (file_id >= 0)
- try {H5.H5Fclose(file_id);} catch (Throwable err) {}
+ try {
+ H5.H5Fclose(file_id);
+ }
+ catch (Throwable err) {
+ }
}
}
}