diff options
Diffstat (limited to 'java/examples/datasets/H5Ex_D_Soint.java')
-rw-r--r-- | java/examples/datasets/H5Ex_D_Soint.java | 110 |
1 files changed, 57 insertions, 53 deletions
diff --git a/java/examples/datasets/H5Ex_D_Soint.java b/java/examples/datasets/H5Ex_D_Soint.java index dd7664f..3eb8e37 100644 --- a/java/examples/datasets/H5Ex_D_Soint.java +++ b/java/examples/datasets/H5Ex_D_Soint.java @@ -30,45 +30,46 @@ import hdf.hdf5lib.HDF5Constants; public class H5Ex_D_Soint { - private static String FILENAME = "H5Ex_D_Soint.h5"; + private static String FILENAME = "H5Ex_D_Soint.h5"; private static String DATASETNAME = "DS1"; - private static final int DIM_X = 32; - private static final int DIM_Y = 64; - private static final int CHUNK_X = 4; - private static final int CHUNK_Y = 8; - private static final int RANK = 2; - private static final int NDIMS = 2; + private static final int DIM_X = 32; + private static final int DIM_Y = 64; + private static final int CHUNK_X = 4; + private static final int CHUNK_Y = 8; + private static final int RANK = 2; + private static final int NDIMS = 2; // Values for the status of space allocation enum H5Z_filter { - H5Z_FILTER_ERROR(HDF5Constants.H5Z_FILTER_ERROR), H5Z_FILTER_NONE(HDF5Constants.H5Z_FILTER_NONE), H5Z_FILTER_DEFLATE( - HDF5Constants.H5Z_FILTER_DEFLATE), H5Z_FILTER_SHUFFLE(HDF5Constants.H5Z_FILTER_SHUFFLE), H5Z_FILTER_FLETCHER32( - HDF5Constants.H5Z_FILTER_FLETCHER32), H5Z_FILTER_SZIP(HDF5Constants.H5Z_FILTER_SZIP), H5Z_FILTER_NBIT( - HDF5Constants.H5Z_FILTER_NBIT), H5Z_FILTER_SCALEOFFSET(HDF5Constants.H5Z_FILTER_SCALEOFFSET), H5Z_FILTER_RESERVED( - HDF5Constants.H5Z_FILTER_RESERVED), H5Z_FILTER_MAX(HDF5Constants.H5Z_FILTER_MAX); + H5Z_FILTER_ERROR(HDF5Constants.H5Z_FILTER_ERROR), + H5Z_FILTER_NONE(HDF5Constants.H5Z_FILTER_NONE), + H5Z_FILTER_DEFLATE(HDF5Constants.H5Z_FILTER_DEFLATE), + H5Z_FILTER_SHUFFLE(HDF5Constants.H5Z_FILTER_SHUFFLE), + H5Z_FILTER_FLETCHER32(HDF5Constants.H5Z_FILTER_FLETCHER32), + H5Z_FILTER_SZIP(HDF5Constants.H5Z_FILTER_SZIP), + H5Z_FILTER_NBIT(HDF5Constants.H5Z_FILTER_NBIT), + H5Z_FILTER_SCALEOFFSET(HDF5Constants.H5Z_FILTER_SCALEOFFSET), + H5Z_FILTER_RESERVED(HDF5Constants.H5Z_FILTER_RESERVED), + H5Z_FILTER_MAX(HDF5Constants.H5Z_FILTER_MAX); private static final Map<Integer, H5Z_filter> lookup = new HashMap<Integer, H5Z_filter>(); - static { + static + { for (H5Z_filter s : EnumSet.allOf(H5Z_filter.class)) lookup.put(s.getCode(), s); } private int code; - H5Z_filter(int layout_type) { - this.code = layout_type; - } + H5Z_filter(int layout_type) { this.code = layout_type; } - public int getCode() { - return this.code; - } + public int getCode() { return this.code; } - public static H5Z_filter get(int code) { - return lookup.get(code); - } + public static H5Z_filter get(int code) { return lookup.get(code); } } - private static boolean checkScaleoffsetFilter() { + private static boolean checkScaleoffsetFilter() + { try { int available = H5.H5Zfilter_avail(HDF5Constants.H5Z_FILTER_SCALEOFFSET); if (available == 0) { @@ -82,8 +83,8 @@ public class H5Ex_D_Soint { try { int filter_info = H5.H5Zget_filter_info(HDF5Constants.H5Z_FILTER_SCALEOFFSET); - if (((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_ENCODE_ENABLED) == 0) - || ((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_DECODE_ENABLED) == 0)) { + if (((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_ENCODE_ENABLED) == 0) || + ((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_DECODE_ENABLED) == 0)) { System.out.println("Scale-Offset filter not available for encoding and decoding."); return false; } @@ -94,13 +95,14 @@ public class H5Ex_D_Soint { return true; } - private static void writeData() { - long file_id = HDF5Constants.H5I_INVALID_HID; + private static void writeData() + { + long file_id = HDF5Constants.H5I_INVALID_HID; long filespace_id = HDF5Constants.H5I_INVALID_HID; - long dataset_id = HDF5Constants.H5I_INVALID_HID; - long dcpl_id = HDF5Constants.H5I_INVALID_HID; - long[] dims = { DIM_X, DIM_Y }; - long[] chunk_dims = { CHUNK_X, CHUNK_Y }; + long dataset_id = HDF5Constants.H5I_INVALID_HID; + long dcpl_id = HDF5Constants.H5I_INVALID_HID; + long[] dims = {DIM_X, DIM_Y}; + long[] chunk_dims = {CHUNK_X, CHUNK_Y}; int[][] dset_data = new int[DIM_X][DIM_Y]; // Initialize data. @@ -111,7 +113,7 @@ public class H5Ex_D_Soint { // Create a new file using the default properties. try { file_id = H5.H5Fcreate(FILENAME, HDF5Constants.H5F_ACC_TRUNC, HDF5Constants.H5P_DEFAULT, - HDF5Constants.H5P_DEFAULT); + HDF5Constants.H5P_DEFAULT); } catch (Exception e) { e.printStackTrace(); @@ -130,7 +132,8 @@ public class H5Ex_D_Soint { try { dcpl_id = H5.H5Pcreate(HDF5Constants.H5P_DATASET_CREATE); if (dcpl_id >= 0) { - H5.H5Pset_scaleoffset(dcpl_id, HDF5Constants.H5Z_SO_INT, HDF5Constants.H5Z_SO_INT_MINBITS_DEFAULT); + H5.H5Pset_scaleoffset(dcpl_id, HDF5Constants.H5Z_SO_INT, + HDF5Constants.H5Z_SO_INT_MINBITS_DEFAULT); H5.H5Pset_chunk(dcpl_id, NDIMS, chunk_dims); } } @@ -142,7 +145,7 @@ public class H5Ex_D_Soint { try { if ((file_id >= 0) && (filespace_id >= 0) && (dcpl_id >= 0)) dataset_id = H5.H5Dcreate(file_id, DATASETNAME, HDF5Constants.H5T_STD_I32LE, filespace_id, - HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT); + HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT); } catch (Exception e) { e.printStackTrace(); @@ -151,8 +154,8 @@ public class H5Ex_D_Soint { // Write the data to the dataset. try { if (dataset_id >= 0) - H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, HDF5Constants.H5S_ALL, - HDF5Constants.H5P_DEFAULT, dset_data); + H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, + HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data); } catch (Exception e) { e.printStackTrace(); @@ -193,10 +196,11 @@ public class H5Ex_D_Soint { } } - private static void readData() { - long file_id = HDF5Constants.H5I_INVALID_HID; - long dataset_id = HDF5Constants.H5I_INVALID_HID; - long dcpl_id = HDF5Constants.H5I_INVALID_HID; + private static void readData() + { + long file_id = HDF5Constants.H5I_INVALID_HID; + long dataset_id = HDF5Constants.H5I_INVALID_HID; + long dcpl_id = HDF5Constants.H5I_INVALID_HID; int[][] dset_data = new int[DIM_X][DIM_Y]; // Open file using the default properties. @@ -229,15 +233,15 @@ public class H5Ex_D_Soint { try { if (dcpl_id >= 0) { // Java lib requires a valid filter_name object and cd_values - int[] flags = { 0 }; - long[] cd_nelmts = { 1 }; - int[] cd_values = { 0 }; - String[] filter_name = { "" }; - int[] filter_config = { 0 }; - int filter_type = -1; - - filter_type = H5 - .H5Pget_filter(dcpl_id, 0, flags, cd_nelmts, cd_values, 120, filter_name, filter_config); + int[] flags = {0}; + long[] cd_nelmts = {1}; + int[] cd_values = {0}; + String[] filter_name = {""}; + int[] filter_config = {0}; + int filter_type = -1; + + filter_type = H5.H5Pget_filter(dcpl_id, 0, flags, cd_nelmts, cd_values, 120, filter_name, + filter_config); System.out.print("Filter type is: "); switch (H5Z_filter.get(filter_type)) { case H5Z_FILTER_DEFLATE: @@ -271,8 +275,8 @@ public class H5Ex_D_Soint { // Read the data using the default properties. try { if (dataset_id >= 0) - H5.H5Dread(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, HDF5Constants.H5S_ALL, - HDF5Constants.H5P_DEFAULT, dset_data); + H5.H5Dread(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, + HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data); } catch (Exception e) { e.printStackTrace(); @@ -316,7 +320,8 @@ public class H5Ex_D_Soint { } } - public static void main(String[] args) { + public static void main(String[] args) + { // Check if Scale-Offset compression is available and can be used // for both compression and decompression. Normally we do not @@ -328,5 +333,4 @@ public class H5Ex_D_Soint { H5Ex_D_Soint.readData(); } } - } |