diff options
author | Allen Byrne <50328838+byrnHDF@users.noreply.github.com> | 2022-04-19 18:08:09 (GMT) |
---|---|---|
committer | GitHub <noreply@github.com> | 2022-04-19 18:08:09 (GMT) |
commit | 03c3a5469542688a52d5a20242d2334408f8ba33 (patch) | |
tree | 46e30ea4aa99b68279fe80c9ad56b1a6bbfcfe7c /java/examples/datasets/H5Ex_D_Checksum.java | |
parent | 32ef796e470da3e4de364d8dc469b03f5f6fafdc (diff) | |
download | hdf5-03c3a5469542688a52d5a20242d2334408f8ba33.zip hdf5-03c3a5469542688a52d5a20242d2334408f8ba33.tar.gz hdf5-03c3a5469542688a52d5a20242d2334408f8ba33.tar.bz2 |
Develop clang format java (#1653)
Diffstat (limited to 'java/examples/datasets/H5Ex_D_Checksum.java')
-rw-r--r-- | java/examples/datasets/H5Ex_D_Checksum.java | 100 |
1 files changed, 53 insertions, 47 deletions
diff --git a/java/examples/datasets/H5Ex_D_Checksum.java b/java/examples/datasets/H5Ex_D_Checksum.java index 781dd68..7b01176 100644 --- a/java/examples/datasets/H5Ex_D_Checksum.java +++ b/java/examples/datasets/H5Ex_D_Checksum.java @@ -30,42 +30,46 @@ import hdf.hdf5lib.H5; import hdf.hdf5lib.HDF5Constants; public class H5Ex_D_Checksum { - private static String FILENAME = "H5Ex_D_Checksum.h5"; + private static String FILENAME = "H5Ex_D_Checksum.h5"; private static String DATASETNAME = "DS1"; - private static final int DIM_X = 32; - private static final int DIM_Y = 64; - private static final int CHUNK_X = 4; - private static final int CHUNK_Y = 8; - private static final int RANK = 2; - private static final int NDIMS = 2; + private static final int DIM_X = 32; + private static final int DIM_Y = 64; + private static final int CHUNK_X = 4; + private static final int CHUNK_Y = 8; + private static final int RANK = 2; + private static final int NDIMS = 2; // Values for the status of space allocation enum H5Z_filter { - H5Z_FILTER_ERROR(-1), H5Z_FILTER_NONE(0), H5Z_FILTER_DEFLATE(1), H5Z_FILTER_SHUFFLE(2), H5Z_FILTER_FLETCHER32(3), H5Z_FILTER_SZIP( - 4), H5Z_FILTER_NBIT(5), H5Z_FILTER_SCALEOFFSET(6), H5Z_FILTER_RESERVED(256), H5Z_FILTER_MAX(65535); + H5Z_FILTER_ERROR(-1), + H5Z_FILTER_NONE(0), + H5Z_FILTER_DEFLATE(1), + H5Z_FILTER_SHUFFLE(2), + H5Z_FILTER_FLETCHER32(3), + H5Z_FILTER_SZIP(4), + H5Z_FILTER_NBIT(5), + H5Z_FILTER_SCALEOFFSET(6), + H5Z_FILTER_RESERVED(256), + H5Z_FILTER_MAX(65535); private static final Map<Integer, H5Z_filter> lookup = new HashMap<Integer, H5Z_filter>(); - static { + static + { for (H5Z_filter s : EnumSet.allOf(H5Z_filter.class)) lookup.put(s.getCode(), s); } private int code; - H5Z_filter(int layout_type) { - this.code = layout_type; - } + H5Z_filter(int layout_type) { this.code = layout_type; } - public int getCode() { - return this.code; - } + public int getCode() { return this.code; } - public static H5Z_filter get(int code) { - return lookup.get(code); - } + public static H5Z_filter get(int code) { return lookup.get(code); } } - private static boolean checkFletcher32Filter() { + private static boolean checkFletcher32Filter() + { try { int available = H5.H5Zfilter_avail(H5Z_filter.H5Z_FILTER_FLETCHER32.getCode()); if (available == 0) { @@ -79,8 +83,8 @@ public class H5Ex_D_Checksum { try { int filter_info = H5.H5Zget_filter_info(HDF5Constants.H5Z_FILTER_FLETCHER32); - if (((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_ENCODE_ENABLED) == 0) - || ((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_DECODE_ENABLED) == 0)) { + if (((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_ENCODE_ENABLED) == 0) || + ((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_DECODE_ENABLED) == 0)) { System.out.println("N-Bit filter not available for encoding and decoding."); return false; } @@ -91,13 +95,14 @@ public class H5Ex_D_Checksum { return true; } - private static void writeChecksum() { - long file_id = HDF5Constants.H5I_INVALID_HID; + private static void writeChecksum() + { + long file_id = HDF5Constants.H5I_INVALID_HID; long filespace_id = HDF5Constants.H5I_INVALID_HID; - long dataset_id = HDF5Constants.H5I_INVALID_HID; - long dcpl_id = HDF5Constants.H5I_INVALID_HID; - long[] dims = { DIM_X, DIM_Y }; - long[] chunk_dims = { CHUNK_X, CHUNK_Y }; + long dataset_id = HDF5Constants.H5I_INVALID_HID; + long dcpl_id = HDF5Constants.H5I_INVALID_HID; + long[] dims = {DIM_X, DIM_Y}; + long[] chunk_dims = {CHUNK_X, CHUNK_Y}; int[][] dset_data = new int[DIM_X][DIM_Y]; // Initialize data. @@ -108,7 +113,7 @@ public class H5Ex_D_Checksum { // Create a new file using default properties. try { file_id = H5.H5Fcreate(FILENAME, HDF5Constants.H5F_ACC_TRUNC, HDF5Constants.H5P_DEFAULT, - HDF5Constants.H5P_DEFAULT); + HDF5Constants.H5P_DEFAULT); } catch (Exception e) { e.printStackTrace(); @@ -140,7 +145,7 @@ public class H5Ex_D_Checksum { try { if ((file_id >= 0) && (filespace_id >= 0) && (dcpl_id >= 0)) dataset_id = H5.H5Dcreate(file_id, DATASETNAME, HDF5Constants.H5T_STD_I32LE, filespace_id, - HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT); + HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT); } catch (Exception e) { e.printStackTrace(); @@ -149,8 +154,8 @@ public class H5Ex_D_Checksum { // Write the data to the dataset. try { if (dataset_id >= 0) - H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, HDF5Constants.H5S_ALL, - HDF5Constants.H5P_DEFAULT, dset_data); + H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, + HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data); } catch (Exception e) { e.printStackTrace(); @@ -191,10 +196,11 @@ public class H5Ex_D_Checksum { } } - private static void readChecksum() { - long file_id = HDF5Constants.H5I_INVALID_HID; - long dataset_id = HDF5Constants.H5I_INVALID_HID; - long dcpl_id = HDF5Constants.H5I_INVALID_HID; + private static void readChecksum() + { + long file_id = HDF5Constants.H5I_INVALID_HID; + long dataset_id = HDF5Constants.H5I_INVALID_HID; + long dcpl_id = HDF5Constants.H5I_INVALID_HID; int[][] dset_data = new int[DIM_X][DIM_Y]; // Open an existing file. @@ -228,14 +234,14 @@ public class H5Ex_D_Checksum { try { if (dcpl_id >= 0) { // Java lib requires a valid filter_name object and cd_values - int[] flags = { 0 }; - long[] cd_nelmts = { 1 }; - int[] cd_values = { 0 }; - String[] filter_name = { "" }; - int[] filter_config = { 0 }; - int filter_type = -1; - filter_type = H5 - .H5Pget_filter(dcpl_id, 0, flags, cd_nelmts, cd_values, 120, filter_name, filter_config); + int[] flags = {0}; + long[] cd_nelmts = {1}; + int[] cd_values = {0}; + String[] filter_name = {""}; + int[] filter_config = {0}; + int filter_type = -1; + filter_type = H5.H5Pget_filter(dcpl_id, 0, flags, cd_nelmts, cd_values, 120, filter_name, + filter_config); System.out.print("Filter type is: "); switch (H5Z_filter.get(filter_type)) { case H5Z_FILTER_DEFLATE: @@ -264,7 +270,7 @@ public class H5Ex_D_Checksum { try { if (dataset_id >= 0) { int status = H5.H5Dread(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL, - HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data); + HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data); // Check if the read was successful. Normally we do not perform // error checking in these examples for the sake of clarity, but in // this case we will make an exception because this is how the @@ -328,7 +334,8 @@ public class H5Ex_D_Checksum { } } - public static void main(String[] args) { + public static void main(String[] args) + { // Check if the Fletcher32 filter is available and can be used for // both encoding and decoding. Normally we do not perform error // checking in these examples for the sake of clarity, but in this @@ -340,5 +347,4 @@ public class H5Ex_D_Checksum { H5Ex_D_Checksum.readChecksum(); } } - } |