summaryrefslogtreecommitdiffstats
path: root/java/examples/datasets/H5Ex_D_Soint.java
diff options
context:
space:
mode:
authorAllen Byrne <50328838+byrnHDF@users.noreply.github.com>2023-12-14 18:50:30 (GMT)
committerGitHub <noreply@github.com>2023-12-14 18:50:30 (GMT)
commit6f2d22d86e6a265535efe22a8bb3872b3b32a52b (patch)
treef79e0d9778b13466282c159dfec67d31faba5b8d /java/examples/datasets/H5Ex_D_Soint.java
parent184445f4a45c602887ca0624c5295a7b6fea2d21 (diff)
downloadhdf5-6f2d22d86e6a265535efe22a8bb3872b3b32a52b.zip
hdf5-6f2d22d86e6a265535efe22a8bb3872b3b32a52b.tar.gz
hdf5-6f2d22d86e6a265535efe22a8bb3872b3b32a52b.tar.bz2
Add variable option syncing for examples (#3885)
Diffstat (limited to 'java/examples/datasets/H5Ex_D_Soint.java')
-rw-r--r--java/examples/datasets/H5Ex_D_Soint.java336
1 files changed, 0 insertions, 336 deletions
diff --git a/java/examples/datasets/H5Ex_D_Soint.java b/java/examples/datasets/H5Ex_D_Soint.java
deleted file mode 100644
index 3eb8e37..0000000
--- a/java/examples/datasets/H5Ex_D_Soint.java
+++ /dev/null
@@ -1,336 +0,0 @@
-/* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *
- * Copyright by The HDF Group. *
- * All rights reserved. *
- * *
- * This file is part of HDF5. The full HDF5 copyright notice, including *
- * terms governing use, modification, and redistribution, is contained in *
- * the COPYING file, which can be found at the root of the source code *
- * distribution tree, or in https://www.hdfgroup.org/licenses. *
- * If you do not have access to either file, you may request a copy from *
- * help@hdfgroup.org. *
- * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
-
-/************************************************************
- This example shows how to read and write data to a dataset
- using the Scale-Offset filter. The program first checks
- if the Scale-Offset filter is available, then if it is it
- writes integers to a dataset using Scale-Offset, then
- closes the file Next, it reopens the file, reads back the
- data, and outputs the type of filter and the maximum value
- in the dataset to the screen.
- ************************************************************/
-package examples.datasets;
-
-import java.util.EnumSet;
-import java.util.HashMap;
-import java.util.Map;
-
-import hdf.hdf5lib.H5;
-import hdf.hdf5lib.HDF5Constants;
-
-public class H5Ex_D_Soint {
-
- private static String FILENAME = "H5Ex_D_Soint.h5";
- private static String DATASETNAME = "DS1";
- private static final int DIM_X = 32;
- private static final int DIM_Y = 64;
- private static final int CHUNK_X = 4;
- private static final int CHUNK_Y = 8;
- private static final int RANK = 2;
- private static final int NDIMS = 2;
-
- // Values for the status of space allocation
- enum H5Z_filter {
- H5Z_FILTER_ERROR(HDF5Constants.H5Z_FILTER_ERROR),
- H5Z_FILTER_NONE(HDF5Constants.H5Z_FILTER_NONE),
- H5Z_FILTER_DEFLATE(HDF5Constants.H5Z_FILTER_DEFLATE),
- H5Z_FILTER_SHUFFLE(HDF5Constants.H5Z_FILTER_SHUFFLE),
- H5Z_FILTER_FLETCHER32(HDF5Constants.H5Z_FILTER_FLETCHER32),
- H5Z_FILTER_SZIP(HDF5Constants.H5Z_FILTER_SZIP),
- H5Z_FILTER_NBIT(HDF5Constants.H5Z_FILTER_NBIT),
- H5Z_FILTER_SCALEOFFSET(HDF5Constants.H5Z_FILTER_SCALEOFFSET),
- H5Z_FILTER_RESERVED(HDF5Constants.H5Z_FILTER_RESERVED),
- H5Z_FILTER_MAX(HDF5Constants.H5Z_FILTER_MAX);
- private static final Map<Integer, H5Z_filter> lookup = new HashMap<Integer, H5Z_filter>();
-
- static
- {
- for (H5Z_filter s : EnumSet.allOf(H5Z_filter.class))
- lookup.put(s.getCode(), s);
- }
-
- private int code;
-
- H5Z_filter(int layout_type) { this.code = layout_type; }
-
- public int getCode() { return this.code; }
-
- public static H5Z_filter get(int code) { return lookup.get(code); }
- }
-
- private static boolean checkScaleoffsetFilter()
- {
- try {
- int available = H5.H5Zfilter_avail(HDF5Constants.H5Z_FILTER_SCALEOFFSET);
- if (available == 0) {
- System.out.println("Scale-Offset filter not available.");
- return false;
- }
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- try {
- int filter_info = H5.H5Zget_filter_info(HDF5Constants.H5Z_FILTER_SCALEOFFSET);
- if (((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_ENCODE_ENABLED) == 0) ||
- ((filter_info & HDF5Constants.H5Z_FILTER_CONFIG_DECODE_ENABLED) == 0)) {
- System.out.println("Scale-Offset filter not available for encoding and decoding.");
- return false;
- }
- }
- catch (Exception e) {
- e.printStackTrace();
- }
- return true;
- }
-
- private static void writeData()
- {
- long file_id = HDF5Constants.H5I_INVALID_HID;
- long filespace_id = HDF5Constants.H5I_INVALID_HID;
- long dataset_id = HDF5Constants.H5I_INVALID_HID;
- long dcpl_id = HDF5Constants.H5I_INVALID_HID;
- long[] dims = {DIM_X, DIM_Y};
- long[] chunk_dims = {CHUNK_X, CHUNK_Y};
- int[][] dset_data = new int[DIM_X][DIM_Y];
-
- // Initialize data.
- for (int indx = 0; indx < DIM_X; indx++)
- for (int jndx = 0; jndx < DIM_Y; jndx++)
- dset_data[indx][jndx] = indx * jndx - jndx;
-
- // Create a new file using the default properties.
- try {
- file_id = H5.H5Fcreate(FILENAME, HDF5Constants.H5F_ACC_TRUNC, HDF5Constants.H5P_DEFAULT,
- HDF5Constants.H5P_DEFAULT);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Create dataspace. Setting maximum size to NULL sets the maximum size to be the current size.
- try {
- filespace_id = H5.H5Screate_simple(RANK, dims, null);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Create the dataset creation property list, add the Scale-Offset
- // filter and set the chunk size.
- try {
- dcpl_id = H5.H5Pcreate(HDF5Constants.H5P_DATASET_CREATE);
- if (dcpl_id >= 0) {
- H5.H5Pset_scaleoffset(dcpl_id, HDF5Constants.H5Z_SO_INT,
- HDF5Constants.H5Z_SO_INT_MINBITS_DEFAULT);
- H5.H5Pset_chunk(dcpl_id, NDIMS, chunk_dims);
- }
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Create the dataset.
- try {
- if ((file_id >= 0) && (filespace_id >= 0) && (dcpl_id >= 0))
- dataset_id = H5.H5Dcreate(file_id, DATASETNAME, HDF5Constants.H5T_STD_I32LE, filespace_id,
- HDF5Constants.H5P_DEFAULT, dcpl_id, HDF5Constants.H5P_DEFAULT);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Write the data to the dataset.
- try {
- if (dataset_id >= 0)
- H5.H5Dwrite(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL,
- HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Close and release resources.
- try {
- if (dcpl_id >= 0)
- H5.H5Pclose(dcpl_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- try {
- if (dataset_id >= 0)
- H5.H5Dclose(dataset_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- try {
- if (filespace_id >= 0)
- H5.H5Sclose(filespace_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Close file
- try {
- if (file_id >= 0)
- H5.H5Fclose(file_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
- }
-
- private static void readData()
- {
- long file_id = HDF5Constants.H5I_INVALID_HID;
- long dataset_id = HDF5Constants.H5I_INVALID_HID;
- long dcpl_id = HDF5Constants.H5I_INVALID_HID;
- int[][] dset_data = new int[DIM_X][DIM_Y];
-
- // Open file using the default properties.
- try {
- file_id = H5.H5Fopen(FILENAME, HDF5Constants.H5F_ACC_RDONLY, HDF5Constants.H5P_DEFAULT);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
- // Open dataset using the default properties.
- try {
- if (file_id >= 0)
- dataset_id = H5.H5Dopen(file_id, DATASETNAME, HDF5Constants.H5P_DEFAULT);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Retrieve dataset creation property list.
- try {
- if (dataset_id >= 0)
- dcpl_id = H5.H5Dget_create_plist(dataset_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Retrieve and print the filter type. Here we only retrieve the
- // first filter because we know that we only added one filter.
- try {
- if (dcpl_id >= 0) {
- // Java lib requires a valid filter_name object and cd_values
- int[] flags = {0};
- long[] cd_nelmts = {1};
- int[] cd_values = {0};
- String[] filter_name = {""};
- int[] filter_config = {0};
- int filter_type = -1;
-
- filter_type = H5.H5Pget_filter(dcpl_id, 0, flags, cd_nelmts, cd_values, 120, filter_name,
- filter_config);
- System.out.print("Filter type is: ");
- switch (H5Z_filter.get(filter_type)) {
- case H5Z_FILTER_DEFLATE:
- System.out.println("H5Z_FILTER_DEFLATE");
- break;
- case H5Z_FILTER_SHUFFLE:
- System.out.println("H5Z_FILTER_SHUFFLE");
- break;
- case H5Z_FILTER_FLETCHER32:
- System.out.println("H5Z_FILTER_FLETCHER32");
- break;
- case H5Z_FILTER_SZIP:
- System.out.println("H5Z_FILTER_SZIP");
- break;
- case H5Z_FILTER_NBIT:
- System.out.println("H5Z_FILTER_NBIT");
- break;
- case H5Z_FILTER_SCALEOFFSET:
- System.out.println("H5Z_FILTER_SCALEOFFSET");
- break;
- default:
- System.out.println("H5Z_FILTER_ERROR");
- }
- System.out.println();
- }
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Read the data using the default properties.
- try {
- if (dataset_id >= 0)
- H5.H5Dread(dataset_id, HDF5Constants.H5T_NATIVE_INT, HDF5Constants.H5S_ALL,
- HDF5Constants.H5S_ALL, HDF5Constants.H5P_DEFAULT, dset_data);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Find the maximum value in the dataset, to verify that it was read correctly.
- int max = dset_data[0][0];
- for (int indx = 0; indx < DIM_X; indx++)
- for (int jndx = 0; jndx < DIM_Y; jndx++) {
- if (max < dset_data[indx][jndx])
- max = dset_data[indx][jndx];
- }
-
- // Print the maximum value.
- System.out.println("Maximum value in " + DATASETNAME + " is: " + max);
-
- // End access to the dataset and release resources used by it.
- try {
- if (dcpl_id >= 0)
- H5.H5Pclose(dcpl_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- try {
- if (dataset_id >= 0)
- H5.H5Dclose(dataset_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
-
- // Close the file.
- try {
- if (file_id >= 0)
- H5.H5Fclose(file_id);
- }
- catch (Exception e) {
- e.printStackTrace();
- }
- }
-
- public static void main(String[] args)
- {
-
- // Check if Scale-Offset compression is available and can be used
- // for both compression and decompression. Normally we do not
- // perform error checking in these examples for the sake of
- // clarity, but in this case we will make an exception because this
- // filter is an optional part of the hdf5 library.
- if (H5Ex_D_Soint.checkScaleoffsetFilter()) {
- H5Ex_D_Soint.writeData();
- H5Ex_D_Soint.readData();
- }
- }
-}