diff options
author | Elena Pourmal <epourmal@hdfgroup.org> | 2006-06-04 22:28:06 (GMT) |
---|---|---|
committer | Elena Pourmal <epourmal@hdfgroup.org> | 2006-06-04 22:28:06 (GMT) |
commit | f5a16e9bfa5ee701d6ee40ddd98cfcb03aad6e1b (patch) | |
tree | 9b911e438ed68bc4e193b77a28d9f8f9cc7faefe | |
parent | 681d0a964cb00bb6a6d0c6ced68a6ea829689f54 (diff) | |
download | hdf5-f5a16e9bfa5ee701d6ee40ddd98cfcb03aad6e1b.zip hdf5-f5a16e9bfa5ee701d6ee40ddd98cfcb03aad6e1b.tar.gz hdf5-f5a16e9bfa5ee701d6ee40ddd98cfcb03aad6e1b.tar.bz2 |
[svn-r12402] Purpose: New features
Description: I am checking in changes for the h5stat tool:
- updated Makefile.am to use tools library
- regenerated Makefile.in
- added command flags to display selected information
- added support for external files and filters
- small output formatting changes
For more information about current features and output see
updated h5stat RFC http://hdf.ncsa.uiuc.edu/RFC/h5stat/h5stat-spec.pdf
Solution:
Platforms tested: mir, heping, shanti, copper 64-bit, copper 32-bit parallel
using files from tools/testfiles directory and some NASA files
from our Web site.
Misc. update:
-rw-r--r-- | tools/misc/Makefile.am | 4 | ||||
-rw-r--r-- | tools/misc/Makefile.in | 9 | ||||
-rw-r--r-- | tools/misc/h5stat.c | 387 |
3 files changed, 318 insertions, 82 deletions
diff --git a/tools/misc/Makefile.am b/tools/misc/Makefile.am index d178984..025ce14 100644 --- a/tools/misc/Makefile.am +++ b/tools/misc/Makefile.am @@ -20,7 +20,7 @@ include $(top_srcdir)/config/commence.am # Include src directory -INCLUDES+=-I$(top_srcdir)/src +INCLUDES+=-I$(top_srcdir)/src -I$(top_srcdir)/tools/lib #test script and program TEST_PROG=h5repart_gentest @@ -46,7 +46,7 @@ CHECK_CLEANFILES+=*.h5 ../testfiles/fst_family*.h5 ../testfiles/scd_family*.h5 DISTCLEANFILES=h5cc testh5repart.sh # All programs rely on hdf5 library and h5tools library -LDADD=$(LIBHDF5) $(LIBH5TOOLS) +LDADD=$(LIBH5TOOLS) $(LIBHDF5) # h5cc needs custom install and uninstall rules, since it may be # named h5pcc if hdf5 is being built in parallel mode. diff --git a/tools/misc/Makefile.in b/tools/misc/Makefile.in index 9f8d801..62a6c2a 100644 --- a/tools/misc/Makefile.in +++ b/tools/misc/Makefile.in @@ -76,8 +76,8 @@ PROGRAMS = $(bin_PROGRAMS) h5debug_SOURCES = h5debug.c h5debug_OBJECTS = h5debug.$(OBJEXT) h5debug_LDADD = $(LDADD) -am__DEPENDENCIES_1 = $(top_builddir)/src/libhdf5.la -am__DEPENDENCIES_2 = $(top_builddir)/tools/lib/libh5tools.la +am__DEPENDENCIES_1 = $(top_builddir)/tools/lib/libh5tools.la +am__DEPENDENCIES_2 = $(top_builddir)/src/libhdf5.la h5debug_DEPENDENCIES = $(am__DEPENDENCIES_1) $(am__DEPENDENCIES_2) h5repart_SOURCES = h5repart.c h5repart_OBJECTS = h5repart.$(OBJEXT) @@ -312,7 +312,8 @@ TIME = time # Path for hl needed in hdf5.h # Include src directory -INCLUDES = -I$(top_srcdir)/hl/src -I$(top_srcdir)/src +INCLUDES = -I$(top_srcdir)/hl/src -I$(top_srcdir)/src \ + -I$(top_srcdir)/tools/lib # Some machines need a command to run executables; this is that command # so that our tests will run. @@ -377,7 +378,7 @@ CLEANFILES = h5redeploy DISTCLEANFILES = h5cc testh5repart.sh # All programs rely on hdf5 library and h5tools library -LDADD = $(LIBHDF5) $(LIBH5TOOLS) +LDADD = $(LIBH5TOOLS) $(LIBHDF5) @BUILD_PARALLEL_CONDITIONAL_FALSE@H5CC_NAME = h5cc # h5cc needs custom install and uninstall rules, since it may be diff --git a/tools/misc/h5stat.c b/tools/misc/h5stat.c index e72f609..db93471 100644 --- a/tools/misc/h5stat.c +++ b/tools/misc/h5stat.c @@ -16,12 +16,22 @@ #include <stdlib.h> #include <string.h> #include "H5private.h" /* Generic Functions */ +#include "h5tools.h" +#include "h5tools_utils.h" +#include "h5tools_ref.h" +#include "h5trav.h" #include "hdf5.h" /* Parameters to control statistics gathered */ #define SIZE_SMALL_GROUPS 10 #define SIZE_SMALL_DSETS 10 +#define H5_NFILTERS_IMPL 8 /* Number of currently implemented filters + one to + accommodate for user-define filters + one + to accomodate datasets whithout any filters */ + + + /* Datatype statistics for datasets */ typedef struct dtype_info_t { hid_t tid; /* ID of datatype */ @@ -58,14 +68,17 @@ typedef struct iter_t { hsize_t max_dset_dims; /* Maximum dimension size of dataset */ unsigned long small_dset_dims[SIZE_SMALL_DSETS]; /* Size of dimensions of small datasets tracked */ unsigned long dset_layouts[H5D_NLAYOUTS]; /* Type of storage for each dataset */ + unsigned long dset_comptype[H5_NFILTERS_IMPL]; /* Number of currently implemented filters */ unsigned long dset_ntypes; /* Number of diff. dataset datatypes found */ dtype_info_t *dset_type_info; /* Pointer to dataset datatype information found */ unsigned dset_dim_nbins; /* Number of bins for dataset dimensions */ unsigned long *dset_dim_bins; /* Pointer to array of bins for dataset dimensions */ ohdr_info_t dset_ohdr_info; /* Object header information for datasets */ hsize_t dset_storage_size; /* Size of raw data for datasets */ + unsigned long nexternal; /* Number of external files for a dataset */ } iter_t; + /* Table containing object id and object name */ static struct { int nalloc; /* number of slots allocated */ @@ -76,6 +89,101 @@ static struct { } *obj; } idtab_g; +const char *progname = "h5stat"; +int d_status = EXIT_SUCCESS; +static int display_all = TRUE; +static int display_file_metadata = FALSE; +static int display_file = FALSE; +static int display_group_metadata = FALSE; +static int display_group = FALSE; +static int display_dset_metadata = FALSE; +static int display_dset = FALSE; +static int display_dtype_metadata = FALSE; +static int display_dtype = FALSE; + +static const char *s_opts ="FfhGgDdT"; +static struct long_options l_opts[] = { + {"help", no_arg, 'h'}, + {"hel", no_arg, 'h'}, + {"file", no_arg, 'f'}, + {"fil", no_arg, 'f'}, + {"fi", no_arg, 'f'}, + {"filemetadata", no_arg, 'F'}, + {"filemetadat", no_arg, 'F'}, + {"filemetada", no_arg, 'F'}, + {"filemetad", no_arg, 'F'}, + {"filemeta", no_arg, 'F'}, + {"filemet", no_arg, 'F'}, + {"fileme", no_arg, 'F'}, + {"filem", no_arg, 'F'}, + {"group", no_arg, 'g'}, + {"grou", no_arg, 'g'}, + {"gro", no_arg, 'g'}, + {"gr", no_arg, 'g'}, + {"groupmetadata", no_arg, 'G'}, + {"groupmetadat", no_arg, 'G'}, + {"groupmetada", no_arg, 'G'}, + {"groupmetad", no_arg, 'G'}, + {"groupmeta", no_arg, 'G'}, + {"groupmet", no_arg, 'G'}, + {"groupme", no_arg, 'G'}, + {"groupm", no_arg, 'G'}, + {"dset", no_arg, 'd'}, + {"dse", no_arg, 'd'}, + {"ds", no_arg, 'd'}, + {"d", no_arg, 'd'}, + {"dsetmetadata", no_arg, 'D'}, + {"dsetmetadat", no_arg, 'D'}, + {"dsetmetada", no_arg, 'D'}, + {"dsetmetad", no_arg, 'D'}, + {"dsetmeta", no_arg, 'D'}, + {"dsetmet", no_arg, 'D'}, + {"dsetme", no_arg, 'D'}, + {"dsetm", no_arg, 'D'}, + {"dtypemetadata", no_arg, 'T'}, + {"dtypemetadat", no_arg, 'T'}, + {"dtypemetada", no_arg, 'T'}, + {"dtypemetad", no_arg, 'T'}, + {"dtypemeta", no_arg, 'T'}, + {"dtypemet", no_arg, 'T'}, + {"dtypeme", no_arg, 'T'}, + {"dtypem", no_arg, 'T'}, + {"dtype", no_arg, 'T'}, + { NULL, 0, '\0' } +}; + +static void +leave(int ret) +{ + h5tools_close(); + exit(ret); +} + + +static void usage(const char *prog) +{ + fflush(stdout); + fprintf(stdout, "\n"); + fprintf(stdout, "This tool is under development. For detailed information\n"); + fprintf(stdout, "please see the specification document at\n"); + fprintf(stdout, "http://hdf.ncsa.uiuc.edu/RFC/h5stat/h5stat-spec.pdf\n"); + fprintf(stdout, "\n"); + fprintf(stdout, "Please send your comments and questions to help@hdfgroup.org\n"); + fprintf(stdout, "\n"); + fprintf(stdout, "Usage: %s [OPTIONS] file\n", prog); + fprintf(stdout, "\n"); + fprintf(stdout, " OPTIONS\n"); + fprintf(stdout, " -h, --help Print a usage message and exit\n"); + fprintf(stdout, " -f, --file Print file information\n"); + fprintf(stdout, " -F, --filemetadata Print file metadata\n"); + fprintf(stdout, " -g, --group Print group information\n"); + fprintf(stdout, " -G, --groupmetadata Print group metadata\n"); + fprintf(stdout, " -d, --dset Print dataset information\n"); + fprintf(stdout, " -D, --dsetmetadata Print dataset metadata\n"); + fprintf(stdout, " -T, --dtypemetadata Print datatype metadata\n"); + fprintf(stdout, "\n"); +} + /*------------------------------------------------------------------------- * Function: ceil_log10 @@ -171,7 +279,7 @@ sym_lookup(H5G_stat_t *sb) return NULL; /*only one name possible*/ objno = (haddr_t)sb->objno[0] | ((haddr_t)sb->objno[1] << (8 * sizeof(long))); for (n=0; n<idtab_g.nobjs; n++) { - if(idtab_g.obj[n].id == objno) + if (idtab_g.obj[n].id==objno) return idtab_g.obj[n].name; } return NULL; @@ -345,6 +453,9 @@ printf("walk: fullname = %s\n", fullname); int ndims; /* Number of dimensions of dataset */ hsize_t storage; /* Size of dataset storage */ unsigned u; /* Local index variable */ + int num_ext; /* Number of external files for a dataset */ + int nfltr; /* Number of filters for a dataset */ + H5Z_filter_t fltr; /* Filter identifier */ /* Gather statistics about this type of object */ iter->uniq_dsets++; @@ -455,7 +566,27 @@ printf("walk: fullname = %s\n", fullname); /* Track the layout type for dataset */ (iter->dset_layouts[lout])++; + + num_ext = H5Pget_external_count(dcpl); + assert (num_ext >= 0); + + if(num_ext) iter->nexternal = iter->nexternal + num_ext; + /* Track different filters */ + + if ((nfltr=H5Pget_nfilters(dcpl)) >= 0) { + + if (nfltr == 0) iter->dset_comptype[0]++; + for (u=0; u < nfltr; u++) { + fltr = H5Pget_filter(dcpl, u, 0, 0, 0, 0, 0, NULL); + if (fltr < (H5_NFILTERS_IMPL-1)) + iter->dset_comptype[fltr]++; + else + iter->dset_comptype[H5_NFILTERS_IMPL-1]++; /*other filters*/ + } + + } /*endif nfltr */ + ret = H5Pclose(dcpl); assert(ret >= 0); @@ -487,29 +618,94 @@ printf("walk: fullname = %s\n", fullname); return 0; } -static void -usage(void) +static int +parse_command_line(int argc, const char *argv[]) { - printf("usage: h5stat <filename>\n"); - exit(1); + int opt; + + + /* parse command line options */ + while ((opt = get_option(argc, argv, s_opts, l_opts)) != EOF) { +parse_start: + switch ((char)opt) { + case 'F': + display_all = FALSE; + display_file_metadata = TRUE; + break; + case 'f': + display_all = FALSE; + display_file = TRUE; + break; + case 'G': + display_all = FALSE; + display_group_metadata = TRUE; + break; + case 'g': + display_all = FALSE; + display_group = TRUE; + break; + case 'T': + display_all = FALSE; + display_dtype_metadata = TRUE; + break; + case 'D': + display_all = FALSE; + display_dset_metadata = TRUE; + break; + case 'd': + display_all = FALSE; + display_dset = TRUE; + break; + case 'h': + usage(progname); + leave(EXIT_SUCCESS); + default: + usage(progname); + leave(EXIT_FAILURE); + } + } + +parse_end: + /* check for file name to be processed */ + if (argc <= opt_ind) { + error_msg(progname, "missing file name\n"); + usage(progname); + leave(EXIT_FAILURE); + } } int -main(int argc, char *argv[]) +main(int argc, const char *argv[]) { - iter_t iter; - hid_t fid; - unsigned long power; /* Temporary "power" for bins */ - size_t dtype_size; /* Size of encoded datatype */ - unsigned long total; /* Total count for various statistics */ - unsigned u; /* Local index variable */ - - if(argc != 2) - usage(); - - printf("Filename: %s\n", argv[1]); + iter_t iter; + const char *fname = NULL; + hid_t fid; + unsigned long power; /* Temporary "power" for bins */ + size_t dtype_size; /* Size of encoded datatype */ + unsigned long total; /* Total count for various statistics */ + unsigned u; /* Local index variable */ + int status; /* Return value for parse_command_line function */ + void *edata; + H5E_auto_stack_t func; + + /* Disable error reporting */ + H5Eget_auto_stack(H5E_DEFAULT, &func, &edata); + H5Eset_auto_stack(H5E_DEFAULT, NULL, NULL); + + /* Initialize h5tools lib */ + h5tools_init(); + status = parse_command_line (argc, argv); + fname = argv[opt_ind]; + + printf("Filename: %s\n", fname); + + fid = H5Fopen(fname, H5F_ACC_RDONLY, H5P_DEFAULT); + if (fid < 0) { + error_msg(progname, "unable to open file \"%s\"\n", fname); + leave(EXIT_FAILURE); + } - if((fid = H5Fopen(argv[1], H5F_ACC_RDONLY, H5P_DEFAULT)) > 0) { + if(fid > 0) { iter.container = "/"; iter.uniq_groups = 0; iter.uniq_dsets = 0; @@ -534,6 +730,8 @@ main(int argc, char *argv[]) iter.small_dset_dims[u] = 0; for(u = 0; u < H5D_NLAYOUTS; u++) iter.dset_layouts[u] = 0; + for(u = 0; u < H5_NFILTERS_IMPL; u++) + iter.dset_comptype[u] = 0; iter.dset_ntypes = 0; iter.dset_type_info = NULL; iter.dset_dim_nbins = 0; @@ -541,61 +739,83 @@ main(int argc, char *argv[]) iter.dset_ohdr_info.total_size = 0; iter.dset_ohdr_info.free_size = 0; iter.dset_storage_size = 0; + iter.nexternal = 0; walk(fid, "/", &iter); H5Fclose(fid); } /* end if */ /* Print information about the file */ - printf("File Hierarchy:\n"); - printf("\t# of unique groups: %lu\n", iter.uniq_groups); - printf("\t# of unique datasets: %lu\n", iter.uniq_dsets); - printf("\t# of unique named dataypes: %lu\n", iter.uniq_types); - printf("\t# of unique links: %lu\n", iter.uniq_links); - printf("\t# of unique other: %lu\n", iter.uniq_others); - printf("\tMax. # of links to object: %lu\n", iter.max_links); - printf("\tMax. depth of hierarchy: %lu\n", iter.max_depth); - HDfprintf(stdout, "\tMax. # of objects in group: %Hu\n", iter.max_fanout); - - printf("Object header size: (total/unused)\n"); - HDfprintf(stdout, "\tGroups: %Hu/%Hu\n", iter.group_ohdr_info.total_size, iter.group_ohdr_info.free_size); - HDfprintf(stdout, "\tDatasets: %Hu/%Hu\n", iter.dset_ohdr_info.total_size, iter.dset_ohdr_info.free_size); - - printf("Small groups:\n"); - total = 0; - for(u = 0; u < SIZE_SMALL_GROUPS; u++) { - if(iter.num_small_groups[u] > 0) { - printf("\t# of groups of size %u: %lu\n", u, iter.num_small_groups[u]); - total += iter.num_small_groups[u]; - } /* end if */ - } /* end for */ - printf("\tTotal # of small groups: %lu\n", total); - - printf("Group bins:\n"); - total = 0; - if(iter.group_bins[0] > 0) { - printf("\t# of groups of size 0: %lu\n", iter.group_bins[0]); - total = iter.group_bins[0]; - } /* end if */ - power = 1; - for(u = 1; u < iter.group_nbins; u++) { - if(iter.group_bins[u] > 0) { - printf("\t# of groups of size %lu - %lu: %lu\n", power, (power * 10) - 1, iter.group_bins[u]); - total += iter.group_bins[u]; - } /* end if */ - power *= 10; - } /* end for */ - printf("\tTotal # of groups: %lu\n", total); + if(display_all) { + display_file = TRUE; + display_file_metadata = TRUE; + display_group = TRUE; + display_group_metadata = TRUE; + display_dset = TRUE; + display_dtype_metadata = TRUE; + } + + if(display_file) { + + printf("File information\n"); + printf("\t# of unique groups: %lu\n", iter.uniq_groups); + printf("\t# of unique datasets: %lu\n", iter.uniq_dsets); + printf("\t# of unique named dataypes: %lu\n", iter.uniq_types); + printf("\t# of unique links: %lu\n", iter.uniq_links); + printf("\t# of unique other: %lu\n", iter.uniq_others); + printf("\tMax. # of links to object: %lu\n", iter.max_links); + printf("\tMax. depth of hierarchy: %lu\n", iter.max_depth); + HDfprintf(stdout, "\tMax. # of objects in group: %Hu\n", iter.max_fanout); + + } + + + if(display_file_metadata) { + printf("Object header size: (total/unused)\n"); + HDfprintf(stdout, "\tGroups: %Hu/%Hu\n", iter.group_ohdr_info.total_size, iter.group_ohdr_info.free_size); + HDfprintf(stdout, "\tDatasets: %Hu/%Hu\n", iter.dset_ohdr_info.total_size, iter.dset_ohdr_info.free_size); + } + + if(display_group) { + + printf("Small groups:\n"); + total = 0; + for(u = 0; u < SIZE_SMALL_GROUPS; u++) { + if(iter.num_small_groups[u] > 0) { + printf("\t# of groups of size %u: %lu\n", u, iter.num_small_groups[u]); + total += iter.num_small_groups[u]; + } /* end if */ + } /* end for */ + printf("\tTotal # of small groups: %lu\n", total); + + printf("Group bins:\n"); + total = 0; + if(iter.group_bins[0] > 0) { + printf("\t# of groups of size 0: %lu\n", iter.group_bins[0]); + total = iter.group_bins[0]; + } /* end if */ + power = 1; + for(u = 1; u < iter.group_nbins; u++) { + if(iter.group_bins[u] > 0) { + printf("\t# of groups of size %lu - %lu: %lu\n", power, (power * 10) - 1, iter.group_bins[u]); + total += iter.group_bins[u]; + } /* end if */ + power *= 10; + } /* end for */ + printf("\tTotal # of groups: %lu\n", total); + + } + if(display_dset) { if(iter.uniq_dsets > 0) { - printf("Dataset dimension info:\n"); + printf("Dataset dimension information:\n"); printf("\tMax. rank of datasets: %lu\n", iter.max_dset_rank); printf("\tDataset ranks:\n"); for(u = 0; u < H5S_MAX_RANK; u++) if(iter.dset_rank_count[u] > 0) printf("\t\t# of dataset with rank %u: %lu\n", u, iter.dset_rank_count[u]); - printf("1-D Dataset info:\n"); + printf("1-D Dataset information:\n"); HDfprintf(stdout, "\tMax. dimension size of 1-D datasets: %Hu\n", iter.max_dset_dims); printf("\tSmall 1-D datasets:\n"); total = 0; @@ -626,28 +846,43 @@ main(int argc, char *argv[]) printf("\t\tTotal # of datasets: %lu\n", total); } /* end if */ - printf("Dataset storage info:\n"); + printf("Dataset storage information:\n"); HDfprintf(stdout, "\tTotal raw data size: %Hu\n", iter.dset_storage_size); - printf("Dataset layout info:\n"); + printf("Dataset layout information:\n"); for(u = 0; u < H5D_NLAYOUTS; u++) printf("\tDataset layout counts[%s]: %lu\n", (u == 0 ? "COMPACT" : (u == 1 ? "CONTIG" : "CHUNKED")), iter.dset_layouts[u]); + printf("\tNumber of external files : %lu\n", iter.nexternal); + printf("Dataset filters information:\n"); + printf("\tNumber of datasets with \n"); + printf("\t NO filter: %lu\n", iter.dset_comptype[H5Z_FILTER_ERROR+1]); + printf("\t GZIP filter: %lu\n", iter.dset_comptype[H5Z_FILTER_DEFLATE]); + printf("\t SHUFFLE filter: %lu\n", iter.dset_comptype[H5Z_FILTER_SHUFFLE]); + printf("\t FLETCHER32 filter: %lu\n", iter.dset_comptype[H5Z_FILTER_FLETCHER32]); + printf("\t SZIP filter: %lu\n", iter.dset_comptype[H5Z_FILTER_SZIP]); + printf("\t NBIT filter: %lu\n", iter.dset_comptype[H5Z_FILTER_NBIT]); + printf("\t SCALEOFFSET filter: %lu\n", iter.dset_comptype[H5Z_FILTER_SCALEOFFSET]); + printf("\t USER-DEFINED filter: %lu\n", iter.dset_comptype[H5_NFILTERS_IMPL-1]); + if(display_dtype_metadata) { + printf("Dataset datatype information:\n"); + printf("\t# of unique datatypes used by datasets: %lu\n", iter.dset_ntypes); + total = 0; + for(u = 0; u < iter.dset_ntypes; u++) { + H5Tencode(iter.dset_type_info[u].tid, NULL, &dtype_size); + printf("\tDataset datatype #%u:\n", u); + printf("\t\tCount (total/named) = (%lu/%lu)\n", iter.dset_type_info[u].count, iter.dset_type_info[u].named); + printf("\t\tSize (desc./elmt) = (%lu/%lu)\n", (unsigned long)dtype_size, (unsigned long)H5Tget_size(iter.dset_type_info[u].tid)); + H5Tclose(iter.dset_type_info[u].tid); + total += iter.dset_type_info[u].count; + } /* end for */ + printf("\tTotal dataset datatype count: %lu\n", total); + } + } /* end if */ + } /* display dset */ + + H5Eset_auto_stack(H5E_DEFAULT, func, edata); - printf("Dataset datatype info:\n"); - printf("\t# of unique datatypes used by datasets: %lu\n", iter.dset_ntypes); - total = 0; - for(u = 0; u < iter.dset_ntypes; u++) { - H5Tencode(iter.dset_type_info[u].tid, NULL, &dtype_size); - printf("\tDataset datatype #%u:\n", u); - printf("\t\tCount (total/named) = (%lu/%lu)\n", iter.dset_type_info[u].count, iter.dset_type_info[u].named); - printf("\t\tSize (desc./elmt) = (%lu/%lu)\n", (unsigned long)dtype_size, (unsigned long)H5Tget_size(iter.dset_type_info[u].tid)); - H5Tclose(iter.dset_type_info[u].tid); - total += iter.dset_type_info[u].count; - } /* end for */ - printf("\tTotal dataset datatype count: %lu\n", total); - } /* end if */ - - return (0); + leave(d_status); } |