HDF5 documents and links Introduction to HDF5 HDF5 User Guide |
And in this document, the
HDF5 Reference Manual
H5 H5A H5D H5E H5F H5G H5I H5P H5R H5S H5T H5Z Tools Datatypes |
These tools enable the user to examine HDF5 file contents, to modify HDF5 file contents in a carefully prescribed manner, and to convert files from HDF4 format to HDF5 format and vice versa.
|
h5dump
[
OPTIONS]
file
h5dump
enables the user to examine
the contents of an HDF5 file and dump those contents, in human
readable form, to an ASCII file.
h5dump
dumps HDF5 file content to standard output.
It can display the contents of the entire HDF5 file or
selected objects, which can be groups, datasets, a subset of a
dataset, links, attributes, or datatypes.
The --header
option displays object header
information only.
Names are the absolute names of the objects. h5dump
displays objects in the order same as the command order. If a
name does not start with a slash, h5dump
begins
searching for the specified object starting at the root group.
If an object is hard linked with multiple names,
h5dump
displays the content of the object in the
first occurrence. Only the link information is displayed in later
occurrences.
h5dump
assigns a name for any unnamed datatype in
the form of
#
oid1:
oid2, where
oid1 and oid2 are the object identifiers
assigned by the library. The unnamed types are displayed within
the root group.
Datatypes are displayed with standard type names. For example,
if a dataset is created with H5T_NATIVE_INT
type
and the standard type name for integer on that machine is
H5T_STD_I32BE
, h5dump
displays
H5T_STD_I32BE
as the type of the dataset.
h5dump
can also dump a subset of a dataset.
This feature operates in much the same way as hyperslabs in HDF5;
the parameters specified on the commnd line are passed to the
function
H5Sselect_hyperslab
and the resulting selection
is displayed.
The h5dump
output is described in detail in the
DDL for HDF5, the
Data Description Language document.
Note: It is not permissable to specify multiple
attributes, datasets, datatypes, groups, or soft links with one
flag. For example, one may not issue the command
WRONG:
h5dump -a /attr1 /attr2 foo.h5
to display both /attr1
and /attr2
.
One must issue the following command:
CORRECT:
h5dump -a /attr1 -a /attr2 foo.h5
It's possible to select the file driver with which to open the HDF5 file by using the --filedriver (-f) command-line option. Acceptable values for the --filedriver option are: "sec2", "family", "split", "multi", and "stream". If the file driver flag isn't specified, then the file will be opened with each driver in turn and in the order specified above until one driver succeeds in opening the file.
--xml
option, h5dump
generates
XML output. This output contains a complete description of the file,
marked up in XML. The XML conforms to the HDF5 Document Type
Definition (DTD) available at
http://hdf.ncsa.uiuc.edu/DTDs/HDF5-File.dtd
.
The XML output is suitable for use with other tools, including the HDF5 Java Tools.
-h
or
--help
-B
or
--bootblock
-H
or
--header
-i
or
--object-ids
-r
or
--string
-V
or
--version
-a P
or
--attribute=P
-d P
or
--dataset=P
-f D
or
--filedriver=D
-g P
or
--group=P
-l P
or
--soft-link=P
-o F
or
--output=F
-t T
or
--datatype=T
-w N
or
--width=N
-x U
or
--xml
-D U
or
--xml-dtd=U
-s L
or
--start=L
-S L
or
--stride=L
-c L
or
--count=L
-k L
or
--block=L
--
--dataset="/foo/mydataset[START;STRIDE;COUNT;BLOCK]"
;
) are required, even when
a parameter value is not specified.
When not specified, default parameter values are used.
/GroupFoo/GroupBar
in the file
quux.h5
:
h5dump -g /GroupFoo/GroupBar quux.h5
Fnord
in the group
/GroupFoo/GroupBar
in the file quux.h5
:
h5dump -d /GroupFoo/GroupBar/Fnord quux.h5
metadata
of the dataset
Fnord
which is in group
/GroupFoo/GroupBar
in the file quux.h5
:
h5dump -a /GroupFoo/GroupBar/Fnord/metadata quux.h5
metadata
which is an
attribute of the root group in the file quux.h5
:
h5dump -a /metadata quux.h5
bobo.h5
:
h5dump --xml bobo.h5 > bobo.h5.xml
/GroupFoo/databar/
in the file quux.h5
h5dump -d /GroupFoo/databar --start="1,1" --stride="2,3"
--count="3,19" --block="1,1" quux.h5
h5dump -d "/GroupFoo/databar[1,1;2,3;3,19;1,1]" quux.h5
h5dump
displays the
following information:
h5ls
[
OPTIONS]
file
[
OBJECTS...]
h5ls
prints selected information about file objects
in the specified format.
-h
or
-?
or
--help
-a
or
--address
-d
or --data
-e
or
--errors
-f
or
--full
-g
or
--group
-l
or
--label
-r
or
--recursive
-s
or
--string
-S
or
--simple
-w
N or
--width=
N
-v
or
--verbose
-V
or
--version
-x
or
--hexdump
%%05d
to open a file family.
printf(3C)
integer format such
as "%05d" to open a file family.
h5repart
[-v]
[-V]
[-[b|m]
N[g|m|k]]
source_file
dest_file
h5repart
splits a single file into a family of
files, joins a family of files into a single file, or copies
one family of files to another while changing the size of the
family members. h5repart
can also be used to
copy a single file to a single file with holes.
Sizes associated with the -b
and -m
options may be suffixed with g
for gigabytes,
m
for megabytes, or k
for kilobytes.
File family names include an integer printf
format such as %d
.
-v
-V
-b
N
-m
N
h5import
infile in_options
[infile in_options ...]
-o outfile
h5import
infile in_options
[infile in_options ...]
-outfile outfile
h5import -h
h5import -help
h5import
converts data
from one or more ASCII or binary files, infile
,
into the same number of HDF5 datasets
in the existing or new HDF5 file, outfile
.
Data conversion is performed in accordance with the
user-specified type and storage properties
specified in in_options
.
The primary objective of h5import
is to
import floating point or integer data.
The utility's design allows for future versions that
accept ASCII text files and store the contents as a
compact array of one-dimensional strings,
but that capability is not implemented in HDF5 Release 1.6.
Input data and options:
Input data can be provided in one of the follwing forms:
infile
,
contains a single n-dimensional
array of values of one of the above types expressed
in the order of fastest-changing dimensions first.
Floating point data in an ASCII input file must be
expressed in the fixed floating form (e.g., 323.56)
h5import
is designed to accept scientific notation
(e.g., 3.23E+02) in an ASCII, but that is not implemented in HDF5 release 1.6.
Each input file can be associated with options specifying the datatype and storage properties. These options can be specified either as command line arguments or in a configuration file. Note that exactly one of these approaches must be used with a single input file.
Command line arguments, best used with simple input files, can be used to specify the class, size, dimensions of the input data and a path identifying the output dataset.
The recommended means of specifying input data options is in a configuration file; this is also the only means of specifying advanced storage features. See further discussion in "The configuration file" below.
The only required option for input data is dimension sizes; defaults are available for all others.
h5import
will accept up to 30 input files in a single call.
Other considerations, such as the maximum length of a command line,
may impose a more stringent limitation.
Output data and options:
The name of the output file is specified following
the -o
or -output
option
in outfile
.
The data from each input file is stored as a separate dataset
in this output file.
outfile
may be an existing file.
If it does not yet exist, h5import
will create it.
Output dataset information and storage properties can be specified only by means of a configuration file.
Dataset path | If the groups in the path leading to the dataset
do not exist, h5import will create them.If no group is specified, the dataset will be created under the root group. If no dataset name is specified, the dataset will be created as dataset1 .h5import does not check for a pre-existing dataset
of the specified or default name; it overwrites any such dataset
without offering an opportunity to preserve it.
| |
Output type | Datatype parameters for output data | |
Output data class | Signed or unsigned integer or floating point | |
Output data size | 8-, 16-, 32-, or 64-bit integer 31- or 64-bit floating point | |
Output architecture | IEEE STD NATIVE (Default)Other architectures are included in the h5import design
but are not implemented in this release.
| |
Output byte order | Little- or big-endian. Relevant only if output architecture is IEEE , UNIX , or STD ;
fixed for other architectures.
| |
Dataset layout and storage properties | Denote how raw data is to be organized on the disk. If none of the following are specified, the default configuration is contiguous layout and with no compression. | |
Layout | Contiguous (Default) Chunked | |
External storage | Allows raw data to be stored in a non-HDF5 file or in an
external HDF5 file. Requires contiguous layout. | |
Compressed | Sets the type of compression and the
level to which the dataset must be compressed. Requires chunked layout. | |
Extendible | Allows the dimensions of the dataset increase over time
and/or to be unlimited. Requires chunked layout. | |
Compressed and extendible | Requires chunked layout. | |
Command-line arguments:
The h5import
syntax for the command-line arguments,
in_options
, is as follows:
h5import infile -d dim_list
[-p pathname]
[-t input_class]
[-s input_size]
[infile ...]
-o outfile or h5import infile -dims dim_list
[-path pathname]
[-type input_class]
[-size input_size]
[infile ...]
-outfile outfile or h5import infile -c config_file
[infile ...]
-outfile outfile
|
-c config_file
option is used with
an input file, no other argument can be used with that input file.
If the -c config_file
option is not used with
an input data file, the -d dim_list
argument
(or -dims dim_list
)
must be used and any combination of the remaining options may be used.
Any arguments used must appear in exactly the order used
in the syntax declarations immediately above.
The configuration file:
A configuration file is specified with the
-c config_file
option:
h5import infile -c config_file
[infile -c config_file2 ...]
-outfile >outfile
|
The configuration file is an ASCII file and must be
organized as "Configuration_Keyword Value" pairs,
with one pair on each line.
For example, the line indicating that
the input data class (configuration keyword INPUT-CLASS
)
is floating point in a text file (value TEXTFP
)
would appear as follows:
INPUT-CLASS TEXTFP
A configuration file may have the following keywords each
followed by one of the following defined values.
One entry for each of the first two keywords,
RANK
and DIMENSION-SIZES
,
is required; all other keywords are optional.
Keyword Value
| Description | ||
---|---|---|---|
RANK
| The number of dimensions in the dataset. (Required) | ||
rank
| An integer specifying the number of dimensions in the dataset. Example: 4 for a 4-dimensional dataset.
| ||
DIMENSION-SIZES
| Sizes of the dataset dimensions. (Required) | ||
dim_sizes
| A string of space-separated integers
specifying the sizes of the dimensions in the dataset.
The number of sizes in this entry must match the value in
the RANK entry.Example: 4 3 4 38 for a 4x3x4x38 dataset.
| ||
PATH
| Path of the output dataset. | ||
path
| The full HDF5 pathname identifying the output dataset
relative to the root group within the output file. I.e., path is a string of optional group names,
each followed by a slash,
and ending with a dataset name.
If the groups in the path do no exist, they will be created.If PATH is not specified, the default
path is /dataset1 .Example: The configuration file entry
dataset1 will be written
in the group grp2/ which is in the group grp1/ ,
a member of the root group in the output file.
| ||
INPUT-CLASS
| A string denoting the type of input data. | ||
TEXTIN
| Input is signed integer data in an ASCII file. | ||
TEXTUIN
| Input is unsigned integer data in an ASCII file. | ||
TEXTFP
| Input is floating point data in fixed notation (e.g., 325.34) in an ASCII file. | ||
TEXTFPE
| Input is floating point data in scientific notation (e.g., 3.2534E+02)
in an ASCII file. (Not implemented in this release.) | ||
IN
| Input is signed integer data in a binary file. | ||
UIN
| Input is unsigned integer data in a binary file. | ||
FP
| Input is floating point data in a binary file. (Default) | ||
STR
| Input is character data in an ASCII file.
With this value, the configuration keywords
RANK , DIMENSION-SIZES ,
OUTPUT-CLASS , OUTPUT-SIZE ,
OUTPUT-ARCHITECTURE , and OUTPUT-BYTE-ORDER
will be ignored.(Not implemented in this release.) | ||
INPUT-SIZE
| An integer denoting the size of the input data, in bits. | ||
8 16 32 64
| For signed and unsigned integer data:
TEXTIN , TEXTUIN ,
IN , or UIN .
(Default: 32 )
| ||
32 64
| For floating point data:
TEXTFP , TEXTFPE ,
or FP .
(Default: 32 )
| ||
OUTPUT-CLASS
| A string denoting the type of output data. | ||
IN
| Output is signed integer data. (Default if INPUT-CLASS is
IN or TEXTIN )
| ||
UIN
| Output is unsigned integer data. (Default if INPUT-CLASS is
UIN or TEXTUIN )
| ||
FP
| Output is floating point data. (Default if INPUT-CLASS is not specified or is
FP , TEXTFP , or TEXTFPE )
| ||
STR
| Output is character data,
to be written as a 1-dimensional array of strings. (Default if INPUT-CLASS is UIN
or TEXTUIN )(Not implemented in this release.) | ||
OUTPUT-SIZE
| An integer denoting the size of the output data, in bits. | ||
8 16 32 64
| For signed and unsigned integer data:
IN or UIN .
(Default: Same as INPUT-SIZE , else 32 )
| ||
32 64
| For floating point data:
FP .
(Default: Same as INPUT-SIZE , else 32 )
| ||
OUTPUT-ARCHITECTURE
| A string denoting the type of output architecture. | ||
STD IEEE INTEL * CRAY * MIPS * ALPHA * NATIVE UNIX *
| See the "Predefined Atomic Types" section
in the "HDF5 Datatypes" chapter
of the HDF5 User's Guide
for a discussion of these architectures. Values marked with an asterisk (*) are not implemented in this release. (Default: NATIVE )
| ||
OUTPUT-BYTE-ORDER
| A string denoting the output byte order. This entry is ignored if the OUTPUT-ARCHITECTURE
is not specified or if it is specified as IEEE ,
UNIX , or STD .
| ||
BE
| Big-endian. (Default) | ||
LE
| Little-endian. | ||
The following options are disabled by default, making the default storage properties no chunking, no compression, no external storage, and no extensible dimensions. | |||
CHUNKED-DIMENSION | Dimension sizes of the chunk for chunked output data. BTW, is this CHUNKED-DIMENSION or CHUNKED-D...-SIZES ?
| ||
chunk_dims
| A string of space-separated integers specifying the
dimension sizes of the chunk for chunked output data.
The number of dimensions must correspond to the value
of RANK .The presence of this field indicates that the output dataset is to be stored in chunked layout; if this configuration field is absent, the dataset will be stored in contiguous layout. | ||
COMPRESSION-TYPE
| Type of compression to be used with chunked storage. Requires that CHUNKED-DIMENSION be specified.
| ||
GZIP
| Gzip compression. Othe compression algorithms are not implemented in this release of h5import .
| ||
COMPRESSION-PARAM
| Compression level. Required if COMPRESSION-TYPE is specified.
Since there is a default, is "required" true?
| ||
1 through 9
| Gzip compression levels:
1 will result in the fastest compression
while 9 will result in the best compression ratio.
(Default: 6)
| ||
EXTERNAL-STORAGE
| Name of an external file in which to create the output dataset. Cannot be used with CHUNKED-DIMENSIONS ,
COMPRESSION-TYPE , OR MAXIMUM-DIMENSIONS .
| ||
external_file
| A string specifying the name of an external file. | ||
MAXIMUM-DIMENSIONS
| Maximum sizes of all dimensions. Requires that CHUNKED-DIMENSION be specified.
| ||
max_dims
| A string of space-separated integers specifying the
maximum size of each dimension of the output dataset.
A value of -1 for any dimension implies
unlimited size for that particular dimension.The number of dimensions must correspond to the value of RANK . | ||
The help
option:
The help option, expressed as one of
h5import -h or h5import -help | |
prints the h5import usage summary | |
h5import -h[elp], OR
| |
then exits. |
infile(s)
in_options
-dims
argument
is required, arguments must used in the order in which they are listed below.
-d dim_list
-dims dim_list
dim_list
is a string of
comma-separated numbers with no spaces
describing the dimensions of the input data.
For example, a 50 x 100 2-dimensional array would be
specified as -dims 50,100
.-p pathname
-pathname pathname
pathname
is a string consisiting of
one or more strings separated by '/' specifying the path
of the dataset in the output file.
If the groups in the path do no exist, they will be created./dataset1
.h5import
does not check for a pre-existing dataset
of the specified or default name; it overwrites any such dataset
without offering an opportunity to preserve it.
-t input_class
-type input_class
input_class
specifies the class of the
input data and determines the class of the output data.FP
.
-s input_size
-size input_size
input_size
specifies the size in bits
of the input data and determines the size of the output data.8
, 16
, 32
, and 64
.32
and 64
.32
.
-c config_file
config_file
specifies a configuration file.infile
and -o outfile
outfile
h5import infile -dims 2,3,4 -type TEXTIN -size 32 -o out1
| |
This command creates a file out1 containing
a single 2x3x4 32-bit integer dataset.
Since no pathname is specified, the dataset is stored
in out1 as /dataset1 .
| |
h5import infile -dims 20,50 -path bin1/dset1 -type FP -size 64 -o out2
| |
This command creates a file out2 containing
a single a 20x50 64-bit floating point dataset.
The dataset is stored in out2 as /bin1/dset1 .
|
outfile
at /work/h5/pkamat/First-set
.PATH work h5 pkamat First-set INPUT-CLASS TEXTFP RANK 3 DIMENSION-SIZES 5 2 4 OUTPUT-CLASS FP OUTPUT-SIZE 64 OUTPUT-ARCHITECTURE IEEE OUTPUT-BYTE-ORDER LE CHUNKED-DIMENSION 2 2 2 MAXIMUM-DIMENSIONS 8 8 -1The next configuration file specifies the following:
NATIVE
format
(as the output architecure is not specified).outfile
at /Second-set
.
PATH Second-set INPUT-CLASS IN RANK 5 DIMENSION-SIZES 6 3 5 2 4 OUTPUT-CLASS IN OUTPUT-SIZE 32 CHUNKED-DIMENSION 2 2 2 2 2 COMPRESSION-TYPE GZIP COMPRESSION-PARAM 7
gif2h5
gif_file h5_file
gif2h5
accepts as input the GIF file gif_file
and produces the HDF5 file h5_file as output.
h52gif
h5_file gif_file
-i
h5_image
[-p
h5_palette]
h52gif
accepts as input the HDF5 file h5_file
and the names of images and associated palettes within that file
as input and produces the GIF file gif_file,
containing those images, as output.
h52gif
expects at least
one h5_image.
You may repeat
-i
h5_image
[-p
h5_palette]
up to 50 times, for a maximum of 50 images.
-i
h5_image
-p
h5_palette
h5toh4 -h
h5toh4
h5file
h4fileh5toh4
h5fileh5toh4 -m
h5file1
h5file2
h5file3 ...
h5toh4
is an HDF5 utility which reads
an HDF5 file, h5file, and converts all
supported objects and pathways to produce an HDF4 file,
h4file. If h4file already exists,
it will be replaced.
If only one file name is given, the name must end in
.h5
and is assumed to represent the
HDF5 input file. h5toh4
replaces the
.h5
suffix with .hdf
to form
the name of the resulting HDF4 file and proceeds as above.
If a file with the name of the intended HDF4 file already
exists, h5toh4
exits with an error without
changing the contents of any file.
The -m
option allows multiple HDF5 file
arguments. Each file name is treated the same as the
single file name case above.
The -h
option causes the following
syntax summary to be displayed:
h5toh4 file.h5 file.hdf h5toh4 file.h5 h5toh4 -m file1.h5 file2.h5 ...
The following HDF5 objects occurring in an HDF5 file are converted to HDF4 objects in the HDF4 file:
Attributes associated with any of the supported HDF5 objects are carried over to the HDF4 objects. Attributes may be of integer, floating point, or fixed length string datatype and they may have up to 32 fixed dimensions.
All datatypes are converted to big-endian. Floating point datatypes are converted to IEEE format.
-h
-m
h4toh5 -h
h4toh5
h4file
h5fileh4toh5
h4fileh4toh5
is a file conversion utility that reads
an HDF4 file, h4file (input.hdf
for example),
and writes an HDF5 file, h5file (output.h5
for example), containing the same data.
If no output file h5file is specified,
h4toh5
uses the input filename to designate
the output file, replacing the extension .hdf
with .h5
.
For example, if the input file scheme3.hdf
is
specified with no output filename, h4toh5
will
name the output file scheme3.h5
.
The -h
option causes a syntax summary
similar to the following to be displayed:
h4toh5 inputfile.hdf outputfile.h5 h5toh4 inputfile.hdf
Each object in the HDF4 file is converted to an equivalent HDF5 object, according to the mapping described in Mapping HDF4 Objects to HDF5 Objects. (If this mapping changes between HDF5 Library releases, a more up-to-date version may be available at Mapping HDF4 Objects to HDF5 Objects on the HDF FTP server.)
In this inital version, h4toh5
converts the following
HDF4 objects:
HDF4 Object | Resulting HDF5 Object |
---|---|
SDS | Dataset |
GR, RI8, and RI24 image | Dataset |
Vdata | Dataset |
Vgroup | Group |
Annotation | Attribute |
Palette | Dataset |
-h
h5cc
[
OPTIONS]
<compile line>
h5cc
can be used in much the same way MPIch is used
to compile an HDF5 program. It takes care of specifying where the
HDF5 header files and libraries are on the commandline.
h5cc
supercedes all other compiler scripts in that
if you've used them to compile the HDF5 library, then
h5cc
also uses those scripts. For example, when
compiling an MPIch program, you use the mpicc
script. If you've built HDF5 using MPIch, then h5cc
uses the MPIch program for compilation.
Some programs use HDF5 in only a few modules. It isn't necessary
to use h5cc
to compile those modules which don't use
HDF5. In fact, since h5cc
is only a convenience
script, you are still able to compile HDF5 modules in the normal
way. In that case, you will have to specify the HDF5 libraries
and include paths yourself.
An example of how to use h5cc
to compile the program
hdf_prog
, which consists of modules
prog1.c
and prog2.c
and uses the HDF5
shared library, would be as follows:
# h5cc -c prog1.c # h5cc -c prog2.c # h5cc -shlib -o hdf_prog prog1.o prog2.o
-help
-echo
-prefix=DIR
DIR
to find the HDF5
lib/
and include/
subdirectories.
-show
-shlib
-noshlib
h5cc
uses the same compiler you used to compile HDF5.
Check your compiler's manual for more information on which
options are needed.
h5cc
.
HDF5_CC
HDF5_CLINKER
HDF5_USE_SHLIB=[yes|no]
h5fc
[
OPTIONS]
<compile line>
h5fc
can be used in much the same way MPIch is used
to compile an HDF5 program. It takes care of specifying where the
HDF5 header files and libraries are on the commandline.
h5fc
supercedes all other compiler scripts in that
if you've used them to compile the HDF5 Fortran library, then
h5fc
also uses those scripts. For example, when
compiling an MPIch program, you use the mpif90
script. If you've built HDF5 using MPIch, then h5fc
uses the MPIch program for compilation.
Some programs use HDF5 in only a few modules. It isn't necessary
to use h5fc
to compile those modules which don't use
HDF5. In fact, since h5fc
is only a convenience
script, you are still able to compile HDF5 Fortran modules in the
normal way. In that case, you will have to specify the HDF5 libraries
and include paths yourself.
An example of how to use h5fc
to compile the program
hdf_prog
, which consists of modules
prog1.f90
and prog2.f90
and uses the HDF5 Fortran library, would be as follows:
# h5fc -c prog1.f90 # h5fc -c prog2.f90 # h5fc -o hdf_prog prog1.o prog2.o
-help
-echo
-prefix=DIR
DIR
to find HDF5
lib/
and include/
subdirectories
-show
h5fc
uses the same compiler you used
to compile HDF5. Check your compiler's manual for
more information on which options are needed.
h5cc
.
HDF5_FC
HDF5_FLINKER
HDF5 documents and links Introduction to HDF5 HDF5 User Guide |
And in this document, the
HDF5 Reference Manual
H5 H5A H5D H5E H5F H5G H5I H5P H5R H5S H5T H5Z Tools Datatypes |