diff options
author | Frank Baker <fbaker@hdfgroup.org> | 2003-06-23 20:32:50 (GMT) |
---|---|---|
committer | Frank Baker <fbaker@hdfgroup.org> | 2003-06-23 20:32:50 (GMT) |
commit | 27b07d75ace268ea680a66ef905c02361ccc18b5 (patch) | |
tree | 6833a8cd1a3360436a13ab7c4a1c59709227b3e3 /doc | |
parent | 38aa393e8acbf0d05d704e48694873d4b351ea3d (diff) | |
download | hdf5-27b07d75ace268ea680a66ef905c02361ccc18b5.zip hdf5-27b07d75ace268ea680a66ef905c02361ccc18b5.tar.gz hdf5-27b07d75ace268ea680a66ef905c02361ccc18b5.tar.bz2 |
[svn-r7085]
Purpose:
Complete "compression" to "filters" revisions.
H5Zregister review feedback.
Description:
Intro -- Finish the "Compression" ==> "Filters" changes.
H5Zregister -- Clarify that the can_apply_func and set_local_func
can be set to null.
-- Define value of filter_id
-- Other smaller revisions.
(all based on review feedback)
Platforms tested:
IE 5, Safari
Diffstat (limited to 'doc')
-rw-r--r-- | doc/html/RM_H5Z.html | 35 |
1 files changed, 20 insertions, 15 deletions
diff --git a/doc/html/RM_H5Z.html b/doc/html/RM_H5Z.html index a6ebe20..f3f3964 100644 --- a/doc/html/RM_H5Z.html +++ b/doc/html/RM_H5Z.html @@ -74,15 +74,15 @@ and customized raw data processing during I/O operations. HDF5 is distributed with a small set of standard filters such as compression (gzip and a shuffling algorithm) and error checking (Fletcher32 checksum). -For further flexibility, the library includes tools enabling a +For further flexibility, the library allows a user application to extend the pipeline through the creation and registration of customized filters. <p> As mentioned above, one set of filters distributed with HDF5 provides built-in methods for raw data compression. The flexibility of the filter pipeline implementation enables the -definition of additional compression methods by a user application. -A compression method<br> +definition of additional filters by a user application. +A filter<br> — is associated with a dataset when the dataset is created,<br> — @@ -91,9 +91,9 @@ A compression method<br> — is applied independently to each chunk of the dataset. <p> -The HDF5 library does not support compression for contiguous datasets +The HDF5 library does not support filters for contiguous datasets because of the difficulty of implementing random access for partial I/O. -Compact dataset compression is not supported because it would not produce +Compact dataset filters are not supported because it would not produce significant results. <p> See <a href="Datasets.html"><cite>The Dataset Interface (H5D)</cite></a> @@ -146,16 +146,16 @@ facilitate moving easily between them.</i> HDF5 library. <p> Making a new filter available to an application is a two-step - process. The first step is to <span class="termEmphasis">define</span> - the three filter callback filter functions described below: - <code>can_applyr_func</code>, <code>set_local_func</code>, and + process. The first step is to write + the three filter callback functions described below: + <code>can_apply_func</code>, <code>set_local_func</code>, and <code>filter_func</code>. - This step can be skipped only when the filter is predefined, as is - the case with the Fletcher32 checksum and shuffle filters that - are distributed with the HDF5 Library. This call to <code>H5Zregister</code>, <span class="termEmphasis">registering</span> the filter with the library, is the second step. + The <code>can_apply_func</code> and <code>set_local_func</code> + fields can be set to <code>NULL</code> + if they are not required for the filter being registered. <p> <code>H5Zregister</code> accepts a single parameter, the <code>filter_class</code> data structure, @@ -172,13 +172,18 @@ facilitate moving easily between them.</i> <p> <code>filter_id</code> is the identifier for the new filter. + This is a user-defined value between + <code>H5Z_FILTER_RESERVED</code> and <code>H5Z_FILTER_MAX</code>, + both of which are defined in the HDF5 source file + <code>H5Zpublic.h</code>. <p> - <code>comment</code> is used for debugging and may be - the null pointer. + <code>comment</code> is used for debugging, + may contain a descriptive name for the filter, + and may be the null pointer. <p> <code>can_apply_func</code>, described in detail below, is a user-defined callback function which determines whether - the combination of the dataset creation property list setting, + the combination of the dataset creation property list values, the datatype, and the dataspace represent a valid combination to apply this filter to. <p> @@ -257,7 +262,7 @@ facilitate moving easily between them.</i> <code>space_id</code>, a dataspace describing the chunk (for chunked dataset storage), which should also not be modified. <p> - The <i>set local</i> callback must set any parameters that are + The <i>set local</i> callback must set any filter parameters that are specific to this dataset, based on the combination of the dataset creation property list values, the datatype, and the dataspace. For example, some filters perform different actions based on |