summaryrefslogtreecommitdiffstats
path: root/test
diff options
context:
space:
mode:
authorLarry Knox <lrknox@hdfgroup.org>2022-04-07 04:49:00 (GMT)
committerLarry Knox <lrknox@hdfgroup.org>2022-04-07 04:49:00 (GMT)
commit91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f (patch)
tree1ca4bfd929700b0743a3ecc653262e836e6436f2 /test
parent9f2b38bb61df6136d214cb7ab248f73d1ff9458c (diff)
downloadhdf5-91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f.zip
hdf5-91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f.tar.gz
hdf5-91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f.tar.bz2
Commit codespell spelling corrections.
Diffstat (limited to 'test')
-rw-r--r--test/SWMR_POSIX_Order_UG.txt4
-rw-r--r--test/SWMR_UseCase_UG.txt6
-rw-r--r--test/accum.c4
-rw-r--r--test/atomic_writer.c2
-rw-r--r--test/big.c6
-rw-r--r--test/btree2.c130
-rw-r--r--test/cache.c66
-rw-r--r--test/cache_api.c8
-rw-r--r--test/cache_common.c10
-rw-r--r--test/cache_image.c12
-rw-r--r--test/cache_tagging.c46
-rw-r--r--test/chunk_info.c4
-rw-r--r--test/cmpd_dset.c2
-rw-r--r--test/cork.c6
-rw-r--r--test/direct_chunk.c14
-rw-r--r--test/dsets.c40
-rw-r--r--test/dt_arith.c22
-rw-r--r--test/dtypes.c12
-rw-r--r--test/earray.c18
-rw-r--r--test/enc_dec_plist_cross_platform.c2
-rw-r--r--test/enum.c2
-rw-r--r--test/external.c6
-rw-r--r--test/farray.c12
-rw-r--r--test/fheap.c8
-rw-r--r--test/file_image.c8
-rw-r--r--test/fillval.c4
-rw-r--r--test/filter_fail.c4
-rw-r--r--test/filter_plugin1_dsets.c2
-rw-r--r--test/filter_plugin3_dsets.c2
-rw-r--r--test/flushrefresh.c12
-rw-r--r--test/gen_bounds.c12
-rw-r--r--test/gen_old_group.c2
-rw-r--r--test/gen_plist.c2
-rw-r--r--test/genall5.c2
-rw-r--r--test/getname.c2
-rw-r--r--test/gheap.c2
-rw-r--r--test/h5test.c2
-rw-r--r--test/hdfs.c2
-rw-r--r--test/links.c8
-rw-r--r--test/mf.c44
-rw-r--r--test/mirror_vfd.c2
-rw-r--r--test/mount.c4
-rw-r--r--test/mtime.c2
-rw-r--r--test/objcopy.c12
-rw-r--r--test/ohdr.c6
-rw-r--r--test/page_buffer.c24
-rw-r--r--test/ros3.c4
-rw-r--r--test/s3comms.c22
-rw-r--r--test/set_extent.c14
-rw-r--r--test/stab.c6
-rw-r--r--test/swmr.c10
-rw-r--r--test/swmr_generator.c2
-rw-r--r--test/swmr_start_write.c2
-rw-r--r--test/tattr.c12
-rw-r--r--test/tcoords.c2
-rw-r--r--test/testmeta.c2
-rw-r--r--test/tfile.c40
-rw-r--r--test/tgenprop.c18
-rw-r--r--test/th5o.c2
-rw-r--r--test/th5s.c2
-rw-r--r--test/tid.c4
-rw-r--r--test/tmisc.c4
-rw-r--r--test/trefer.c2
-rw-r--r--test/tselect.c40
-rw-r--r--test/tsohm.c10
-rw-r--r--test/ttsafe_attr_vlen.c2
-rw-r--r--test/ttsafe_rec_rw_lock.c14
-rw-r--r--test/tvltypes.c2
-rw-r--r--test/twriteorder.c2
-rw-r--r--test/unlink.c6
-rw-r--r--test/use_common.c4
-rw-r--r--test/use_disable_mdc_flushes.c2
-rw-r--r--test/vds.c16
-rw-r--r--test/vds_env.c2
-rw-r--r--test/vfd.c20
75 files changed, 435 insertions, 435 deletions
diff --git a/test/SWMR_POSIX_Order_UG.txt b/test/SWMR_POSIX_Order_UG.txt
index 2771af1..a4190bc 100644
--- a/test/SWMR_POSIX_Order_UG.txt
+++ b/test/SWMR_POSIX_Order_UG.txt
@@ -3,7 +3,7 @@ POSIX Write Order Test Instructions
Purpose
-------
-This documents shows the requirments, implementaion design and instructions
+This documents shows the requirements, implementation design and instructions
of building and running the POSIX Write Order test. The name of the
test is twriteorder and it resides in the test/ directory.
@@ -44,7 +44,7 @@ and they are ready for the reader processes to access.
If the system, in which the writer and reader processes run, the readers
will always get all chain-linked blocks correctly. If the order of write
-is not maintained, some reader processes may found unexpect block data.
+is not maintained, some reader processes may found unexpected block data.
Building the Tests
------------------
diff --git a/test/SWMR_UseCase_UG.txt b/test/SWMR_UseCase_UG.txt
index 1e3d1e6..677fdc4 100644
--- a/test/SWMR_UseCase_UG.txt
+++ b/test/SWMR_UseCase_UG.txt
@@ -2,7 +2,7 @@
User Guide for SWMR Use Case Programs
2. Purpose:
- This is a User Guide of the SWMR Use Case programs. It descibes the use
+ This is a User Guide of the SWMR Use Case programs. It describes the use
case program and explain how to run them.
2.1. Author and Dates:
@@ -87,7 +87,7 @@
It opens the HDF5 data file without the SWMR access mode (0 means
off). This likely will result in error. This option is provided for
- users to see the effect of the neede SWMR access mode for concurrent
+ users to see the effect of the needed SWMR access mode for concurrent
access.
3.4. Test Shell Script:
@@ -176,7 +176,7 @@
It opens the HDF5 data file without the SWMR access mode (0 means
off). This likely will result in error. This option is provided for
- users to see the effect of the neede SWMR access mode for concurrent
+ users to see the effect of the needed SWMR access mode for concurrent
access.
4.4. Test Shell Script:
diff --git a/test/accum.c b/test/accum.c
index 47aae39..73652a3 100644
--- a/test/accum.c
+++ b/test/accum.c
@@ -1348,7 +1348,7 @@ test_accum_adjust(H5F_t *f)
/* ==> Accumulator will try to resize, but see that it's getting too big */
/* ==> Size of new block is less than than half maximum size of accumulator */
/* ==> New block being appended to accumulator */
- /* ==> We can slide the dirty region down, to accomodate the request */
+ /* ==> We can slide the dirty region down, to accommodate the request */
/* ==> Max Buffer Size - (dirty offset + adjust size) >= 2 * size) */
/* ==> Need to adjust location of accumulator while appending */
/* ==> Accumulator will need to be reallocated */
@@ -1394,7 +1394,7 @@ test_accum_adjust(H5F_t *f)
/* ==> Accumulator will try to resize, but see that it's getting too big */
/* ==> Size of new block is less than than half maximum size of accumulator */
/* ==> New block being appended to accumulator */
- /* ==> We can slide the dirty region down, to accomodate the request */
+ /* ==> We can slide the dirty region down, to accommodate the request */
/* ==> Max Buffer Size - (dirty offset + adjust size) < 2 * size) */
/* ==> Need to adjust location of accumulator while appending */
if (accum_write((1024 * 1024) - 5, 10, wbuf) < 0)
diff --git a/test/atomic_writer.c b/test/atomic_writer.c
index 5a6310a..33252f1 100644
--- a/test/atomic_writer.c
+++ b/test/atomic_writer.c
@@ -112,7 +112,7 @@ int
main(int argc, char *argv[])
{
int fd = -1; /* file descriptor */
- ssize_t bytes_wrote; /* the nubmer of bytes written */
+ ssize_t bytes_wrote; /* the number of bytes written */
unsigned int *buf = NULL; /* buffer to hold written data */
unsigned int n, u, i; /* local index variable */
int temp; /* temporary variable */
diff --git a/test/big.c b/test/big.c
index 2bbae67..9f62b3a 100644
--- a/test/big.c
+++ b/test/big.c
@@ -43,7 +43,7 @@
* that disk space is allocated only when the contents are actually written.
* E.g., If one creates a new file, seeks forward 10 million bytes, writes
* 1 bytes and closes the file, then a sparse file, will show file size of
- * 10 million bytes but actaully uses only couple disk blocks, much smaller
+ * 10 million bytes but actually uses only couple disk blocks, much smaller
* than the formal file size.)
*
* One more consideration is that we want to distinguish an HDF5 library
@@ -70,7 +70,7 @@
#endif
/* Define Small, Large, Extra Large, Huge File which
- * corrspond to less than 2GB, 2GB, 4GB, and tens of GB file size.
+ * correspond to less than 2GB, 2GB, 4GB, and tens of GB file size.
* NO_FILE stands for "no file" to be tested.
*/
typedef enum fsizes_t { SFILE, LFILE, XLFILE, HUGEFILE, NO_FILE } fsizes_t;
@@ -609,7 +609,7 @@ usage(void)
"Examples:\n"
"\tbig -fsize 2.1e9 \t# test with file size just under 2GB\n"
"\tbig -fsize 2.2e9 \t# test with file size just above 2GB\n"
- "\tBe sure the file system can support the file size requested\n",
+ "\the sure the file system can support the file size requested\n",
(hsize_t)FAMILY_SIZE);
}
diff --git a/test/btree2.c b/test/btree2.c
index 3e03366..ec99229 100644
--- a/test/btree2.c
+++ b/test/btree2.c
@@ -696,7 +696,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_insert(bt2, &record) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
idx = 41;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
@@ -707,7 +707,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_find(bt2, &idx, NULL, NULL) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in B-tree with 1 record */
+ /* Attempt to find existent record in B-tree with 1 record */
idx = 42;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
TEST_ERROR
@@ -716,7 +716,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_find(bt2, &idx, NULL, NULL) != TRUE)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
idx = 0;
H5E_BEGIN_TRY
{
@@ -764,18 +764,18 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_insert(bt2, &record) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
idx = 41;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several record */
+ /* Attempt to find existent record in level-0 B-tree with several record */
idx = 56;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
idx = 0;
H5E_BEGIN_TRY
{
@@ -935,23 +935,23 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (idx != (INSERT_SPLIT_ROOT_NREC + 2))
TEST_ERROR
- /* Attempt to find non-existant record in level-1 B-tree */
+ /* Attempt to find non-existent record in level-1 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_SPLIT_ROOT_NREC + 10;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in root of level-1 B-tree */
+ /* Attempt to find existent record in root of level-1 B-tree */
idx = 33;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
- /* Attempt to find existant record in leaf of level-1 B-tree */
+ /* Attempt to find existent record in leaf of level-1 B-tree */
idx = 56;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
- /* Attempt to index non-existant record in level-1 B-tree */
+ /* Attempt to index non-existent record in level-1 B-tree */
idx = 0;
H5E_BEGIN_TRY
{
@@ -1671,13 +1671,13 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx != ((INSERT_SPLIT_ROOT_NREC * 29) + 5))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_SPLIT_ROOT_NREC * 30;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
idx = 948;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
@@ -1687,7 +1687,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
idx = 505;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
@@ -1697,7 +1697,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
idx = 555;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
@@ -1707,7 +1707,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
idx = 0;
H5E_BEGIN_TRY
{
@@ -3060,7 +3060,7 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
if (idx != INSERT_MANY)
TEST_ERROR
- /* Attempt to find non-existant record in level-4 B-tree */
+ /* Attempt to find non-existent record in level-4 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_MANY * 2;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
@@ -3071,12 +3071,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
/* Pick random record */
idx = (hsize_t)(HDrandom() % INSERT_MANY);
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
} /* end for */
- /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */
+ /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY * 3), find_cb, NULL);
@@ -3099,12 +3099,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
/* Pick random record */
idx = (hsize_t)(HDrandom() % INSERT_MANY);
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
/* (in increasing order) */
if (H5B2_index(bt2, H5_ITER_INC, idx, find_cb, &idx) < 0)
FAIL_STACK_ERROR
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
/* (in decreasing order) */
if (H5B2_index(bt2, H5_ITER_DEC, idx, find_dec_cb, &idx) < 0)
FAIL_STACK_ERROR
@@ -3210,7 +3210,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3226,7 +3226,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in B-tree with 1 record */
+ /* Attempt to find existent record in B-tree with 1 record */
find.key = 42;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
FAIL_STACK_ERROR
@@ -3240,7 +3240,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL);
@@ -3275,7 +3275,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &modify, modify_rec_cb, &modify) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3305,7 +3305,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL);
@@ -3360,7 +3360,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3369,14 +3369,14 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several records */
+ /* Attempt to find existent record in level-0 B-tree with several records */
find.key = 56;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
TEST_ERROR
if (find.val != 12)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL);
@@ -3448,7 +3448,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, modify_rec_cb, &modify) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
find.key = 41;
find.val = (hsize_t)-1;
@@ -3457,14 +3457,14 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several record */
+ /* Attempt to find existent record in level-0 B-tree with several record */
find.key = 56;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
TEST_ERROR
if (find.val != 22)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL);
@@ -3692,7 +3692,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (idx.key != (INSERT_SPLIT_ROOT_NREC_REC + 2))
TEST_ERROR
- /* Attempt to find non-existant record in level-1 B-tree */
+ /* Attempt to find non-existent record in level-1 B-tree */
/* (Should not be found, but not fail) */
find.key = 800;
find.val = (hsize_t)-1;
@@ -3701,7 +3701,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-1 B-tree */
+ /* Attempt to find existent record in root of level-1 B-tree */
find.key = 33;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -3711,7 +3711,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (find.val != 67)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-1 B-tree */
+ /* Attempt to find existent record in leaf of level-1 B-tree */
find.key = 56;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -3721,7 +3721,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (find.val != 113)
TEST_ERROR
- /* Attempt to index non-existant record in level-1 B-tree */
+ /* Attempt to index non-existent record in level-1 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC + 2), index_rec_cb, NULL);
@@ -4478,7 +4478,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4487,7 +4487,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1347;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4500,7 +4500,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 513;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4513,7 +4513,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4526,7 +4526,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -4643,7 +4643,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4652,7 +4652,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1344;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4665,7 +4665,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 512;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4678,7 +4678,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4691,7 +4691,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -4818,7 +4818,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 6))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4827,7 +4827,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1345;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4840,7 +4840,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 513;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4853,7 +4853,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4866,7 +4866,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -5051,7 +5051,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
if (iter.key != INSERT_MANY_REC)
TEST_ERROR
- /* Attempt to find non-existant record in level-4 B-tree */
+ /* Attempt to find non-existent record in level-4 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_MANY_REC * 2;
find.val = (hsize_t)-1;
@@ -5066,14 +5066,14 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)(HDrandom() % INSERT_MANY_REC);
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
FAIL_STACK_ERROR
if (find.val != (find.key * 2))
TEST_ERROR
} /* end for */
- /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */
+ /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY_REC * 3), find_rec_cb, NULL);
@@ -5102,7 +5102,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)-1;
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
/* (in increasing order) */
if (H5B2_index(bt2, H5_ITER_INC, idx, index_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5115,7 +5115,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)-1;
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
/* (in decreasing order) */
if (H5B2_index(bt2, H5_ITER_DEC, idx, index_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5234,7 +5234,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
- TESTING("B-tree remove: non-existant record from 1 record B-tree");
+ TESTING("B-tree remove: non-existent record from 1 record B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5261,7 +5261,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!H5F_addr_defined(root_addr))
TEST_ERROR
- /* Attempt to remove a non-existant record from a B-tree with 1 record */
+ /* Attempt to remove a non-existent record from a B-tree with 1 record */
record = 0;
H5E_BEGIN_TRY
{
@@ -5275,7 +5275,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
/* Attempt to remove a record from a B-tree with 1 record */
- TESTING("B-tree remove: existant record from 1 record B-tree");
+ TESTING("B-tree remove: existent record from 1 record B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5339,8 +5339,8 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
- /* Attempt to remove a non-existant record from a level-0 B-tree with mult. record */
- TESTING("B-tree remove: non-existant record from level-0 B-tree");
+ /* Attempt to remove a non-existent record from a level-0 B-tree with mult. record */
+ TESTING("B-tree remove: non-existent record from level-0 B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5359,7 +5359,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
/* Attempt to remove a record from a level-0 B-tree with mult. record */
- TESTING("B-tree remove: mult. existant records from level-0 B-tree");
+ TESTING("B-tree remove: mult. existent records from level-0 B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5517,7 +5517,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt
herr_t ret; /* Generic error return value */
/* B-tree record removal tests */
- TESTING("B-tree remove: non-existant record from level-1 B-tree");
+ TESTING("B-tree remove: non-existent record from level-1 B-tree");
/* Create the file for the test */
if (create_file(&file, &f, fapl) < 0)
@@ -5550,7 +5550,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt
if (!H5F_addr_defined(root_addr))
TEST_ERROR
- /* Attempt to remove a non-existant record from a B-tree with 1 record */
+ /* Attempt to remove a non-existent record from a B-tree with 1 record */
record = (INSERT_SPLIT_ROOT_NREC * 2) + 1;
H5E_BEGIN_TRY
{
@@ -9605,7 +9605,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa
/*
* Test modifying records
*/
- TESTING("B-tree modify: attempt to modify non-existant record");
+ TESTING("B-tree modify: attempt to modify non-existent record");
/* Create the file for the test */
if (create_file(&file, &f, fapl) < 0)
@@ -9628,7 +9628,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa
if (bt2_stat.depth != 2)
TEST_ERROR
- /* Attempt to modify a non-existant record */
+ /* Attempt to modify a non-existent record */
record = 3;
modify = 4;
H5E_BEGIN_TRY
diff --git a/test/cache.c b/test/cache.c
index aa58777..95a985b 100644
--- a/test/cache.c
+++ b/test/cache.c
@@ -2774,8 +2774,8 @@ write_permitted_check(int
* Test the behaviour with different flags.
*
* This test was added primarily to test basic insert
- * pinned entry functionallity, but I through in explicit
- * tests for other functionallity that is tested implicitly
+ * pinned entry functionality, but I through in explicit
+ * tests for other functionality that is tested implicitly
* elsewhere.
*
* Return: void
@@ -2893,7 +2893,7 @@ check_insert_entry(unsigned paged)
}
}
- /* Thats all we can get from H5C_get_entry_status().
+ /* That's all we can get from H5C_get_entry_status().
* Now start looking at the cache data structures directly.
*/
@@ -5131,7 +5131,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr)
*
* 10/8/07 -- JRM
* Added a resize operation to this test to satisfy the new
- * requiremnt that any resize of an entry on flush will always
+ * requirement that any resize of an entry on flush will always
* be accompanied by a resize. Note that as a result, this
* test becomes redundant with later tests.
*/
@@ -5223,7 +5223,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr)
*
* 10/8/07 -- JRM
* Added a resize operation to this test to satisfy the new
- * requiremnt that any resize of an entry on flush will always
+ * requirement that any resize of an entry on flush will always
* be accompanied by a resize. Note that as a result, this
* test becomes redundant with later tests.
*/
@@ -9601,7 +9601,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
*
* (VET, 9) Y 10 KB N N - dirty (VET, 8)
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -10154,7 +10154,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
* returns to the serialize function for (VET, 9).
*
* When the serialize function for (VET, 9) is done with (VET, 8), it
- * calls H5C_unprotect(VET, 8), which markes (VET, 8) as dirty and
+ * calls H5C_unprotect(VET, 8), which marks (VET, 8) as dirty and
* unprotected, and places it at the head of the LRU.
*
* (VET, 0) is the next item on the LRU -- it is clean and is therefore
@@ -10553,7 +10553,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
* Added tests for pinned entries.
*
* JRM -- 5/17/06
- * Complete reqrite of pinned entry tests to accomodate
+ * Complete reqrite of pinned entry tests to accommodate
* the new H5C_mark_entry_dirty() call.
*
*-------------------------------------------------------------------------
@@ -12978,7 +12978,7 @@ check_expunge_entry(unsigned paged)
* Function: check_multiple_read_protect()
*
* Purpose: Verify that multiple, simultaneous read protects of a
- * single entry perform as expectd.
+ * single entry perform as expected.
*
* Return: void
*
@@ -13324,7 +13324,7 @@ check_multiple_read_protect(unsigned paged)
if (pass && (entry_ptr->header.is_pinned)) {
pass = FALSE;
- failure_mssg = "enty (0,4) still pinned.\n";
+ failure_mssg = "entry (0,4) still pinned.\n";
}
unprotect_entry(file_ptr, 0, 4, /* (0,4) unprotect */
@@ -15527,7 +15527,7 @@ check_destroy_protected_err(unsigned paged)
* Function: check_duplicate_insert_err()
*
* Purpose: Verify that an attempt to insert and entry that is
- * alread in the cache will generate an error.
+ * already in the cache will generate an error.
*
* Return: void
*
@@ -18083,7 +18083,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- force the hit rate to 100% again -- should be steady
+ /* eighth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -18179,7 +18179,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- if (pass) { /* twelth epoch -- at minimum size so no more ageouts */
+ if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */
rpt_fcn_called = FALSE;
i = 0;
while (pass && (i < 1000)) {
@@ -18460,7 +18460,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- still 100% hit rate
+ /* eighth epoch -- still 100% hit rate
*/
if (pass) {
@@ -18563,7 +18563,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* twelth epoch -- force the hit rate to 100% again -- should be steady
+ /* twelfth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -18939,7 +18939,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- force the hit rate to 100% again -- should be steady
+ /* eighth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -19039,7 +19039,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- if (pass) { /* twelth epoch -- at minimum size so no more ageouts */
+ if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */
rpt_fcn_called = FALSE;
i = 0;
@@ -19690,7 +19690,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* twelth epoch -- hit rate 1.0 -- decrease as before.
+ /* twelfth epoch -- hit rate 1.0 -- decrease as before.
*/
if (pass) {
@@ -20703,7 +20703,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* protect and unprotect two more large entries -- shouldnt trigger a
+ /* protect and unprotect two more large entries -- shouldn't trigger a
* flash cache size increase.
*/
if (pass) {
@@ -23263,7 +23263,7 @@ check_auto_cache_resize_disable(unsigned paged)
* mode (add space), which is triggered whenever the size of a newly
* loaded or inserted entry, or the delta between old and new entry
* sizes exceeds some fraction of the current maximum cache size, and
- * the cache doesn't have enough free space to accomodate the new/
+ * the cache doesn't have enough free space to accommodate the new/
* resize entry without performing evictions. The range of permissible
* values for the flash_threshold (0.1 to 1.0 as of this writing), and
* for the flash_multiple (0.1 to 10.0) do not permit the facility to
@@ -23272,7 +23272,7 @@ check_auto_cache_resize_disable(unsigned paged)
* current max_cache_size equal to max_size.
*
* We have already tested the latter in check_auto_cache_resize(), so
- * we need only thest the former here. Do this by disabling flash
+ * we need only test the former here. Do this by disabling flash
* cache size increments via the flash_incr_mode, and then creating
* situations that would trigger flash cache size increases were that
* code enabled.
@@ -26345,7 +26345,7 @@ check_auto_cache_resize_input_errs(unsigned paged)
/*-------------------------------------------------------------------------
* Function: check_auto_cache_resize_aux_fcns()
*
- * Purpose: Verify that the auxilary functions associated with
+ * Purpose: Verify that the auxiliary functions associated with
* the automatic cache resize capability are operating
* correctly. These functions are:
*
@@ -26422,13 +26422,13 @@ check_auto_cache_resize_aux_fcns(unsigned paged)
};
if (paged)
- TESTING("automatic cache resize auxilary functions (paged aggregation)")
+ TESTING("automatic cache resize auxiliary functions (paged aggregation)")
else
- TESTING("automatic cache resize auxilary functions")
+ TESTING("automatic cache resize auxiliary functions")
pass = TRUE;
- /* allocate a cache, and then test the various auxilary functions.
+ /* allocate a cache, and then test the various auxiliary functions.
*/
if (pass) {
@@ -27075,7 +27075,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged)
if (file_ptr == NULL) {
pass = FALSE;
- failure_mssg = "bad return from cache intialization.\n";
+ failure_mssg = "bad return from cache initialization.\n";
}
else
cache_ptr = file_ptr->shared->cache;
@@ -27294,7 +27294,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged)
H5C__DIRTIED_FLAG); /* unsigned int flags */
}
- /* Expected staus is that after each insertion, the entry
+ /* Expected status is that after each insertion, the entry
* inserted 26 insertions ago has been flushed, and the
* entry currently getting inserted is now in the cache and
* dirty.
@@ -29699,7 +29699,7 @@ check_flush_deps_err(unsigned paged)
result =
H5C_destroy_flush_dependency(&((entries[entry_type])[0]), &((entries[entry_type])[1]));
if (result != FAIL)
- CACHE_ERROR("Destroying [non-existant] dependency when parent isn't protected")
+ CACHE_ERROR("Destroying [non-existent] dependency when parent isn't protected")
break;
/* Verify that parent entry has flush dependency */
@@ -33055,7 +33055,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged)
if (file_ptr == NULL) {
pass = FALSE;
- failure_mssg = "bad return from cache intialization.\n";
+ failure_mssg = "bad return from cache initialization.\n";
}
else
cache_ptr = file_ptr->shared->cache;
@@ -33565,7 +33565,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged)
* an entry that is no longer in the cache, and which may have
* been deleted.
*
- * This function contans tests for correct handling on this
+ * This function contains tests for correct handling on this
* situation.
*
* Do nothing if pass is FALSE on entry.
@@ -33796,7 +33796,7 @@ cedds__expunge_dirty_entry_in_flush_test(H5F_t *file_ptr)
*
* (HET, 3) Y 16 KB Y N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -34046,7 +34046,7 @@ cedds__H5C_make_space_in_cache(H5F_t *file_ptr)
*
* (HET, 3) Y 16 KB N N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -34461,7 +34461,7 @@ cedds__H5C__autoadjust__ageout__evict_aged_out_entries(H5F_t *file_ptr)
*
* (MET, 1-31) Y 64 KB N N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
diff --git a/test/cache_api.c b/test/cache_api.c
index 55aaa0a..f79a531 100644
--- a/test/cache_api.c
+++ b/test/cache_api.c
@@ -662,7 +662,7 @@ check_file_mdc_api_calls(unsigned paged, hid_t fcpl_id)
* configured as per the default both by looking at its internal
* configuration, and via the H5Fget_mdc_config() call.
*
- * Then set serveral different configurations, and verify that
+ * Then set several different configurations, and verify that
* they took as per above.
*/
@@ -1054,7 +1054,7 @@ mdc_api_call_smoke_check(int express_test, unsigned paged, hid_t fcpl_id)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -1885,7 +1885,7 @@ check_file_mdc_api_errs(unsigned paged, hid_t fcpl_id)
pass = TRUE;
- /* Create a file for test purposes, and veify that its metadata cache
+ /* Create a file for test purposes, and verify that its metadata cache
* set to the default MDC configuration.
*/
@@ -2254,7 +2254,7 @@ main(void)
} /* end if */
/* Test with paged aggregation enabled or not */
- /* The "my_fcpl" passed to each test has the paged or non-paged strategy set up accordinly */
+ /* The "my_fcpl" passed to each test has the paged or non-paged strategy set up accordingly */
for (paged = FALSE; paged <= TRUE; paged++) {
hid_t my_fcpl = fcpl_id;
diff --git a/test/cache_common.c b/test/cache_common.c
index 59fda94..61c5565 100644
--- a/test/cache_common.c
+++ b/test/cache_common.c
@@ -1523,7 +1523,7 @@ add_flush_op(int target_type, int target_idx, int op_code, int type, int idx, hb
* Otherwise, set up a pinned entry dependency so we can
* test the pinned entry modifications to the flush routine.
*
- * Given the types and indicies of the pinned and pinning
+ * Given the types and indices of the pinned and pinning
* entries, add the pinned entry to the list of pinned
* entries in the pinning entry, increment the
* pinning reference count of the pinned entry, and
@@ -1768,7 +1768,7 @@ execute_flush_op(H5F_t *file_ptr, struct test_entry_t *entry_ptr, struct flush_o
case FLUSH_OP__EXPUNGE:
/* the expunge flush op exists to allow us to simulate the
- * case in which an entry is removed from the cashe as the
+ * case in which an entry is removed from the cache as the
* the result of the flush of a second entry. At present,
* this can only happen via the take ownership flag, but
* we will make this test feature more general to as to make
@@ -2313,7 +2313,7 @@ verify_clean(void)
* Function: verify_entry_status
*
* Purpose: Verify that a list of entries have the expected status.
- * If any discrepencies are found, set the failure message
+ * If any discrepancies are found, set the failure message
* and set pass to FALSE.
*
* Do nothing if pass is FALSE on entry.
@@ -3833,7 +3833,7 @@ row_major_scan_forward(H5F_t *file_ptr, int32_t max_index, int32_t lag, hbool_t
HDassert((!cache_ptr->slist_enabled) ||
(cache_ptr->slist_size == cache_ptr->dirty_index_size));
} /* end if */
- } /* end elsef */
+ } /* end elseif */
if (verbose)
HDfprintf(stdout, "\n");
@@ -5241,7 +5241,7 @@ resize_configs_are_equal(const H5C_auto_size_ctl_t *a, const H5C_auto_size_ctl_t
* has both internal and external configuration matching
* *config_ptr.
*
- * Do nothin on success. On failure, set pass to FALSE, and
+ * Do nothing on success. On failure, set pass to FALSE, and
* load an error message into failue_mssg. Note that
* failure_msg is assumed to be at least 128 bytes in length.
*
diff --git a/test/cache_image.c b/test/cache_image.c
index 90fe7af..d2313ba 100644
--- a/test/cache_image.c
+++ b/test/cache_image.c
@@ -136,7 +136,7 @@ create_datasets(hid_t file_id, int min_dset, int max_dset)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -470,7 +470,7 @@ delete_datasets(hid_t file_id, int min_dset, int max_dset)
* Set pass to FALSE and issue a suitable failure
* message if either the file contains a metadata cache image
* superblock extension and mdci_sbem_expected is TRUE, or
- * vise versa.
+ * vice versa.
*
* If mdci_sbem_expected is TRUE, also verify that the metadata
* cache has been advised of this.
@@ -544,7 +544,7 @@ open_hdf5_file(hbool_t create_file, hbool_t mdci_sbem_expected, hbool_t read_onl
if (show_progress)
HDfprintf(stdout, "%s: cp = %d.\n", fcn_name, cp++);
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = h5_fileaccess();
@@ -879,7 +879,7 @@ attempt_swmr_open_hdf5_file(const hbool_t create_file, const hbool_t set_mdci_fa
H5AC_cache_image_config_t cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE,
H5AC__CACHE_IMAGE__ENTRY_AGEOUT__NONE};
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = h5_fileaccess();
@@ -1505,7 +1505,7 @@ check_cache_image_ctl_flow_1(hbool_t single_file_vfd)
* processed as part of the first protect operation after the
* superblock is loaded.)
*
- * In this particular test, we preform the following operations:
+ * In this particular test, we perform the following operations:
*
* 1) Create a HDF5 file with the cache image FAPL entry.
*
@@ -7001,7 +7001,7 @@ cache_image_api_error_check_4(hbool_t single_file_vfd)
* image.
*
* The objective of this test is to create a test file
- * with both non-empty self referential presistant
+ * with both non-empty self referential persistent
* free space managers, and a cache image, and then
* verify that this situation is handled correctly if
* H5Fget_free_sections() is called before the metadata
diff --git a/test/cache_tagging.c b/test/cache_tagging.c
index d03caec..b01fc34 100644
--- a/test/cache_tagging.c
+++ b/test/cache_tagging.c
@@ -443,7 +443,7 @@ check_file_creation_tags(hid_t fcpl_id, int type)
/* Variable Declarations */
hid_t fid = -1; /* File Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose test outout */
+ int verbose = FALSE; /* verbose test output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -547,7 +547,7 @@ check_file_open_tags(hid_t fcpl, int type)
/* Variable Declarations */
hid_t fid = -1; /* File Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag; /* Root Group Tag */
@@ -677,7 +677,7 @@ check_group_creation_tags(void)
hid_t fid = -1; /* File Identifier */
hid_t gid = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = HADDR_UNDEF; /* Root Group Tag */
@@ -800,7 +800,7 @@ check_multi_group_creation_tags(void)
hid_t fid = -1; /* File Identifier */
hid_t gid = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
char gname[16]; /* group name buffer */
int i = 0; /* iterator */
@@ -952,7 +952,7 @@ check_link_iteration_tags(void)
hid_t sid = -1; /* Group Identifier */
hid_t did = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int i = 0; /* iterator */
haddr_t root_tag = 0; /* Root Group Tag Value */
@@ -1094,7 +1094,7 @@ check_dense_attribute_tags(void)
hid_t did = -1; /* Group Identifier */
hid_t dcpl = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int i = 0; /* iterator */
hid_t fapl = -1; /* File access property list */
@@ -1455,7 +1455,7 @@ check_attribute_creation_tags(hid_t fcpl, int type)
hid_t gid = -1; /* Group Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0; /* Root group tag */
@@ -1620,7 +1620,7 @@ check_attribute_open_tags(hid_t fcpl, int type)
hid_t gid = -1; /* Group Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -1787,7 +1787,7 @@ check_attribute_rename_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -2000,7 +2000,7 @@ check_attribute_delete_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -2191,7 +2191,7 @@ check_dataset_creation_tags(hid_t fcpl, int type)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2351,7 +2351,7 @@ check_dataset_creation_earlyalloc_tags(hid_t fcpl, int type)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2517,7 +2517,7 @@ check_dataset_open_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2669,7 +2669,7 @@ check_dataset_write_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2836,7 +2836,7 @@ check_attribute_write_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -3022,7 +3022,7 @@ check_dataset_read_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3184,7 +3184,7 @@ check_dataset_size_retrieval(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3348,7 +3348,7 @@ check_dataset_extend_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3795,7 +3795,7 @@ check_link_removal_tags(hid_t fcpl, int type)
hid_t sid = -1; /* Dataspace Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3984,7 +3984,7 @@ check_link_getname_tags(void)
hid_t sid = -1; /* Dataspace Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -4161,7 +4161,7 @@ check_external_link_creation_tags(void)
hid_t fid2 = -1; /* File Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -4288,7 +4288,7 @@ check_external_link_open_tags(void)
hid_t gid = -1; /* Dataspace Identifier */
hid_t xid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
H5O_native_info_t ninfo; /* Native object info struct */
hid_t fapl = -1; /* File access prop list */
@@ -4579,7 +4579,7 @@ main(void)
/* Create a standard file creation property list */
fcpl_default = H5Pcreate(H5P_FILE_CREATE);
- /* Create an fcpl with shared messages and file space managment enabled */
+ /* Create an fcpl with shared messages and file space management enabled */
fcpl_shmesg_all = H5Pcreate(H5P_FILE_CREATE);
H5Pset_shared_mesg_nindexes(fcpl_shmesg_all, 1);
H5Pset_shared_mesg_index(fcpl_shmesg_all, 0, H5O_SHMESG_ALL_FLAG, 20);
diff --git a/test/chunk_info.c b/test/chunk_info.c
index 5c16a7d..f60cddd 100644
--- a/test/chunk_info.c
+++ b/test/chunk_info.c
@@ -291,7 +291,7 @@ index_type_str(H5D_chunk_index_t idx_type)
/*-------------------------------------------------------------------------
* Function: verify_selected_chunks (helper function)
*
- * Purpose: Reads the chunks within the boundery {start,end} and verify
+ * Purpose: Reads the chunks within the boundary {start,end} and verify
* the values against the populated data.
*
* Return: Success: SUCCEED
@@ -321,7 +321,7 @@ verify_selected_chunks(hid_t dset, hid_t plist, const hsize_t *start, const hsiz
for (jj = 0; jj < CHUNK_NY; jj++)
expected_buf[n][ii][jj] = (int)(ii * jj) + 1;
- /* Read each chunk within the boundery of {start,end} and verify the
+ /* Read each chunk within the boundary of {start,end} and verify the
values against the expected data */
chk_index = 0;
for (ii = start[0]; ii < end[0]; ii++)
diff --git a/test/cmpd_dset.c b/test/cmpd_dset.c
index 3235056..024b6c0 100644
--- a/test/cmpd_dset.c
+++ b/test/cmpd_dset.c
@@ -713,7 +713,7 @@ test_compound(char *filename, hid_t fapl)
/*
*######################################################################
* Step 12: Write an array into the middle third of the dataset
- * initializeing only members `b' and `d' to -1.
+ * initializing only members `b' and `d' to -1.
*/
TESTING("hyperslab part initialized write");
diff --git a/test/cork.c b/test/cork.c
index 1b99384..f4b8d1c 100644
--- a/test/cork.c
+++ b/test/cork.c
@@ -97,7 +97,7 @@ verify_old_dset_cork(void)
hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */
hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */
hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */
- int ** buf = NULL; /* Data bufer (pointers to fake 2D array) */
+ int ** buf = NULL; /* Data buffer (pointers to fake 2D array) */
int * buf_data = NULL; /* Data buffer (actual data) */
int i = 0, j = 0; /* Local index variables */
H5O_info2_t oinfo, oinfo2, oinfo3; /* Object metadata information */
@@ -509,7 +509,7 @@ verify_dset_cork(hbool_t swmr, hbool_t new_format)
hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */
hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */
hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */
- int ** buf = NULL; /* Data bufer (pointers to fake 2D array) */
+ int ** buf = NULL; /* Data buffer (pointers to fake 2D array) */
int * buf_data = NULL; /* Data buffer (actual data) */
int i = 0, j = 0; /* Local index variables */
H5O_info2_t oinfo, oinfo2, oinfo3; /* Object metadata information */
@@ -1883,7 +1883,7 @@ test_dset_cork(hbool_t swmr, hbool_t new_format)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
hid_t fapl = H5I_INVALID_HID; /* File access property list */
- hid_t gid = H5I_INVALID_HID; /* Groupd ID */
+ hid_t gid = H5I_INVALID_HID; /* Grouped ID */
hid_t did1 = H5I_INVALID_HID, did2 = H5I_INVALID_HID; /* Dataset IDs */
hid_t tid1 = H5I_INVALID_HID, tid2 = H5I_INVALID_HID; /* Datatype IDs */
hid_t sid = H5I_INVALID_HID; /* Dataspace ID */
diff --git a/test/direct_chunk.c b/test/direct_chunk.c
index b2b7cba..d4a1e7a 100644
--- a/test/direct_chunk.c
+++ b/test/direct_chunk.c
@@ -686,7 +686,7 @@ filter_bogus1(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= (int)ADD_ON;
buf_left -= sizeof(int);
@@ -722,7 +722,7 @@ filter_bogus2(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ /= (int)FACTOR;
buf_left -= sizeof(int);
@@ -1427,7 +1427,7 @@ error:
* Function: test_direct_chunk_read_no_cache
*
* Purpose: Test the basic functionality of H5Dread_chunk with the
- * chunk cache diabled.
+ * chunk cache disabled.
*
* Return: Success: 0
* Failure: 1
@@ -1963,7 +1963,7 @@ error:
* Function: test_read_unallocated_chunk
*
* Purpose: Tests the H5Dread_chunk and H5Dget_chunk_storage_size with valid
- * offets to chunks that have not been written to the dataset and are
+ * offsets to chunks that have not been written to the dataset and are
* not allocated in the chunk storage on disk.
*
* Return: Success: 0
@@ -2014,7 +2014,7 @@ test_read_unallocated_chunk(hid_t file)
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
FAIL_STACK_ERROR;
- /* Write a single chunk to intialize the chunk storage */
+ /* Write a single chunk to initialize the chunk storage */
HDmemset(direct_buf, 0, CHUNK_NX * CHUNK_NY * sizeof(int));
offset[0] = 0;
offset[1] = 0;
@@ -2031,7 +2031,7 @@ test_read_unallocated_chunk(hid_t file)
offset[0] = i * CHUNK_NX;
offset[1] = j * CHUNK_NY;
- /* Read a non-existant chunk using the direct read function. */
+ /* Read a non-existent chunk using the direct read function. */
H5E_BEGIN_TRY
{
status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, &direct_buf);
@@ -2042,7 +2042,7 @@ test_read_unallocated_chunk(hid_t file)
if (status != -1)
TEST_ERROR
- /* Query the size of the non-existant chunk */
+ /* Query the size of the non-existent chunk */
direct_chunk_nbytes = ULONG_MAX;
H5E_BEGIN_TRY
{
diff --git a/test/dsets.c b/test/dsets.c
index 0423898..b1ebf58 100644
--- a/test/dsets.c
+++ b/test/dsets.c
@@ -2009,7 +2009,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2076,7 +2076,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2140,7 +2140,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2207,7 +2207,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -3242,7 +3242,7 @@ test_nbit_int(hid_t file)
power = HDpow(2.0, (double)(precision - 1));
orig_data[i][j] = (int)(((long long)HDrandom() % (long long)power) << offset);
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -3820,7 +3820,7 @@ test_nbit_compound(hid_t file)
orig_data[i][j].s = (short)(((long long)HDrandom() % (long long)power) << offset[2]);
orig_data[i][j].f = float_val[i][j];
- /* some even-numbered integer values are negtive */
+ /* some even-numbered integer values are negative */
if ((i * size[1] + j + 1) % 2 == 0) {
orig_data[i][j].i = -orig_data[i][j].i;
orig_data[i][j].s = (short)-orig_data[i][j].s;
@@ -4534,7 +4534,7 @@ test_nbit_int_size(hid_t file)
goto error;
}
- /* Initiliaze data buffer with random data within correct range
+ /* Initialize data buffer with random data within correct range
* corresponding to the memory datatype's precision and offset.
*/
for (i = 0; i < DSET_DIM1; i++)
@@ -4747,7 +4747,7 @@ test_nbit_flt_size(hid_t file)
} /* end if */
/*
- * Initiliaze data buffer with random data
+ * Initialize data buffer with random data
*/
for (i = 0; i < DSET_DIM1; i++)
for (j = 0; j < DSET_DIM2; j++)
@@ -4906,7 +4906,7 @@ test_scaleoffset_int(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = HDrandom() % 10000;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5050,7 +5050,7 @@ test_scaleoffset_int_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (int)HDrandom() % 10000;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -5172,7 +5172,7 @@ test_scaleoffset_float(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = (float)(HDrandom() % 100000) / 1000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5318,7 +5318,7 @@ test_scaleoffset_float_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (float)(HDrandom() % 100000) / 1000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -5439,7 +5439,7 @@ test_scaleoffset_double(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = (float)(HDrandom() % 10000000) / 10000000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5585,7 +5585,7 @@ test_scaleoffset_double_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (float)(HDrandom() % 10000000) / 10000000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -8897,7 +8897,7 @@ test_chunk_cache(hid_t fapl)
if (H5Pclose(dapl2) < 0)
FAIL_STACK_ERROR
- /* Similary, test use of H5Dcreate2 with H5P_DEFAULT */
+ /* Similarly, test use of H5Dcreate2 with H5P_DEFAULT */
if (H5Dclose(dsid) < 0)
FAIL_STACK_ERROR
if ((dsid = H5Dcreate2(fid, "dset2", H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT)) < 0)
@@ -11982,7 +11982,7 @@ error:
* Purpose: Checks that a file created with either:
* (a) SWMR-write + non-latest-format
* (b) write + latest format
- * will generate datset with latest chunk indexing type.
+ * will generate dataset with latest chunk indexing type.
*
* Return: Success: 0
* Failure: -1
@@ -14716,7 +14716,7 @@ test_versionbounds(void)
hid_t dcpl = -1; /* Dataset creation property list */
hid_t srcspace = -1; /* Source dataspaces */
hid_t vspace = -1; /* Virtual dset dataspaces */
- hid_t srcdset = -1; /* Source datset */
+ hid_t srcdset = -1; /* Source dataset */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[1] = {3}; /* Data space current size */
char srcfilename[FILENAME_BUF_SIZE];
@@ -14852,7 +14852,7 @@ error:
/*-----------------------------------------------------------------------------
* Function: test_object_header_minimization_dcpl
*
- * Purpose: Test the "datset object header minimization" property as part of
+ * Purpose: Test the "dataset object header minimization" property as part of
* the DCPL.
*
* Return: Success/pass: 0
@@ -15195,7 +15195,7 @@ main(void)
double rdcc_w0;
int nerrors = 0;
const char *envval;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
int i;
/* Don't run this test using certain file drivers */
@@ -15203,7 +15203,7 @@ main(void)
if (envval == NULL)
envval = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0);
/* Set the random # seed */
diff --git a/test/dt_arith.c b/test/dt_arith.c
index 3f47153..e6927f5 100644
--- a/test/dt_arith.c
+++ b/test/dt_arith.c
@@ -27,7 +27,7 @@
#define FP_EPSILON 0.000001F
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
@@ -223,7 +223,7 @@ static int without_hardware_g = 0;
/*The number of values if multiplied by 10 for each step.*/ \
num_norm = (SRC_MAX_10_EXP - SRC_MIN_10_EXP); \
/*Reduce the number of values by 2^factor. MULTIPLY=10^(2^factor). Using this algorithm \
- *instead of arithmatic operation to avoid any conversion*/ \
+ *instead of arithmetic operation to avoid any conversion*/ \
num_norm >>= factor; \
\
/*Total number of values*/ \
@@ -640,7 +640,7 @@ test_particular_fp_integer(void)
float src_f = (float)INT_MAX;
int dst_i;
int fill_value = 13;
- int endian; /*endianess */
+ int endian; /*endianness */
unsigned int fails_this_test = 0;
size_t j;
@@ -818,7 +818,7 @@ test_derived_flt(void)
size_t src_size, dst_size;
unsigned char *buf = NULL, *saved_buf = NULL;
int * aligned = NULL;
- int endian; /*endianess */
+ int endian; /*endianness */
size_t nelmts = NTESTELEM;
unsigned int fails_this_test = 0;
const size_t max_fails = 40; /*max number of failures*/
@@ -1267,7 +1267,7 @@ test_derived_integer(void)
char filename[1024];
size_t src_size, dst_size;
unsigned char *buf = NULL, *saved_buf = NULL;
- int endian; /*endianess */
+ int endian; /*endianness */
size_t nelmts = NTESTELEM;
unsigned int fails_this_test = 0;
const size_t max_fails = 40; /*max number of failures*/
@@ -1591,7 +1591,7 @@ test_conv_int_1(const char *name, hid_t src, hid_t dst)
dtype_t src_type, dst_type; /*data types */
const char * src_type_name = NULL; /*source type name */
const char * dst_type_name = NULL; /*destination type name */
- int endian; /*machine endianess */
+ int endian; /*machine endianness */
size_t src_size, dst_size; /*type sizes */
unsigned char * buf = NULL; /*buffer for conversion */
unsigned char * saved = NULL; /*original values */
@@ -2831,8 +2831,8 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst)
int overflow = 0; /*overflow occurred */
int uflow = 0; /*underflow debug counters*/
size_t j, k; /*counters */
- int sendian; /* source type endianess */
- int dendian; /* Destination type endianess */
+ int sendian; /* source type endianness */
+ int dendian; /* Destination type endianness */
size_t dst_ebias; /* Destination type's exponent bias */
size_t src_epos; /* Source type's exponent position */
size_t src_esize; /* Source type's exponent size */
@@ -3147,7 +3147,7 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst)
/*
* Assume same if both results are NaN. There are many NaN bit
- * patterns and the software doesn't attemt to emulate the
+ * patterns and the software doesn't attempt to emulate the
* hardware in this regard. Instead, software uses a single bit
* pattern for NaN by setting the significand to all ones.
*/
@@ -3426,8 +3426,8 @@ test_conv_int_fp(const char *name, int run_test, hid_t src, hid_t dst)
dtype_t dst_type; /*data types */
const char * src_type_name = NULL; /*source type name */
const char * dst_type_name = NULL; /*destination type name */
- int sendian; /*source endianess */
- int dendian; /*destination endianess */
+ int sendian; /*source endianness */
+ int dendian; /*destination endianness */
size_t src_size, dst_size; /*type sizes */
unsigned char * buf = NULL; /*buffer for conversion */
unsigned char * saved = NULL; /*original values */
diff --git a/test/dtypes.c b/test/dtypes.c
index bf2ad5a..59ad0e3 100644
--- a/test/dtypes.c
+++ b/test/dtypes.c
@@ -29,7 +29,7 @@
#define ARRAY_DIM 4
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
@@ -2637,7 +2637,7 @@ test_compound_13(void)
data_out.x[u] = (unsigned char)u;
data_out.y = 99.99F;
- /* Set latest_format in access propertly list to enable the latest
+ /* Set latest_format in access property list to enable the latest
* compound datatype format.
*/
if ((fapl_id = H5Pcreate(H5P_FILE_ACCESS)) < 0)
@@ -3792,7 +3792,7 @@ error:
*
* Modifications:
* Raymond Lu
- * Wednesday, Febuary 9, 2005
+ * Wednesday, February 9, 2005
* Added test for H5Tenum_valueof, H5Tenum_nameof, and
* H5Tget_member_value.
*-------------------------------------------------------------------------
@@ -6845,7 +6845,7 @@ test_set_order(void)
{
hid_t dtype; /* Datatype ID */
H5T_order_t order; /* Byte order */
- hsize_t dims[2] = {3, 4}; /* Array dimenstions */
+ hsize_t dims[2] = {3, 4}; /* Array dimensions */
herr_t ret; /* Generic return value */
TESTING("H5Tset/get_order");
@@ -7093,7 +7093,7 @@ test_set_order_compound(hid_t fapl)
hid_t file = -1;
hid_t cmpd = -1, memb_cmpd = -1, memb_array1 = -1, memb_array2 = -1, cmpd_array = -1;
hid_t vl_id = -1;
- hsize_t dims[2] = {3, 4}; /* Array dimenstions */
+ hsize_t dims[2] = {3, 4}; /* Array dimensions */
char filename[1024];
herr_t ret; /* Generic return value */
@@ -7948,7 +7948,7 @@ test_deprec(hid_t fapl)
unsigned u; /* Local index variable */
herr_t status; /* Generic routine value */
- TESTING("deprected API routines for datatypes");
+ TESTING("deprecated API routines for datatypes");
/* Create an array datatype with an atomic base type */
/* (dimension permutations allowed, but not stored) */
diff --git a/test/earray.c b/test/earray.c
index 65664fb..4ed5b7f 100644
--- a/test/earray.c
+++ b/test/earray.c
@@ -1284,7 +1284,7 @@ typedef struct eiter_fw_t {
/*-------------------------------------------------------------------------
* Function: eiter_fw_init
*
- * Purpose: Initialize element interator (forward iteration)
+ * Purpose: Initialize element iterator (forward iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1428,7 +1428,7 @@ eiter_fw_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa
/*-------------------------------------------------------------------------
* Function: eiter_fw_term
*
- * Purpose: Shut down element interator (forward iteration)
+ * Purpose: Shut down element iterator (forward iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1472,7 +1472,7 @@ typedef struct eiter_rv_t {
/*-------------------------------------------------------------------------
* Function: eiter_rv_init
*
- * Purpose: Initialize element interator (reverse iteration)
+ * Purpose: Initialize element iterator (reverse iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1653,7 +1653,7 @@ eiter_rv_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa
/*-------------------------------------------------------------------------
* Function: eiter_rv_term
*
- * Purpose: Shut down element interator (reverse iteration)
+ * Purpose: Shut down element iterator (reverse iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1694,7 +1694,7 @@ typedef struct eiter_rnd_t {
/*-------------------------------------------------------------------------
* Function: eiter_rnd_init
*
- * Purpose: Initialize element interator (random iteration)
+ * Purpose: Initialize element iterator (random iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1803,7 +1803,7 @@ eiter_rnd_max(const void *in_eiter)
/*-------------------------------------------------------------------------
* Function: eiter_rnd_term
*
- * Purpose: Shut down element interator (random iteration)
+ * Purpose: Shut down element iterator (random iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1843,7 +1843,7 @@ static const earray_iter_t ea_iter_rnd = {
/*-------------------------------------------------------------------------
* Function: eiter_rnd2_init
*
- * Purpose: Initialize element interator (random #2 iteration)
+ * Purpose: Initialize element iterator (random #2 iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1926,7 +1926,7 @@ typedef struct eiter_cyc_t {
/*-------------------------------------------------------------------------
* Function: eiter_cyc_init
*
- * Purpose: Initialize element interator (cyclic iteration)
+ * Purpose: Initialize element iterator (cyclic iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -2019,7 +2019,7 @@ eiter_cyc_max(const void *in_eiter)
/*-------------------------------------------------------------------------
* Function: eiter_cyc_term
*
- * Purpose: Shut down element interator (cyclic iteration)
+ * Purpose: Shut down element iterator (cyclic iteration)
*
* Return: Success: 0
* Failure: -1
diff --git a/test/enc_dec_plist_cross_platform.c b/test/enc_dec_plist_cross_platform.c
index 9ea5cf9..3030d7c 100644
--- a/test/enc_dec_plist_cross_platform.c
+++ b/test/enc_dec_plist_cross_platform.c
@@ -28,7 +28,7 @@ int
main(void)
{
if (VERBOSE_MED)
- HDprintf("Encode/Decode property list endianess\n");
+ HDprintf("Encode/Decode property list endianness\n");
/******* ENCODE/DECODE DCPLS *****/
TESTING("Default DCPL Encoding/Decoding");
diff --git a/test/enum.c b/test/enum.c
index 3923892..512fd9d 100644
--- a/test/enum.c
+++ b/test/enum.c
@@ -552,7 +552,7 @@ test_value_dsnt_exist(void)
if ((datatype_id = H5Tenum_create(H5T_NATIVE_INT)) < 0)
goto error;
- /* These calls should fail, since no memebrs exist yet */
+ /* These calls should fail, since no members exist yet */
if (H5Tenum_valueof(datatype_id, "SAX", &val) >= 0)
goto error;
val = 3;
diff --git a/test/external.c b/test/external.c
index 1e4e32a..11185b3 100644
--- a/test/external.c
+++ b/test/external.c
@@ -1217,7 +1217,7 @@ test_path_relative_cwd(hid_t fapl)
}
H5E_END_TRY;
if (dset3 >= 0)
- FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded");
+ FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded");
/* Read the entire dataset and compare with the original */
HDmemset(whole, 0, sizeof(whole));
@@ -1233,7 +1233,7 @@ test_path_relative_cwd(hid_t fapl)
if (H5Dclose(dset) < 0)
FAIL_STACK_ERROR
- /* Open dataset (use a differend prefix than for create.
+ /* Open dataset (use a different prefix than for create.
* This works because the dataset was closed.
*/
if (H5Pset_efile_prefix(dapl2, "${ORIGIN}/.") < 0)
@@ -1254,7 +1254,7 @@ test_path_relative_cwd(hid_t fapl)
}
H5E_END_TRY;
if (dset3 >= 0)
- FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded");
+ FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded");
/* Read the entire dataset and compare with the original */
HDmemset(whole, 0, sizeof(whole));
diff --git a/test/farray.c b/test/farray.c
index fdf384b..44dda99 100644
--- a/test/farray.c
+++ b/test/farray.c
@@ -969,7 +969,7 @@ typedef struct fiter_fw_t {
/*-------------------------------------------------------------------------
* Function: fiter_fw_init
*
- * Purpose: Initialize element interator (forward iteration)
+ * Purpose: Initialize element iterator (forward iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1021,7 +1021,7 @@ fiter_fw_next(void *_fiter)
/*-------------------------------------------------------------------------
* Function: fiter_term
*
- * Purpose: Shut down element interator (simple iterators)
+ * Purpose: Shut down element iterator (simple iterators)
*
* Return: Success: 0
* Failure: -1
@@ -1055,7 +1055,7 @@ typedef struct fiter_rv_t {
/*-------------------------------------------------------------------------
* Function: fiter_rv_init
*
- * Purpose: Initialize element interator (reverse iteration)
+ * Purpose: Initialize element iterator (reverse iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1120,7 +1120,7 @@ typedef struct fiter_rnd_t {
/*-------------------------------------------------------------------------
* Function: fiter_rnd_init
*
- * Purpose: Initialize element interator (random iteration)
+ * Purpose: Initialize element iterator (random iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1193,7 +1193,7 @@ fiter_rnd_next(void *_fiter)
/*-------------------------------------------------------------------------
* Function: fiter_rnd_term
*
- * Purpose: Shut down element interator (random iteration)
+ * Purpose: Shut down element iterator (random iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1235,7 +1235,7 @@ typedef struct fiter_cyc_t {
/*-------------------------------------------------------------------------
* Function: fiter_cyc_init
*
- * Purpose: Initialize element interator (cyclic iteration)
+ * Purpose: Initialize element iterator (cyclic iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
diff --git a/test/fheap.c b/test/fheap.c
index c98034f..5ca45fc 100644
--- a/test/fheap.c
+++ b/test/fheap.c
@@ -2470,7 +2470,7 @@ test_id_limits(hid_t fapl, H5HF_create_t *cparam, hid_t fcpl)
size_t id_len; /* Size of fractal heap IDs */
size_t tiny_max_len; /* Max. length of tiny objects */
hbool_t tiny_len_extended; /* Do tiny objects use two bytes for the length? */
- hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */
+ hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */
/* Set the filename to use for this test (dependent on fapl) */
h5_fixname(FILENAME[0], fapl, filename, sizeof(filename));
@@ -13786,7 +13786,7 @@ test_filtered_huge(hid_t fapl, H5HF_create_t *cparam, fheap_test_param_t *tparam
fheap_heap_state_t state; /* State of fractal heap */
unsigned deflate_level; /* Deflation level */
size_t old_actual_id_len = 0; /* Old actual ID length */
- hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */
+ hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */
hbool_t pline_init = FALSE; /* Whether the I/O pipeline has been initialized */
/* Test description */
const char *base_desc = "insert 'huge' object into heap with I/O filters, then remove %s";
@@ -16362,7 +16362,7 @@ main(void)
unsigned num_pb_fs = 1; /* The number of settings to test for page buffering and file space handling */
int ExpressMode; /* Express testing level */
const char *envval; /* Environment variable */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
hbool_t api_ctx_pushed = FALSE; /* Whether API context pushed */
/* Don't run this test using certain file drivers */
@@ -16370,7 +16370,7 @@ main(void)
if (envval == NULL)
envval = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0);
/* Reset library */
diff --git a/test/file_image.c b/test/file_image.c
index 0373468..6950bfb 100644
--- a/test/file_image.c
+++ b/test/file_image.c
@@ -145,7 +145,7 @@ test_properties(void)
/* Check that sizes are the same, and that the buffers are identical but separate */
VERIFY(temp_size == size, "Sizes of buffers don't match");
- VERIFY(temp2 != NULL, "Recieved buffer not set");
+ VERIFY(temp2 != NULL, "Received buffer not set");
VERIFY(temp2 != buffer, "Retrieved buffer is the same as original");
VERIFY(temp2 != temp, "Retrieved buffer is the same as previously retrieved buffer");
VERIFY(0 == HDmemcmp(temp2, buffer, size), "Buffers contain different data");
@@ -260,7 +260,7 @@ free_cb(void *ptr, H5FD_file_image_op_t op, void *udata)
* Function: udata_copy_cb
*
* Purpose: This function allows calls to the udata_copy callback to be tracked.
- * No copying actualy takes place; it is easier to deal with only one
+ * No copying actually takes place; it is easier to deal with only one
* instance of the udata.
*
* Returns: A pointer to the same udata that was passed in.
@@ -954,7 +954,7 @@ test_get_file_image(const char *test_banner, const int file_name_num, hid_t fapl
err = H5Fclose(core_file_id);
VERIFY(err == SUCCEED, "H5Fclose(core_file_id) failed.");
- /* dicard core fapl */
+ /* discard core fapl */
err = H5Pclose(core_fapl_id);
VERIFY(err == SUCCEED, "H5Pclose(core_fapl_id) failed.");
@@ -1078,7 +1078,7 @@ test_get_file_image_error_rejection(void)
VERIFY(bytes_read < 0, "H5Fget_file_image(2 -- test 1) succeeded.");
/* Call H5Fget_file_image() with good buffer and buffer size,
- * but non-existant file_id. Should fail.
+ * but non-existent file_id. Should fail.
*/
H5E_BEGIN_TRY
{
diff --git a/test/fillval.c b/test/fillval.c
index 8de6ef1..b2e88bc 100644
--- a/test/fillval.c
+++ b/test/fillval.c
@@ -1918,7 +1918,7 @@ test_extend_cases(hid_t file, hid_t _dcpl, const char *dset_name, const hsize_t
if ((fspace = H5Dget_space(dset)) < 0)
TEST_ERROR
- /* Create dataspace for single element sized bufer */
+ /* Create dataspace for single element sized buffer */
if ((mspace = H5Screate_simple(5, one, NULL)) < 0)
TEST_ERROR
@@ -2142,7 +2142,7 @@ test_extend(hid_t fapl, const char *base_name, H5D_layout_t layout)
#if 1
/*
- * Remove this when contiguous datasets can be exended to some
+ * Remove this when contiguous datasets can be extended to some
* predetermined fininte size, even if it's just in the slowest varying
* dimension. If it's removed prematurely then you'll get one of the
* errors described above or `unable to select fill value region'.
diff --git a/test/filter_fail.c b/test/filter_fail.c
index a1bcf28..64cc41a 100644
--- a/test/filter_fail.c
+++ b/test/filter_fail.c
@@ -160,7 +160,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled)
points[i] = i;
/* Write data. If the chunk cache is enabled, H5Dwrite should succeed. If it is
- * diabled, H5Dwrite should fail. */
+ * disabled, H5Dwrite should fail. */
if (cache_enabled) {
if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, sid, H5P_DEFAULT, points) < 0)
TEST_ERROR
@@ -186,7 +186,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled)
TEST_ERROR
/* Close dataset. If the chunk cache is enabled, the flushing of chunks should fail
- * during H5Dclose. If it is diabled, H5Dwrite should fail but H5Dclose should succeed. */
+ * during H5Dclose. If it is disabled, H5Dwrite should fail but H5Dclose should succeed. */
if (cache_enabled) {
H5E_BEGIN_TRY
{
diff --git a/test/filter_plugin1_dsets.c b/test/filter_plugin1_dsets.c
index 5d6c1ef..29b6e73 100644
--- a/test/filter_plugin1_dsets.c
+++ b/test/filter_plugin1_dsets.c
@@ -79,7 +79,7 @@ add_sub_value(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_value
value = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) {
- /* READ - Substract the given value from all the data values */
+ /* READ - Subtract the given value from all the data values */
while (buf_left > 0) {
*int_ptr++ -= value;
buf_left -= sizeof(int);
diff --git a/test/filter_plugin3_dsets.c b/test/filter_plugin3_dsets.c
index 618ce06..1d399e5 100644
--- a/test/filter_plugin3_dsets.c
+++ b/test/filter_plugin3_dsets.c
@@ -108,7 +108,7 @@ add_sub_value_hdf5(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
value = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) {
- /* READ - Substract the given value from all the data values */
+ /* READ - Subtract the given value from all the data values */
while (buf_left > 0) {
*int_ptr++ -= value;
buf_left -= sizeof(int);
diff --git a/test/flushrefresh.c b/test/flushrefresh.c
index 6bf6260..0010cfd 100644
--- a/test/flushrefresh.c
+++ b/test/flushrefresh.c
@@ -63,10 +63,10 @@
#define NOT_FLUSHED "NOT_FLUSHED"
/* Error Handling */
-/* For errors occuring in the main process, use the standard TEST_ERROR macro.
+/* For errors occurring in the main process, use the standard TEST_ERROR macro.
For errors occurring in the spawned process (from the test script), use
the PROCESS_ERROR macro, which will send a signal to the main process so the
- main process can propogate errors correctly. */
+ main process can propagate errors correctly. */
FILE *errorfile;
#define ERRFILE "flushrefresh_ERROR"
#define PROCESS_ERROR \
@@ -222,7 +222,7 @@ test_flush(void)
* object, a series of verifications will occur on each object in the file.
*
* Each verification consists of spawning off a new process and determining
- * if the object can be opened and its information retreived in said
+ * if the object can be opened and its information retrieved in said
* alternate process. It reports the results, which are compared to an
* expected value (either that the object can be found on disk, or that it
* cannot).
@@ -695,7 +695,7 @@ test_refresh(void)
* that an attribute has been added, and is only visible after a
* successful call to a H5*refresh function).
*
- * As with the flush case, the implemention is a bit tricky as it's
+ * As with the flush case, the implementation is a bit tricky as it's
* dealing with signals going back and forth between the two processes
* to ensure the timing is correct, but basically, an example:
*
@@ -1166,7 +1166,7 @@ error:
* Purpose: This function is used to communicate with the verification
* process spawned by the start_refresh_verification_process
* function. It gives it the go-ahead to call H5*refresh
- * on an object and conlcude the refresh verification.
+ * on an object and conclude the refresh verification.
*
* Return: 0 on Success, 1 on Failure
*
@@ -1371,7 +1371,7 @@ error:
*
* Purpose: This function checks the status of external verification
* processes to see if they've succeeded. It checks for the
- * existance of flushrefresh_ERROR file. If present, that indicates
+ * existence of flushrefresh_ERROR file. If present, that indicates
* an external verification process has failed, and this function
* thus fails as well. If not present, then nothing else has
* failed, and this function succeeds.
diff --git a/test/gen_bounds.c b/test/gen_bounds.c
index 8148f16..87a9c18 100644
--- a/test/gen_bounds.c
+++ b/test/gen_bounds.c
@@ -858,12 +858,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_REG_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
@@ -886,12 +886,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_IRR_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
@@ -915,12 +915,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_PT_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT, H5P_DEFAULT,
H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
diff --git a/test/gen_old_group.c b/test/gen_old_group.c
index 1b4e0b2..15e67a4 100644
--- a/test/gen_old_group.c
+++ b/test/gen_old_group.c
@@ -19,7 +19,7 @@
* empty "symbol table" group.
*
* This file is used in the v1.7 branch (after the "compact group"
- * checkin) to test compatability. Compile and run this
+ * checkin) to test compatibility. Compile and run this
* program (with the 1.6.x branch), it will generate a file*
* called "group_old.h5". You need to move it to the test
* directory in the HDF5 v1.7 source tree. The test/stab.c
diff --git a/test/gen_plist.c b/test/gen_plist.c
index 10c6423..25d48cf 100644
--- a/test/gen_plist.c
+++ b/test/gen_plist.c
@@ -81,7 +81,7 @@ main(void)
H5AC_cache_image_config_t my_cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE,
-1};
- /* check endianess */
+ /* check endianness */
{
short int word = 0x0001;
char * byte = (char *)&word;
diff --git a/test/genall5.c b/test/genall5.c
index 4fa6c01..53d1cdd 100644
--- a/test/genall5.c
+++ b/test/genall5.c
@@ -14,7 +14,7 @@
/* Programmer: John Mainzer
* 9/23/15
*
- * This file contains a heavily edited and functionaly reduce
+ * This file contains a heavily edited and functionally reduce
* version of the test code first written by Quincey in a file
* of the same name.
*/
diff --git a/test/getname.c b/test/getname.c
index 200c426..698079d 100644
--- a/test/getname.c
+++ b/test/getname.c
@@ -1526,7 +1526,7 @@ test_main(hid_t file_id, hid_t fapl)
{
char name[NAME_BUF_SIZE]; /* Buffer to hold name and its size */
- /* Get name for non commited datatype, it should fail */
+ /* Get name for non committed datatype, it should fail */
H5E_BEGIN_TRY
{
if (H5Iget_name(type_id, name, NAME_BUF_SIZE) > 0)
diff --git a/test/gheap.c b/test/gheap.c
index fa761b4..c83ff68 100644
--- a/test/gheap.c
+++ b/test/gheap.c
@@ -353,7 +353,7 @@ error:
* Function: test_4
*
* Purpose: Tests the H5HG_remove() feature by writing lots of objects
- * and occassionally removing some. When we're done they're all
+ * and occasionally removing some. When we're done they're all
* removed.
*
* Return: Success: 0
diff --git a/test/h5test.c b/test/h5test.c
index bb2f5cc..0a7f2fd 100644
--- a/test/h5test.c
+++ b/test/h5test.c
@@ -1828,7 +1828,7 @@ h5_send_message(const char *send, const char *arg1, const char *arg2)
* signal file from disk, and only continues once it has
* successfully done so (i.e., only after another process has
* called the "h5_send_message" function to write the signal file).
- * This functon will then immediately remove the file (i.e.,
+ * This function will then immediately remove the file (i.e.,
* to indicate that it has been received and can be reused),
* and then exits, allowing the calling function to continue.
*
diff --git a/test/hdfs.c b/test/hdfs.c
index 5465e77..be67ee4 100644
--- a/test/hdfs.c
+++ b/test/hdfs.c
@@ -1454,7 +1454,7 @@ test_noops_and_autofails(void)
JSVERIFY(SUCCEED, H5FDlock(file, FALSE), "lock issue")
JSVERIFY(SUCCEED, H5FDunlock(file), "unlock issue")
/* Lock/unlock with null file or similar error crashes tests.
- * HDassert in calling heirarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
+ * HDassert in calling hierarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
*/
/************
diff --git a/test/links.c b/test/links.c
index 2de1fca..2aeef14 100644
--- a/test/links.c
+++ b/test/links.c
@@ -8199,7 +8199,7 @@ external_link_dangling(hid_t fapl, hbool_t new_format)
if (status >= 0) {
H5_FAILED();
HDputs(
- " Retreiving name of object by index through dangling file external link should have failed.");
+ " Retrieving name of object by index through dangling file external link should have failed.");
} /* end if */
/* Close root group */
@@ -8672,7 +8672,7 @@ external_link_abstar(hid_t fapl, hbool_t new_format)
}
H5E_END_TRY;
- /* should be able to find the target file with abolute path */
+ /* should be able to find the target file with absolute path */
if (gid < 0) {
H5_FAILED();
HDputs(" Should have found the file in tmp_links directory.");
@@ -16591,7 +16591,7 @@ link_filters(hid_t fapl, hbool_t new_format)
TEST_ERROR
filesize_unfiltered = h5_get_file_size(filename, fapl);
- /* Set deflate fitler */
+ /* Set deflate filter */
if (H5Pset_deflate(fcpl, 6) < 0)
TEST_ERROR
@@ -22556,7 +22556,7 @@ main(void)
for (minimize_dset_oh = 0; minimize_dset_oh <= 1; minimize_dset_oh++) {
if (minimize_dset_oh) {
- HDprintf("\n-Testing with minimzed dataset object headers-\n");
+ HDprintf("\n-Testing with minimized dataset object headers-\n");
dcpl_g = H5Pcreate(H5P_DATASET_CREATE);
if (0 > dcpl_g)
TEST_ERROR
diff --git a/test/mf.c b/test/mf.c
index 9eb7807..facfeb7 100644
--- a/test/mf.c
+++ b/test/mf.c
@@ -223,7 +223,7 @@ test_mf_eoa(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MM_alloc() of file allocation");
@@ -371,7 +371,7 @@ test_mf_eoa_shrink(const char *env_h5_drvr, hid_t fapl)
haddr_t addr = 0;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_shrink() of file allocation: test 1");
@@ -658,7 +658,7 @@ test_mf_eoa_extend(const char *env_h5_drvr, hid_t fapl)
htri_t was_extended;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_extend() of file allocation: test 1");
@@ -1934,7 +1934,7 @@ test_mf_fs_extend(hid_t fapl)
if (check_stats(f, f->shared->fs_man[H5FD_MEM_SUPER], &state))
TEST_ERROR
- /* Remove section A from free-space manger */
+ /* Remove section A from free-space manager */
if (H5MF__find_sect(f, H5FD_MEM_SUPER, (hsize_t)(TBLOCK_SIZE30 - 10), f->shared->fs_man[H5FD_MEM_SUPER],
&tmp) != TRUE)
TEST_ERROR
@@ -2012,7 +2012,7 @@ test_mf_fs_absorb(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
H5MF_free_section_t *sect_node = NULL;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("A free-space section absorbs an aggregator: test 1");
@@ -2188,7 +2188,7 @@ test_mf_aggr_alloc1(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2;
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 1");
@@ -2339,7 +2339,7 @@ test_mf_aggr_alloc2(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, addr3;
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 2");
@@ -2497,7 +2497,7 @@ test_mf_aggr_alloc3(const char *env_h5_drvr, hid_t fapl)
hsize_t ma_size = 0, new_ma_size = 0;
haddr_t sdata_addr = HADDR_UNDEF;
hsize_t sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator: test 3");
@@ -2665,7 +2665,7 @@ test_mf_aggr_alloc4(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, saddr1, saddr2, saddr3;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 4");
@@ -2814,7 +2814,7 @@ test_mf_aggr_alloc5(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, addr3;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 5");
@@ -2952,7 +2952,7 @@ test_mf_aggr_alloc6(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
H5FS_stat_t state;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 6");
@@ -3126,7 +3126,7 @@ test_mf_aggr_alloc7(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, sdata_size = 0;
H5FS_stat_t state;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 7");
@@ -3297,7 +3297,7 @@ test_mf_aggr_extend(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
htri_t was_extended;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_extend() of meta/sdata aggregator: test 1");
@@ -3595,7 +3595,7 @@ test_mf_aggr_absorb(const char *env_h5_drvr, hid_t fapl)
haddr_t new_sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
hsize_t sdata_size = 0, new_sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_shrink() of meta/sdata aggregator: test 1");
@@ -5838,7 +5838,7 @@ error:
*
* Alignment = 4096 aggr->alloc_size = 2048
*
- * Allocate first block (30) from meta_aggr: (meta_aggr is emtpy)
+ * Allocate first block (30) from meta_aggr: (meta_aggr is empty)
* request-size is > meta_aggr->size and < meta_aggr->alloc_size
* Result:
* A block of meta_aggr->alloc_size is allocated from file allocation
@@ -7105,7 +7105,7 @@ test_mf_fs_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format)
else
TESTING("File's free-space is going away with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
if (contig_addr_vfd) {
@@ -7296,7 +7296,7 @@ test_mf_strat_thres_persist(const char *env_h5_drvr, hid_t fapl, hbool_t new_for
else
TESTING("File space strategy/persisting/threshold with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Set the filename to use for this test (dependent on fapl) */
@@ -7466,7 +7466,7 @@ test_mf_strat_thres_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format
else
TESTING("File space merge/shrink for section size < threshold with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Set the filename to use for this test (dependent on fapl) */
@@ -8024,7 +8024,7 @@ test_page_try_shrink(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1; /* Address for small metadata block */
haddr_t saddr1; /* Address for small raw data block */
haddr_t gaddr1; /* Address for large data block */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t status; /* status from shrinking */
h5_stat_size_t file_size; /* File size */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8154,7 +8154,7 @@ test_page_small_try_extend(const char *env_h5_drvr, hid_t fapl)
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t addr1, addr2, addr3; /* Addresses for small metadata blocks */
haddr_t saddr1; /* Address for small raw data block */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t was_extended; /* Whether the block can be extended or not */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8333,7 +8333,7 @@ test_page_large_try_extend(const char *env_h5_drvr, hid_t fapl)
hid_t fcpl = -1; /* File creation property list */
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for large data blocks */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t was_extended; /* Whether the block can be extended or not */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8496,7 +8496,7 @@ test_page_large(const char *env_h5_drvr, hid_t fapl)
hid_t fcpl = -1; /* File creation property list ID */
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for blocks */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
H5FS_stat_t fs_stat; /* Information for free-space manager */
h5_stat_size_t file_size; /* File size */
char filename[FILENAME_LEN]; /* Filename to use */
diff --git a/test/mirror_vfd.c b/test/mirror_vfd.c
index 3556cd2..7047146 100644
--- a/test/mirror_vfd.c
+++ b/test/mirror_vfd.c
@@ -128,7 +128,7 @@ mybzero(void *dest, size_t size)
* Function: _populate_filepath
*
* Purpose: Given a directory name and a base name, concatenate the two and
- * run h5fixname() to get the "actual" path to the intented target.
+ * run h5fixname() to get the "actual" path to the intended target.
* `h5suffix' should be FALSE to keep the base name unaltered;
* TRUE will append the '.h5' h5suffix to the basename...
* FALSE -> h5fixname_no_suffix(), TRUE -> h5fixname()
diff --git a/test/mount.c b/test/mount.c
index c5ac9b3..b23de30 100644
--- a/test/mount.c
+++ b/test/mount.c
@@ -753,7 +753,7 @@ test_move(hid_t fapl)
H5E_END_TRY;
if (status >= 0) {
H5_FAILED();
- HDputs(" Moving an object across files should't have been possible");
+ HDputs(" Moving an object across files shouldn't have been possible");
TEST_ERROR
} /* end if */
@@ -1835,7 +1835,7 @@ test_mount_after_unmount(hid_t fapl)
if (HDstrcmp(objname, "/X/M/Y") != 0)
TEST_ERROR
- /* Rename object in file #3 that is "disconnected" from name hiearchy */
+ /* Rename object in file #3 that is "disconnected" from name hierarchy */
/* (It is "disconnected" because it's parent file has been unmounted) */
if (H5Lmove(gidAMX, "M/Y", gidAMX, "M/Z", H5P_DEFAULT, H5P_DEFAULT) < 0)
TEST_ERROR
diff --git a/test/mtime.c b/test/mtime.c
index 43dc502..fdc499a 100644
--- a/test/mtime.c
+++ b/test/mtime.c
@@ -116,7 +116,7 @@ main(void)
if (0 == oi1.ctime) {
SKIPPED();
HDputs(" The modification time could not be decoded on this OS.");
- HDputs(" Modification times will be mantained in the file but");
+ HDputs(" Modification times will be maintained in the file but");
HDputs(" cannot be queried on this system. See H5O_mtime_decode().");
return 0;
}
diff --git a/test/objcopy.c b/test/objcopy.c
index 9aae621..c70bbea 100644
--- a/test/objcopy.c
+++ b/test/objcopy.c
@@ -2789,7 +2789,7 @@ test_copy_dataset_simple_empty(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): empty and openend contiguous dataset");
+ TESTING("H5Ocopy(): empty and opened contiguous dataset");
}
else {
TESTING("H5Ocopy(): empty contiguous dataset");
@@ -8679,7 +8679,7 @@ test_copy_soft_link(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t dst_fa
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): openend object through soft link");
+ TESTING("H5Ocopy(): opened object through soft link");
}
else {
TESTING("H5Ocopy(): object through soft link");
@@ -9674,7 +9674,7 @@ test_copy_dataset_contig_named_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl
char src_filename[NAME_BUF_SIZE];
char dst_filename[NAME_BUF_SIZE];
- TESTING("H5Ocopy(): contigous dataset with named VLEN datatype");
+ TESTING("H5Ocopy(): contiguous dataset with named VLEN datatype");
/* set initial data values */
for (i = 0; i < DIM_SIZE_1; i++) {
@@ -10434,10 +10434,10 @@ test_copy_dataset_contig_vl_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): contigous and opened dataset with nested VLEN datatype");
+ TESTING("H5Ocopy(): contiguous and opened dataset with nested VLEN datatype");
}
else {
- TESTING("H5Ocopy(): contigous dataset with nested VLEN datatype");
+ TESTING("H5Ocopy(): contiguous dataset with nested VLEN datatype");
}
/* set initial data values */
@@ -13763,7 +13763,7 @@ test_copy_cdt_hier_merge(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t d
TEST_ERROR
/*
- * Test 3: copy datsets in /g0 at SRC to DST group /uncopied
+ * Test 3: copy datasets in /g0 at SRC to DST group /uncopied
*/
if (H5Ocopy(fid_src, NAME_GROUP_TOP "/" SRC_DSET1, fid_dst, NAME_GROUP_UNCOPIED "/" SRC_DSET1, ocpypl_id,
H5P_DEFAULT) < 0)
diff --git a/test/ohdr.c b/test/ohdr.c
index e0b5e88..ecea18e 100644
--- a/test/ohdr.c
+++ b/test/ohdr.c
@@ -584,7 +584,7 @@ test_unknown(unsigned bogus_id, char *filename, hid_t fapl)
TESTING("object with unknown header message & 'shareable' flag set");
- /* Open the dataset with the unknown header message, adn "shareable" flag */
+ /* Open the dataset with the unknown header message, and "shareable" flag */
if ((did = H5Dopen2(loc_bogus, "Dataset5", H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR
if (H5Dclose(did) < 0)
@@ -1089,13 +1089,13 @@ test_minimized_dset_ohdr_size_comparisons(hid_t fapl_id)
/* IDs for non-minimized file open */
hid_t file_f_id = -1; /* lower 'f' for standard file setting */
hid_t dset_f_x_id = -1; /* 'x' for default */
- hid_t dset_f_N_id = -1; /* 'N' for explcit non-minimized dset */
+ hid_t dset_f_N_id = -1; /* 'N' for explicit non-minimized dset */
hid_t dset_f_Y_id = -1; /* 'Y' for minimized dset */
/* IDs for minimized file open */
hid_t file_F_id = -1; /* upper 'F' for minimized file setting */
hid_t dset_F_x_id = -1; /* 'x' for default */
- hid_t dset_F_N_id = -1; /* 'N' for explcit non-minimized dset */
+ hid_t dset_F_N_id = -1; /* 'N' for explicit non-minimized dset */
hid_t dset_F_Y_id = -1; /* 'Y' for minimized dset */
char filename_a[512] = "";
diff --git a/test/page_buffer.c b/test/page_buffer.c
index 37f2acc..8918639 100644
--- a/test/page_buffer.c
+++ b/test/page_buffer.c
@@ -75,7 +75,7 @@ const char *FILENAME[] = {"filepaged", NULL};
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -215,7 +215,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -375,7 +375,7 @@ error:
* Any data mis-matches or unexpected failures or successes
* reported by the HDF5 library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -549,7 +549,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -827,7 +827,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -1081,7 +1081,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -1502,7 +1502,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 2)
TEST_ERROR;
- /* adding more meta entires should replace meta entries since raw data
+ /* adding more meta entries should replace meta entries since raw data
* is at its minimum
*/
if (H5F_block_write(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 600), sizeof(int) * 100, data) < 0)
@@ -1517,7 +1517,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 2)
TEST_ERROR;
- /* bring existing raw entires up the LRU */
+ /* bring existing raw entries up the LRU */
if (H5F_block_read(f, H5FD_MEM_DRAW, raw_addr + (sizeof(int) * 750), sizeof(int) * 100, data) < 0)
FAIL_STACK_ERROR;
@@ -1533,7 +1533,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 3)
TEST_ERROR;
- /* adding 2 meta entries should replace 2 entires at the bottom of the LRU */
+ /* adding 2 meta entries should replace 2 entries at the bottom of the LRU */
if (H5F_block_read(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 98), sizeof(int) * 100, data) < 0)
FAIL_STACK_ERROR;
@@ -1716,7 +1716,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -2019,7 +2019,7 @@ error:
* At present, page buffering should be disabled in parallel
* builds. Verify this.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: John Mainzer
@@ -2134,7 +2134,7 @@ error:
*
* Purpose: Main function for the page buffer tests.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
diff --git a/test/ros3.c b/test/ros3.c
index 2d4bcf0..1f56d70 100644
--- a/test/ros3.c
+++ b/test/ros3.c
@@ -1456,7 +1456,7 @@ test_noops_and_autofails(void)
JSVERIFY(SUCCEED, H5FDlock(file, FALSE), NULL)
JSVERIFY(SUCCEED, H5FDunlock(file), NULL)
/* Lock/unlock with null file or similar error crashes tests.
- * HDassert in calling heirarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
+ * HDassert in calling hierarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
*/
/************
@@ -1772,7 +1772,7 @@ main(void)
bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL");
if (bucket_url_env == NULL || bucket_url_env[0] == '\0') {
- HDprintf("WARNING: S3 bucket url is not defined in enviornment "
+ HDprintf("WARNING: S3 bucket url is not defined in environment "
"variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n");
}
else {
diff --git a/test/s3comms.c b/test/s3comms.c
index 22d647f..03b8ce6 100644
--- a/test/s3comms.c
+++ b/test/s3comms.c
@@ -244,7 +244,7 @@ jserr_str(const char *expected, const char *actual, const char *reason)
}
#ifdef JSVERIFY_EXP_ACT
-/* VERIFY rountines with paramter order (<expected>, <actual> [, <msg> ])
+/* VERIFY routines with parameter order (<expected>, <actual> [, <msg> ])
*/
/*----------------------------------------------------------------------------
@@ -435,7 +435,7 @@ error:
*
* Purpose:
*
- * Demonstrate the construction of a Canoncial Request (and Signed Headers)
+ * Demonstrate the construction of a Canonical Request (and Signed Headers)
*
* Elided / not yet implemented:
* Query strings
@@ -730,7 +730,7 @@ test_hrb_init_request(void)
FALSE,
},
{
- "null verb substitues to GET",
+ "null verb substitutes to GET",
NULL,
"/MYPATH/MYFILE.tiff",
"/MYPATH/MYFILE.tiff",
@@ -1144,9 +1144,9 @@ test_hrb_node_set(void)
for (mock_i = 0; test->given[mock_i] != NULL; mock_i += 2) {
const char *name = test->given[mock_i];
- const char *valu = test->given[mock_i + 1];
+ const char *value = test->given[mock_i + 1];
- FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, valu))
+ FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, value))
}
/********
* TEST *
@@ -1163,10 +1163,10 @@ test_hrb_node_set(void)
mock_i = 0;
while (test->expected[mock_i] != NULL && node != NULL) {
const char *name = test->expected[mock_i];
- const char *valu = test->expected[mock_i + 1];
+ const char *value = test->expected[mock_i + 1];
JSVERIFY_STR(name, node->name, NULL)
- JSVERIFY_STR(valu, node->value, NULL)
+ JSVERIFY_STR(value, node->value, NULL)
mock_i += 2;
node = node->next;
@@ -1559,7 +1559,7 @@ test_parse_url(void)
NULL,
"a=b&d=b",
},
- "QUERY with implict PATH",
+ "QUERY with implicit PATH",
},
{
"http://[5]/path?a=b&d=b",
@@ -2014,7 +2014,7 @@ error:
* H5FD_s3comms_s3r_read << called by getsize(), multiple times working
* H5FD_s3comms_s3r_close
*
- * Shows most basic curl interation.
+ * Shows most basic curl iteration.
*
* Programmer: Jacob Smith
* 2017-10-06
@@ -2610,7 +2610,7 @@ main(void)
bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL");
if (bucket_url_env == NULL || bucket_url_env[0] == '\0') {
- HDprintf("WARNING: S3 bucket url is not defined in enviornment "
+ HDprintf("WARNING: S3 bucket url is not defined in environment "
"variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n");
}
else {
@@ -2618,7 +2618,7 @@ main(void)
s3_test_bucket_defined = TRUE;
}
- /* tests ordered rougly by dependence */
+ /* tests ordered roughly by dependence */
nerrors += test_macro_format_credential() < 0 ? 1 : 0;
nerrors += test_trim() < 0 ? 1 : 0;
nerrors += test_nlowercase() < 0 ? 1 : 0;
diff --git a/test/set_extent.c b/test/set_extent.c
index 3078c2d..44bf8cb 100644
--- a/test/set_extent.c
+++ b/test/set_extent.c
@@ -115,12 +115,12 @@ main(void)
unsigned chunk_cache; /* Whether to enable chunk caching */
int nerrors = 0;
const char *env_h5_drvr; /* File Driver value from environment */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
env_h5_drvr = HDgetenv("HDF5_DRIVER");
if (env_h5_drvr == NULL)
env_h5_drvr = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Initialize random number seed */
@@ -649,7 +649,7 @@ test_rank1(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
HDprintf("\n");
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++)
if (buf_s[i] != buf_o[i]) {
HDprintf("buf_s[%d] = %d\n", i, buf_s[i]);
@@ -1055,7 +1055,7 @@ test_rank2(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
}
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
if (buf_s[i][j] != buf_o[i][j]) {
@@ -1536,7 +1536,7 @@ test_rank3(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
HDprintf("\n");
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
for (k = 0; k < (int)dims_r[2]; k++) {
@@ -1734,7 +1734,7 @@ test_external(hid_t fapl)
hsize_t dims_r[RANK2]; /* read dimensions */
hsize_t maxdims[RANK2] = {DIME0, DIM1}; /* only the first dimension can be extendible */
int buf_o[DIM0][DIM1]; /* original buffer, for writing */
- int buf_s[DIMS0][DIMS1]; /* shrinked buffer, for reading */
+ int buf_s[DIMS0][DIMS1]; /* shrunk buffer, for reading */
int buf_e[DIME0][DIM1]; /* extended buffer, for writing, dimension 1 is the original */
int buf_ro[DIM0][DIM1]; /* original buffer for reading */
int i, j;
@@ -1946,7 +1946,7 @@ test_external(hid_t fapl)
}
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
if (buf_s[i][j] != buf_o[i][j]) {
diff --git a/test/stab.c b/test/stab.c
index 798619d..b11abd5 100644
--- a/test/stab.c
+++ b/test/stab.c
@@ -125,7 +125,7 @@ test_misc(hid_t fcpl, hid_t fapl, hbool_t new_format)
if (H5Gclose(g3) < 0)
TEST_ERROR
- /* Open all groups with absolute names to check for exsistence */
+ /* Open all groups with absolute names to check for existence */
if ((g1 = H5Gopen2(fid, "/test_1a", H5P_DEFAULT)) < 0)
TEST_ERROR
if ((g2 = H5Gopen2(fid, "/test_1a/sub_1", H5P_DEFAULT)) < 0)
@@ -1416,7 +1416,7 @@ main(void)
hid_t fcpl, fcpl2; /* File creation property list ID */
unsigned new_format; /* Whether to use the new format or not */
const char *env_h5_drvr; /* File Driver value from environment */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
int nerrors = 0;
/* Get the VFD to use */
@@ -1424,7 +1424,7 @@ main(void)
if (env_h5_drvr == NULL)
env_h5_drvr = "nomatch";
- /* VFD that does not support contigous address space */
+ /* VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Reset library */
diff --git a/test/swmr.c b/test/swmr.c
index b609fb6..5294267 100644
--- a/test/swmr.c
+++ b/test/swmr.c
@@ -3666,7 +3666,7 @@ test_append_flush_dataset_chunked(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dcpl = -1; /* A copy of dataset creation property */
hid_t dapl = -1; /* A copy of dataset access property */
@@ -3889,7 +3889,7 @@ test_append_flush_dataset_fixed(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dapl = -1; /* A copy of dataset access property */
hid_t ddapl = -1; /* The dataset access property of the opened dataset */
@@ -4107,7 +4107,7 @@ test_append_flush_dataset_multiple(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dcpl = -1; /* A copy of dataset creation property */
hid_t dapl1 = -1; /* A copy of dataset access property */
@@ -7412,7 +7412,7 @@ main(void)
/*
* Modify the following routines to test for files:
* H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format)
- * --both result in v3 superblock and latest version suppport
+ * --both result in v3 superblock and latest version support
*/
nerrors += test_start_swmr_write(fapl, TRUE);
nerrors += test_start_swmr_write(fapl, FALSE);
@@ -7446,7 +7446,7 @@ main(void)
*
* Modify the following 2 routines to test for files:
* H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format)
- * --both result in v3 superblock and latest version suppport
+ * --both result in v3 superblock and latest version support
*/
nerrors += test_file_lock_swmr_same(fapl);
nerrors += test_file_lock_swmr_concur(fapl);
diff --git a/test/swmr_generator.c b/test/swmr_generator.c
index a455f3f..93cfa0f 100644
--- a/test/swmr_generator.c
+++ b/test/swmr_generator.c
@@ -221,7 +221,7 @@ gen_skeleton(const char *filename, hbool_t verbose, hbool_t swmr_write, int comp
if (verbose)
HDfprintf(stderr, "Closing objects\n");
- /* Close everythign */
+ /* Close everything */
if (H5Pclose(dcpl) < 0)
return -1;
if (H5Sclose(sid) < 0)
diff --git a/test/swmr_start_write.c b/test/swmr_start_write.c
index 3e84dcc..d046b40 100644
--- a/test/swmr_start_write.c
+++ b/test/swmr_start_write.c
@@ -125,7 +125,7 @@ create_file(const char *filename, hbool_t verbose, FILE *verbose_file, unsigned
*
* Parameters:
* fid: file ID for the SWMR test file
- * comp_level: the compresssion level
+ * comp_level: the compression level
* index_type: The chunk index type (b1 | b2 | ea | fa)
* verbose: whether verbose console output is desired.
* verbose_file: file pointer for verbose output
diff --git a/test/tattr.c b/test/tattr.c
index e75eeb5..974367c 100644
--- a/test/tattr.c
+++ b/test/tattr.c
@@ -5858,7 +5858,7 @@ test_attr_info_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for query on non-existant attribute */
+ /* Check for query on non-existent attribute */
ret = H5Aget_info_by_idx(my_dataset, ".", H5_INDEX_CRT_ORDER, H5_ITER_INC, (hsize_t)0, &ainfo,
H5P_DEFAULT);
VERIFY(ret, FAIL, "H5Aget_info_by_idx");
@@ -6346,7 +6346,7 @@ test_attr_delete_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for deleting non-existant attribute */
+ /* Check for deleting non-existent attribute */
ret = H5Adelete_by_idx(my_dataset, ".", idx_type, order, (hsize_t)0, H5P_DEFAULT);
VERIFY(ret, FAIL, "H5Adelete_by_idx");
@@ -8001,7 +8001,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for opening a non-existant attribute on an object with no attributes */
+ /* Check for opening a non-existent attribute on an object with no attributes */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -8040,7 +8040,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for opening a non-existant attribute on an object with compact attribute storage */
+ /* Check for opening a non-existent attribute on an object with compact attribute storage */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -8121,7 +8121,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
VERIFY(name_count, (max_compact * 2), "H5O__attr_dense_info_test");
} /* end if */
- /* Check for opening a non-existant attribute on an object with dense attribute storage */
+ /* Check for opening a non-existent attribute on an object with dense attribute storage */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -10989,7 +10989,7 @@ test_attr_bug8(hid_t fcpl, hid_t fapl)
** chunk. Do this twice: once with only attributes in the
** object header chunk and once with a (small) soft link in
** the chunk as well. In both cases, the shrunk chunk will
-** initally be too small and a new NULL message must be
+** initially be too small and a new NULL message must be
** created.
**
****************************************************************/
diff --git a/test/tcoords.c b/test/tcoords.c
index fef8689..b2d68d3 100644
--- a/test/tcoords.c
+++ b/test/tcoords.c
@@ -17,7 +17,7 @@
*
* Test the element coordinates for dataspace selection. For
* chunked dataset, when the hyperslab selection of some
- * dimensions is full, the library optimize it by "flattenning"
+ * dimensions is full, the library optimize it by "flattening"
* the fully selected dimensions. This program tests if the
* coordinates of selected elements are correctly calculated.
*
diff --git a/test/testmeta.c b/test/testmeta.c
index 7cb3000..a3b5160 100644
--- a/test/testmeta.c
+++ b/test/testmeta.c
@@ -15,7 +15,7 @@
* This program illustrates assertion errors when linked
* to HDF5 1.4.1-post2 or 1.4.2-pre3 debug library.
*
- * If the assertion errors are ignored, the program eventially causes
+ * If the assertion errors are ignored, the program eventually causes
* an error in H5Gcreate2 when writing object 83381.
*
* When writing in single file mode, the assertion errors still occur
diff --git a/test/tfile.c b/test/tfile.c
index 8c3f9b0..d49a64c 100644
--- a/test/tfile.c
+++ b/test/tfile.c
@@ -225,7 +225,7 @@ test_file_create(void)
/* First ensure the file does not exist */
HDremove(FILE1);
- /* Try opening a non-existant file */
+ /* Try opening a non-existent file */
fid1 = H5Fopen(FILE1, H5F_ACC_RDWR, H5P_DEFAULT);
VERIFY(fid1, FAIL, "H5Fopen");
@@ -1774,7 +1774,7 @@ test_file_is_accessible(const char *env_h5_drvr)
**
** test_file_ishdf5(): low-level file test routine.
** This test checks whether the H5Fis_hdf5() routine is working
-** correctly in variuous situations.
+** correctly in various situations.
**
*****************************************************************/
#ifndef H5_NO_DEPRECATED_SYMBOLS
@@ -1998,7 +1998,7 @@ test_file_delete(hid_t fapl_id)
**
** test_file_open_dot(): low-level file test routine.
** This test checks whether opening objects with "." for a name
-** works correctly in variuous situations.
+** works correctly in various situations.
**
*****************************************************************/
static void
@@ -2380,7 +2380,7 @@ test_file_double_root_open(void)
grp2_id = H5Gopen2(file2_id, "/", H5P_DEFAULT);
CHECK(grp2_id, FAIL, "H5Gopen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Gclose(grp1_id);
CHECK(ret, FAIL, "H5Gclose");
ret = H5Gclose(grp2_id);
@@ -2419,7 +2419,7 @@ test_file_double_group_open(void)
grp2_id = H5Gopen2(file2_id, GRP_NAME, H5P_DEFAULT);
CHECK(grp2_id, FAIL, "H5Gopen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Gclose(grp1_id);
CHECK(ret, FAIL, "H5Gclose");
ret = H5Gclose(grp2_id);
@@ -2468,7 +2468,7 @@ test_file_double_dataset_open(void)
ret = H5Sclose(space_id);
CHECK(ret, FAIL, "H5Sclose");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Dclose(dset1_id);
CHECK(ret, FAIL, "H5Dclose");
ret = H5Dclose(dset2_id);
@@ -2869,7 +2869,7 @@ test_file_double_datatype_open(void)
type2_id = H5Topen2(file2_id, TYPE_NAME, H5P_DEFAULT);
CHECK(type2_id, FAIL, "H5Topen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Tclose(type1_id);
CHECK(ret, FAIL, "H5Tclose");
ret = H5Tclose(type2_id);
@@ -2914,7 +2914,7 @@ test_userblock_file_size(void)
ret = H5Pset_userblock(fcpl2_id, USERBLOCK_SIZE);
CHECK(ret, FAIL, "H5Pset_userblock");
- /* Create files. Onyl file2 with have a userblock. */
+ /* Create files. Only file2 with have a userblock. */
file1_id = H5Fcreate(FILE1, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
CHECK(file1_id, FAIL, "H5Fcreate");
file2_id = H5Fcreate(FILE2, H5F_ACC_TRUNC, fcpl2_id, H5P_DEFAULT);
@@ -3835,7 +3835,7 @@ test_userblock_alignment_paged(void)
* Strategy is H5F_FILE_SPACE_NONE; fsp_size = 1024
* H5Pset_alignment() is 16
* Outcome:
- * Should succed:
+ * Should succeed:
* userblock (512) is integral multiple of alignment (16)
*/
/* Create file creation property list with user block */
@@ -3874,7 +3874,7 @@ test_userblock_alignment_paged(void)
* H5Pset_alignment() is 3
* Reopen the file; H5Pset_alignment() is 1024
* Outcome:
- * Should succed:
+ * Should succeed:
* Userblock (512) is the same as alignment (512);
* The H5Pset_alignment() calls have no effect
*/
@@ -3943,7 +3943,7 @@ test_filespace_info(const char *env_h5_drvr)
hsize_t fs_threshold; /* Free-space section threshold--iteration variable */
hsize_t fsp_size; /* File space page size */
char filename[FILENAME_LEN]; /* Filename to use */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
herr_t ret; /* Return value */
/* Output message about test being performed */
@@ -5185,7 +5185,7 @@ test_filespace_1_10_0_compatible(void)
** Verify that the trunk can open, read and modify these files--
** 1) They are initially created (via gen_filespace.c) in the trunk
** with combinations of file space strategies, default/non-default
-** threshold, and file spacing paging enabled/disbled.
+** threshold, and file spacing paging enabled/disabled.
** The library creates the file space info message with
** "mark if unknown" in these files.
** 2) They are copied to the 1.8 branch, and are opened/read/modified
@@ -6345,7 +6345,7 @@ test_libver_bounds_dataset(hid_t fapl)
VERIFY(dset->shared->dcpl_cache.fill.version, H5O_fill_ver_bounds[low], "H5O_fill_ver_bounds");
}
- /* Verify filter pipleline message version */
+ /* Verify filter pipeline message version */
VERIFY(dset->shared->dcpl_cache.pline.version, H5O_pline_ver_bounds[low], "H5O_pline_ver_bounds");
/* Close the dataset */
@@ -6398,7 +6398,7 @@ test_libver_bounds_dataset(hid_t fapl)
ret = H5Sclose(sid);
CHECK(ret, FAIL, "H5Sclose");
- /* Close the datset creation property list */
+ /* Close the dataset creation property list */
ret = H5Pclose(dcpl);
CHECK(ret, FAIL, "H5Pclose");
@@ -6418,7 +6418,7 @@ test_libver_bounds_dataset(hid_t fapl)
/* Loop through all the combinations of low/high bounds in new_fapl */
/* Open the file with the fapl and create the chunked dataset */
- /* Verify the dataset's layout, fill value and filter pipleline message versions */
+ /* Verify the dataset's layout, fill value and filter pipeline message versions */
for (low = H5F_LIBVER_EARLIEST; low < H5F_LIBVER_NBOUNDS; low++) {
for (high = H5F_LIBVER_EARLIEST; high < H5F_LIBVER_NBOUNDS; high++) {
H5E_BEGIN_TRY
@@ -6523,7 +6523,7 @@ test_libver_bounds_dataset(hid_t fapl)
** new fapl:
** --Open the same file in (a) with the fapl
** --Create a chunked dataset, a compact dataset and
-** a contigous dataset
+** a contiguous dataset
** --Verify the dataspace message version for these
** three datasets
** --Delete the three datasets and the dataspaces
@@ -6872,7 +6872,7 @@ static void
test_libver_bounds_datatype_check(hid_t fapl, hid_t tid)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
- hid_t new_fapl = H5I_INVALID_HID; /* File acess property list */
+ hid_t new_fapl = H5I_INVALID_HID; /* File access property list */
hid_t dcpl = H5I_INVALID_HID; /* Dataset creation property list */
hid_t dtid = H5I_INVALID_HID; /* Datatype ID for the dataset */
hid_t str_tid = H5I_INVALID_HID; /* String datatype ID */
@@ -7132,7 +7132,7 @@ test_libver_bounds_attributes(hid_t fapl)
fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, H5P_DEFAULT, fapl);
CHECK(fid, H5I_INVALID_HID, "H5Fcreate");
- /* Integer datatpye */
+ /* Integer datatype */
tid = H5Tcopy(H5T_NATIVE_INT);
CHECK(tid, H5I_INVALID_HID, "H5Tcopy");
@@ -7241,7 +7241,7 @@ test_libver_bounds_attributes(hid_t fapl)
fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, fcpl, fapl);
CHECK(fid, H5I_INVALID_HID, "H5Fcreate");
- /* Create an integer datatye */
+ /* Create an integer datatype */
tid = H5Tcopy(H5T_NATIVE_INT);
CHECK(tid, H5I_INVALID_HID, "H5Tcopy");
@@ -8037,7 +8037,7 @@ test_file(void)
test_libver_macros(); /* Test the macros for library version comparison */
test_libver_macros2(); /* Test the macros for library version comparison */
test_incr_filesize(); /* Test H5Fincrement_filesize() and H5Fget_eoa() */
- test_min_dset_ohdr(); /* Test datset object header minimization */
+ test_min_dset_ohdr(); /* Test dataset object header minimization */
#ifndef H5_NO_DEPRECATED_SYMBOLS
test_file_ishdf5(env_h5_drvr); /* Test detecting HDF5 files correctly */
test_deprec(); /* Test deprecated routines */
diff --git a/test/tgenprop.c b/test/tgenprop.c
index 55e5a9b..f54fbbc 100644
--- a/test/tgenprop.c
+++ b/test/tgenprop.c
@@ -107,7 +107,7 @@ test_genprop_basic_class(void)
ret = H5Pequal(cid2, H5P_ROOT);
VERIFY(ret, 1, "H5Pequal");
- /* Make certain false postives aren't being returned */
+ /* Make certain false positives aren't being returned */
ret = H5Pequal(cid2, H5P_FILE_CREATE);
VERIFY(ret, 0, "H5Pequal");
@@ -185,7 +185,7 @@ test_genprop_basic_class_prop(void)
CHECK_I(ret, "H5Pget_nprops");
VERIFY(nprops, 0, "H5Pget_nprops");
- /* Check the existance of the first property (should fail) */
+ /* Check the existence of the first property (should fail) */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 0, "H5Pexist");
@@ -199,7 +199,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister2");
- /* Check the existance of the first property */
+ /* Check the existence of the first property */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -223,7 +223,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister2");
- /* Check the existance of the second property */
+ /* Check the existence of the second property */
ret = H5Pexist(cid1, PROP2_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -242,7 +242,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
CHECK_I(ret, "H5Pregister2");
- /* Check the existance of the third property */
+ /* Check the existence of the third property */
ret = H5Pexist(cid1, PROP3_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1942,7 +1942,7 @@ test_genprop_deprec_class(void)
CHECK_I(ret, "H5Pget_nprops");
VERIFY(nprops, 0, "H5Pget_nprops");
- /* Check the existance of the first property (should fail) */
+ /* Check the existence of the first property (should fail) */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 0, "H5Pexist");
@@ -1954,7 +1954,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister1");
- /* Check the existance of the first property */
+ /* Check the existence of the first property */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1976,7 +1976,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister1");
- /* Check the existance of the second property */
+ /* Check the existence of the second property */
ret = H5Pexist(cid1, PROP2_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1994,7 +1994,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
CHECK_I(ret, "H5Pregister1");
- /* Check the existance of the third property */
+ /* Check the existence of the third property */
ret = H5Pexist(cid1, PROP3_NAME);
VERIFY(ret, 1, "H5Pexist");
diff --git a/test/th5o.c b/test/th5o.c
index ebb6c6e..1afbe17 100644
--- a/test/th5o.c
+++ b/test/th5o.c
@@ -594,7 +594,7 @@ test_h5o_refcount(void)
CHECK(ret, FAIL, "H5Oget_info_by_name3");
VERIFY(oinfo.rc, 1, "reference count in H5Oget_info_by_name3");
- /* Increment the reference counts and then close the file to make sure the increment is permanant */
+ /* Increment the reference counts and then close the file to make sure the increment is permanent */
ret = H5Oincr_refcount(grp);
CHECK(ret, FAIL, "H5Oincr_refcount");
ret = H5Oincr_refcount(dtype);
diff --git a/test/th5s.c b/test/th5s.c
index 51216dc..1a6f56e 100644
--- a/test/th5s.c
+++ b/test/th5s.c
@@ -2022,7 +2022,7 @@ test_h5s_encode_points(H5F_libver_t low, H5F_libver_t high)
{
hid_t fapl = -1; /* File access property list ID */
hid_t sid; /* Dataspace ID */
- hsize_t numparticles = 8388608; /* Used to calculate dimenion size */
+ hsize_t numparticles = 8388608; /* Used to calculate dimension size */
unsigned num_dsets = 513; /* used to calculate dimension size */
hsize_t total_particles = numparticles * num_dsets;
hsize_t vdsdims[1] = {total_particles}; /* Dimension size */
diff --git a/test/tid.c b/test/tid.c
index c7e40a0..ab06840 100644
--- a/test/tid.c
+++ b/test/tid.c
@@ -41,7 +41,7 @@ basic_id_test(void)
int num_ref;
hsize_t num_members;
- /* Try to register an ID with ficticious types */
+ /* Try to register an ID with fictitious types */
H5E_BEGIN_TRY
arrayID = H5Iregister((H5I_type_t)420, testObj);
H5E_END_TRY
@@ -58,7 +58,7 @@ basic_id_test(void)
if (arrayID != H5I_INVALID_HID)
goto out;
- /* Try to access IDs with ficticious types */
+ /* Try to access IDs with fictitious types */
H5E_BEGIN_TRY
testPtr = H5Iobject_verify((hid_t)100, (H5I_type_t)0);
H5E_END_TRY
diff --git a/test/tmisc.c b/test/tmisc.c
index 3655bed..9d7d180 100644
--- a/test/tmisc.c
+++ b/test/tmisc.c
@@ -1280,7 +1280,7 @@ test_misc8(void)
for (v = 0; v < MISC8_DIM1; v++)
*tdata++ = (int)(((u * MISC8_DIM1) + v) % 13);
- /* Create a file acccess property list */
+ /* Create a file access property list */
fapl = H5Pcreate(H5P_FILE_ACCESS);
CHECK(fapl, FAIL, "H5Pcreate");
@@ -3892,7 +3892,7 @@ test_misc21(void)
/****************************************************************
**
-** test_misc22(): Test SZIP bits-per-pixel paramter.
+** test_misc22(): Test SZIP bits-per-pixel parameter.
** This should be set according to the datatype.
** Tests for precision and offset combo's.
**
diff --git a/test/trefer.c b/test/trefer.c
index 9b6c415..0e25ec2 100644
--- a/test/trefer.c
+++ b/test/trefer.c
@@ -2732,7 +2732,7 @@ test_reference_compat_conv(void)
hid_t sid1, sid2, sid3; /* Dataspace IDs */
hid_t tid1, tid2; /* Datatype ID */
hsize_t dims1[] = {SPACE1_DIM1}, dims2[] = {SPACE2_DIM1, SPACE2_DIM2},
- dims3[] = {SPACE1_DIM1}; /* Purposedly set dimension larger to test NULL references */
+ dims3[] = {SPACE1_DIM1}; /* Purposely set dimension larger to test NULL references */
hsize_t start[SPACE2_RANK]; /* Starting location of hyperslab */
hsize_t stride[SPACE2_RANK]; /* Stride of hyperslab */
hsize_t count[SPACE2_RANK]; /* Element count of hyperslab */
diff --git a/test/tselect.c b/test/tselect.c
index 31bb9c4..7d32407 100644
--- a/test/tselect.c
+++ b/test/tselect.c
@@ -1879,7 +1879,7 @@ verify_select_hyper_contig_dr__run_test(const uint16_t *cube_buf, size_t H5_ATTR
uint16_t expected_value; /* Expected value in dataset */
unsigned i, j, k, l, m; /* Local index variables */
size_t s; /* Local index variable */
- hbool_t mis_match; /* Flag to indicate mis-match in expected value */
+ hbool_t mis_match; /* Flag to indicate mismatch in expected value */
HDassert(cube_buf);
HDassert(cube_size > 0);
@@ -2357,7 +2357,7 @@ test_select_hyper_contig_dr__run_test(int test_num, const uint16_t *cube_buf, co
/* Now write the contents of the in memory small cube to slices of
* the on disk cube. After each write, read the on disk cube
- * into memeory, and verify that it contains the expected
+ * into memory, and verify that it contains the expected
* data. Verify that H5Sselect_shape_same() returns true on
* the memory and file selections.
*/
@@ -2742,7 +2742,7 @@ test_select_hyper_checker_board_dr__select_checker_board(hid_t tgt_n_cube_sid, u
i++;
} while ((i <= 1) && (0 >= sel_offset));
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Some how, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -3401,7 +3401,7 @@ test_select_hyper_checker_board_dr__run_test(int test_num, const uint16_t *cube_
/* Now write checker board selections of the entries in memory
* small cube to slices of the on disk cube. After each write,
- * read the on disk large cube * into memeory, and verify that
+ * read the on disk large cube * into memory, and verify that
* it contains the expected * data. Verify that
* H5Sselect_shape_same() returns true on the memory and file
* selections.
@@ -5256,7 +5256,7 @@ test_select_hyper_union_stagger(void)
error = H5Fclose(file_id);
CHECK(error, FAIL, "H5Fclose");
- /* Initialize intput buffer */
+ /* Initialize input buffer */
HDmemset(data_out, 0, 7 * 7 * sizeof(int));
/* Open file */
@@ -11254,7 +11254,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i
CHECK(ret, FAIL, "H5Sselect_hyperslab");
} /* end if */
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Some how, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -11349,7 +11349,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i
CHECK(ret, FAIL, "H5Sselect_hyperslab");
} /* end if */
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Again, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -12130,7 +12130,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 2-D hyperslab selection\n"));
/* Create 2-D dataspace */
@@ -12222,7 +12222,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 3-D hyperslab selection\n"));
@@ -12324,7 +12324,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 4-D hyperslab selection\n"));
@@ -12443,7 +12443,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 5-D hyperslab selection\n"));
@@ -12572,7 +12572,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* We use 5-D to test a special case with
rebuilding routine TRUE, FALSE and TRUE */
@@ -12619,7 +12619,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* Adding some selections to make it real irregular */
start5[3] = 1;
@@ -12647,7 +12647,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* Add more selections to make it regular again */
start5[3] = 5;
@@ -12675,7 +12675,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
H5Sclose(sid_reg1);
CHECK(ret, FAIL, "H5Sclose");
@@ -12981,7 +12981,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_update_diminfo");
} /* end if */
- /* Add block parially overlapping first, with OR */
+ /* Add block partially overlapping first, with OR */
start1[0] = 4;
count1[0] = 1;
block1[0] = 2;
@@ -13026,7 +13026,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_update_diminfo");
} /* end if */
- /* Add equally sized block parially overlapping current, with XOR */
+ /* Add equally sized block partially overlapping current, with XOR */
start1[0] = 3;
count1[0] = 1;
block1[0] = 5;
@@ -13060,7 +13060,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_rebuild");
} /* end if */
- /* Add differently sized block parially overlapping current, with XOR */
+ /* Add differently sized block partially overlapping current, with XOR */
start1[0] = 4;
count1[0] = 1;
block1[0] = 5;
@@ -14236,7 +14236,7 @@ test_hyper_regular(void)
**
****************************************************************/
static void
-test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enblocks, hsize_t *eblock1,
+test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t endpoints, hssize_t enblocks, hsize_t *eblock1,
hsize_t *eblock2)
{
hid_t lim_sid;
@@ -14261,7 +14261,7 @@ test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enb
/* Check number of elements */
npoints = H5Sget_select_npoints(lim_sid);
CHECK(npoints, FAIL, "H5Sget_select_npoints");
- VERIFY(npoints, enpoints, "H5Sget_select_npoints");
+ VERIFY(npoints, endpoints, "H5Sget_select_npoints");
/* Get selection type */
sel_type = H5Sget_select_type(lim_sid);
diff --git a/test/tsohm.c b/test/tsohm.c
index b4ece0b..9678f46 100644
--- a/test/tsohm.c
+++ b/test/tsohm.c
@@ -636,7 +636,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_
if (GetTestExpress() > 1)
test_file_closing = FALSE;
- /* Intialize wdata */
+ /* Initialize wdata */
HDmemset(&wdata, 0, sizeof(wdata));
wdata.i1 = 11;
HDstrcpy(wdata.str, "string");
@@ -649,7 +649,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_
wdata.i8 = 88;
wdata.f1 = 0.0F;
- /* Intialize rdata */
+ /* Initialize rdata */
HDmemset(&rdata, 0, sizeof(rdata));
dtype1_id = make_dtype_1();
@@ -762,7 +762,7 @@ error:
* Function: getsize_testsize1
*
* Purpose: Creates a test file, populates it, and returns its file size.
- * Oject header information from the "first" dataset in the file
+ * Object header information from the "first" dataset in the file
* is stored in pointer `oinfo`.
*
* Programmer: Jacob Smith
@@ -1551,7 +1551,7 @@ size2_dump_struct(const char *name, size2_helper_struct *sizes)
/*-------------------------------------------------------------------------
* Function: size2_helper
*
- * Purpose: A helper functon for test_sohm_size2.
+ * Purpose: A helper function for test_sohm_size2.
*
* Creates a file using the given fcpl, then creates lots
* of different kinds of messages within the file and
@@ -1931,7 +1931,7 @@ size2_helper(hid_t fcpl_id, int test_file_closing, size2_helper_struct *ret_size
/*-------------------------------------------------------------------------
* Function: size2_verify
*
- * Purpose: A helper functon to verify the file created by size2_helper.
+ * Purpose: A helper function to verify the file created by size2_helper.
*
* Runs various tests (not exhaustive) to ensure that the
* file FILENAME actually has the structure that size2_helper
diff --git a/test/ttsafe_attr_vlen.c b/test/ttsafe_attr_vlen.c
index ecd0aa6..9fafaca 100644
--- a/test/ttsafe_attr_vlen.c
+++ b/test/ttsafe_attr_vlen.c
@@ -136,7 +136,7 @@ tts_attr_vlen_thread(void H5_ATTR_UNUSED *client_data)
gid = H5Gopen2(fid, "/", H5P_DEFAULT);
CHECK(gid, H5I_INVALID_HID, "H5Gopen");
- /* Open the attribte */
+ /* Open the attribute */
aid = H5Aopen(gid, "root_attr", H5P_DEFAULT);
CHECK(aid, H5I_INVALID_HID, "H5Aopen");
diff --git a/test/ttsafe_rec_rw_lock.c b/test/ttsafe_rec_rw_lock.c
index 0597578..a5033f7 100644
--- a/test/ttsafe_rec_rw_lock.c
+++ b/test/ttsafe_rec_rw_lock.c
@@ -136,7 +136,7 @@ void *tts_rw_lock_smoke_check_test_thread(void *_udata);
*
* 8) Drop the read lock a second time.
*
- * 9) Verify the exptected stats, and then reset them.
+ * 9) Verify the expected stats, and then reset them.
*
* 10) Obtain a write lock.
*
@@ -170,7 +170,7 @@ void *tts_rw_lock_smoke_check_test_thread(void *_udata);
*
* 25) Shut down the recursive R/W lock.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -242,7 +242,7 @@ tts_rec_rw_lock_smoke_check_1(void)
result = H5TS_rw_unlock(&rec_rw_lock);
CHECK_I(result, "H5TS_rw_unlock -- 3");
- /* 9) Verify the exptected stats, and then reset them. */
+ /* 9) Verify the expected stats, and then reset them. */
result = H5TS_rw_lock_get_stats(&rec_rw_lock, &stats);
CHECK_I(result, "H5TS_rw_lock_get_stats -- 2");
@@ -442,7 +442,7 @@ cleanup_rec_rw_lock_smoke_check_1(void)
* number of recursive lock and unlock calls is random, as is the
* order of read and write locks if both are enabled.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -593,7 +593,7 @@ tts_rw_lock_smoke_check_test_thread(void *_udata)
* repeats the process until the spcified number of read locks have
* been acquired and dropped.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -839,7 +839,7 @@ cleanup_rec_rw_lock_smoke_check_2(void)
* repeats the process until the spcified number of write locks have
* been acquired and dropped.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -1085,7 +1085,7 @@ cleanup_rec_rw_lock_smoke_check_3(void)
* repeats the process until the spcified number of locks have
* been acquired and dropped.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
diff --git a/test/tvltypes.c b/test/tvltypes.c
index 03a8ad3..0c4cf68 100644
--- a/test/tvltypes.c
+++ b/test/tvltypes.c
@@ -131,7 +131,7 @@ test_vltypes_dataset_create(void)
dcpl = H5Pcreate(H5P_DATASET_CREATE);
CHECK(dcpl, FAIL, "H5Pcreate");
- /* Set fill value writting time to be NEVER */
+ /* Set fill value writing time to be NEVER */
ret = H5Pset_fill_time(dcpl, H5D_FILL_TIME_NEVER);
CHECK(ret, FAIL, "H5Pset_fill_time");
diff --git a/test/twriteorder.c b/test/twriteorder.c
index 9cff00a..c58b030 100644
--- a/test/twriteorder.c
+++ b/test/twriteorder.c
@@ -54,7 +54,7 @@
*
* If the system, in which the writer and reader processes run, the readers
* will always get all chain-linked blocks correctly. If the order of write
- * is not maintained, some reader processes may found unexpect block data.
+ * is not maintained, some reader processes may found unexpected block data.
*
*************************************************************/
diff --git a/test/unlink.c b/test/unlink.c
index 94bd035..19b7b37 100644
--- a/test/unlink.c
+++ b/test/unlink.c
@@ -617,7 +617,7 @@ test_filespace(hid_t fapl)
/* Create common objects for datasets */
- /* Create dataset creation property list for contigous storage */
+ /* Create dataset creation property list for contiguous storage */
if ((contig_dcpl = H5Pcreate(H5P_DATASET_CREATE)) < 0)
TEST_ERROR
@@ -1099,7 +1099,7 @@ test_filespace(hid_t fapl)
PASSED();
- /* Create simple group hiearchy, remove it & verify file size */
+ /* Create simple group hierarchy, remove it & verify file size */
TESTING(" simple group hierarchy");
/* Create file */
@@ -1138,7 +1138,7 @@ test_filespace(hid_t fapl)
PASSED();
- /* Create complex group hiearchy, remove it & verify file size */
+ /* Create complex group hierarchy, remove it & verify file size */
TESTING(" complex group hierarchy");
/* Create file */
diff --git a/test/use_common.c b/test/use_common.c
index 8f78f3b..7b5df53 100644
--- a/test/use_common.c
+++ b/test/use_common.c
@@ -298,7 +298,7 @@ write_uc_file(hbool_t tosend, hid_t file_id, options_t *opts)
return -1;
}
- /* verify chunk_dims against set paramenters */
+ /* verify chunk_dims against set parameters */
if (chunk_dims[0] != opts->chunkdims[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) {
HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n",
(unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1],
@@ -533,7 +533,7 @@ read_uc_file(hbool_t towait, options_t *opts)
/* quit when all nplanes have been read */
loops_waiting_for_plane = 0;
while (nplanes_seen < opts->nplanes) {
- /* print progress message according to if new planes are availalbe */
+ /* print progress message according to if new planes are available */
if (nplanes_seen < dims[0]) {
if (loops_waiting_for_plane) {
/* end the previous message */
diff --git a/test/use_disable_mdc_flushes.c b/test/use_disable_mdc_flushes.c
index f28e574..d40e56d 100644
--- a/test/use_disable_mdc_flushes.c
+++ b/test/use_disable_mdc_flushes.c
@@ -360,7 +360,7 @@ write_file(void)
return -1;
}
- /* verify chunk_dims against set paramenters */
+ /* verify chunk_dims against set parameters */
if (chunk_dims[0] != chunkdims_g[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) {
HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n",
(unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1],
diff --git a/test/vds.c b/test/vds.c
index ac9bb5d..79424ea 100644
--- a/test/vds.c
+++ b/test/vds.c
@@ -196,7 +196,7 @@ vds_select_equal(hid_t space1, hid_t space2)
if (nblocks1 != nblocks2)
return FALSE;
- /* Allocate block lists. Do not return directly afer
+ /* Allocate block lists. Do not return directly after
* allocating, to make sure buffers are freed. */
if (NULL ==
(buf1 = (hsize_t *)HDmalloc((size_t)2 * (size_t)rank1 * (size_t)nblocks1 * sizeof(*buf1))))
@@ -479,7 +479,7 @@ test_api_get_ex_dcpl(test_api_config_t config, hid_t fapl, hid_t dcpl, hid_t *ex
/* Verify examination DCPL is equal to original DCPL. Do not compare the
* plist to itself, and do not do the comparison if we reopened the file,
- * because in that case the extent of the source dset will not be corrent.
+ * because in that case the extent of the source dset will not be current.
*/
if ((*ex_dcpl != dcpl) && (config != TEST_API_REOPEN_FILE)) {
if ((tri_ret = H5Pequal(dcpl, *ex_dcpl)) < 0)
@@ -1044,7 +1044,7 @@ test_api(test_api_config_t config, hid_t fapl, H5F_libver_t low)
if ((vspace[i] = H5Screate_simple(2, dims, NULL)) < 0)
TEST_ERROR
- /* Select row in virual dataspace */
+ /* Select row in virtual dataspace */
start[0] = (hsize_t)i;
if (H5Sselect_hyperslab(vspace[i], H5S_SELECT_SET, start, NULL, count, block) < 0)
TEST_ERROR
@@ -1182,7 +1182,7 @@ test_vds_prefix_first(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
int buf[10][26]; /* Write and expected read buffer */
@@ -1461,7 +1461,7 @@ test_basic_io(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
hsize_t start[4]; /* Hyperslab start */
@@ -4389,7 +4389,7 @@ test_unlim(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {10, 10}; /* Data space current size */
hsize_t mdims[2] = {10, 20}; /* Data space maximum size */
@@ -7403,7 +7403,7 @@ test_printf(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[2] = {-1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {10, 0}; /* Data space current size */
hsize_t mdims[2] = {10, 20}; /* Data space maximum size */
@@ -11111,7 +11111,7 @@ test_all(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[3] = {-1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {6, 6}; /* Data space current size */
hsize_t mdims[2] = {10, 10}; /* Data space maximum size */
diff --git a/test/vds_env.c b/test/vds_env.c
index 1f00444..db54ea0 100644
--- a/test/vds_env.c
+++ b/test/vds_env.c
@@ -58,7 +58,7 @@ test_vds_prefix_second(unsigned config, hid_t fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
int buf[10][26]; /* Write and expected read buffer */
diff --git a/test/vfd.c b/test/vfd.c
index e90a589..b0f37c3 100644
--- a/test/vfd.c
+++ b/test/vfd.c
@@ -186,7 +186,7 @@ test_sec2(void)
if (os_file_handle == NULL)
FAIL_PUTS_ERROR("NULL os-specific vfd/file handle was returned from H5Fget_vfd_handle");
- /* There is no garantee the size of metadata in file is constant.
+ /* There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* Currently it should be around 2 KB.
@@ -382,7 +382,7 @@ test_core(void)
if (os_file_handle == NULL)
FAIL_PUTS_ERROR("NULL os-specific vfd/file handle was returned from H5Fget_vfd_handle");
- /* There is no garantee the size of metadata in file is constant.
+ /* There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* TODO: Needs justification of why is this is a reasonable size.
@@ -529,7 +529,7 @@ test_core(void)
} /* end if */
/* Check file size API.
- * There is no garantee the size of metadata in file is constant.
+ * There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* TODO: Needs justification of why is this is a reasonable size.
@@ -1101,7 +1101,7 @@ error:
* See if we can open files created with v1.6 library.
* The source file was created by the test/file_handle.c
* of the v1.6 library. Then tools/misc/h5repart.c was
- * used to concantenated. The command was "h5repart -m 5k
+ * used to concatenated. The command was "h5repart -m 5k
* family_file%05d.h5 family_v16_%05d.h5".
*
* Return: Success: 0
@@ -1369,7 +1369,7 @@ H5_GCC_DIAG_ON("format-nonliteral")
/*-------------------------------------------------------------------------
* Function: test_multi
*
- * Purpose: Tests the file handle interface for MUTLI driver
+ * Purpose: Tests the file handle interface for MULTI driver
*
* Return: SUCCEED/FAIL
*
@@ -2524,7 +2524,7 @@ done:
* Function: driver_is_splitter_compatible
*
* Purpose: Determine whether the driver set in the FAPL ID is compatible
- * with the Splitter VFD -- specificially, Write-Only channel.
+ * with the Splitter VFD -- specifically, Write-Only channel.
*
* Return: Success: 0
* Failure: -1
@@ -3097,7 +3097,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3117,7 +3117,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3148,7 +3148,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3179,7 +3179,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");