summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
authorLarry Knox <lrknox@hdfgroup.org>2022-04-07 04:49:00 (GMT)
committerLarry Knox <lrknox@hdfgroup.org>2022-04-07 04:49:00 (GMT)
commit91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f (patch)
tree1ca4bfd929700b0743a3ecc653262e836e6436f2
parent9f2b38bb61df6136d214cb7ab248f73d1ff9458c (diff)
downloadhdf5-91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f.zip
hdf5-91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f.tar.gz
hdf5-91a5eb2cd9fef54b13402e2a11e10ce98f8a3d1f.tar.bz2
Commit codespell spelling corrections.
-rw-r--r--.github/workflows/codespell.yml15
-rwxr-xr-xbin/buildhdf52
-rwxr-xr-xbin/cmakehdf54
-rwxr-xr-xbin/genparser2
-rw-r--r--bin/h5cc.in6
-rwxr-xr-xbin/newer2
-rwxr-xr-xbin/warnhist2
-rw-r--r--c++/src/H5Attribute.cpp2
-rw-r--r--c++/src/H5DataSet.cpp2
-rw-r--r--c++/src/H5DataSpace.cpp2
-rw-r--r--c++/src/H5DataType.cpp4
-rw-r--r--c++/src/H5DcreatProp.cpp2
-rw-r--r--c++/src/H5File.cpp2
-rw-r--r--c++/src/H5Group.cpp2
-rw-r--r--c++/src/H5Location.cpp2
-rw-r--r--c++/src/H5Location.h2
-rw-r--r--c++/src/H5PredType.cpp2
-rw-r--r--c++/src/H5PropList.cpp2
-rw-r--r--c++/test/tobject.cpp2
-rw-r--r--c++/test/ttypes.cpp2
-rw-r--r--config/BlankForm4
-rw-r--r--config/cmake_ext_mod/HDFUseCXX.cmake2
-rw-r--r--configure.ac6
-rw-r--r--examples/h5_mount.c2
-rw-r--r--examples/h5_shared_mesg.c4
-rw-r--r--examples/h5_vds-percival-unlim-maxmin.c2
-rw-r--r--examples/testh5cc.sh.in2
-rw-r--r--fortran/examples/compound.f9014
-rw-r--r--fortran/examples/h5_cmprss.f902
-rw-r--r--fortran/examples/h5_crtatt.f902
-rw-r--r--fortran/examples/h5_extend.f904
-rw-r--r--fortran/examples/refobjexample.f902
-rw-r--r--fortran/examples/testh5fc.sh.in2
-rw-r--r--fortran/src/H5Af.c8
-rw-r--r--fortran/src/H5Aff.F9014
-rw-r--r--fortran/src/H5Df.c6
-rw-r--r--fortran/src/H5Dff.F906
-rw-r--r--fortran/src/H5Fff.F902
-rw-r--r--fortran/src/H5Gf.c8
-rw-r--r--fortran/src/H5Gff.F9032
-rw-r--r--fortran/src/H5Lff.F906
-rw-r--r--fortran/src/H5Pf.c30
-rw-r--r--fortran/src/H5Pff.F9016
-rw-r--r--fortran/src/H5Sf.c8
-rw-r--r--fortran/src/H5Sff.F902
-rw-r--r--fortran/src/H5Tf.c6
-rw-r--r--fortran/src/H5Tff.F904
-rw-r--r--fortran/src/H5_buildiface.F908
-rw-r--r--fortran/src/H5_ff.F904
-rw-r--r--fortran/src/H5f90global.F904
-rw-r--r--fortran/src/H5f90proto.h2
-rw-r--r--fortran/src/h5fc.in6
-rw-r--r--fortran/test/H5_test_buildiface.F902
-rw-r--r--fortran/test/fflush1.F902
-rw-r--r--fortran/test/fflush2.F904
-rw-r--r--fortran/test/tH5A.F902
-rw-r--r--fortran/test/tH5A_1_8.F9010
-rw-r--r--fortran/test/tH5D.F9016
-rw-r--r--fortran/test/tH5F.F902
-rw-r--r--fortran/test/tH5G_1_8.F902
-rw-r--r--fortran/test/tH5MISC_1_8.F902
-rw-r--r--fortran/test/tH5P.F906
-rw-r--r--fortran/test/tH5P_F03.F9012
-rw-r--r--fortran/test/tH5R.F902
-rw-r--r--fortran/test/tH5S.F9016
-rw-r--r--fortran/test/tH5Sselect.F9012
-rw-r--r--fortran/test/tH5T.F9018
-rw-r--r--fortran/test/tH5VL.F904
-rw-r--r--fortran/test/tH5Z.F902
-rw-r--r--hl/c++/test/ptableTest.cpp4
-rw-r--r--hl/examples/ex_image1.c4
-rw-r--r--hl/fortran/src/H5DSff.F902
-rw-r--r--hl/fortran/src/H5HL_buildiface.F902
-rw-r--r--hl/fortran/src/H5LTff.F904
-rw-r--r--hl/fortran/src/H5TBfc.c4
-rw-r--r--hl/fortran/src/H5TBff.F902
-rw-r--r--hl/fortran/test/tsttable.F908
-rw-r--r--hl/src/H5DO.c2
-rw-r--r--hl/src/H5DS.c8
-rw-r--r--hl/test/test_ds.c28
-rw-r--r--hl/test/test_file_image.c2
-rw-r--r--hl/test/test_h5do_compat.c2
-rw-r--r--hl/test/test_image.c4
-rw-r--r--hl/test/test_ld.c6
-rw-r--r--hl/test/test_packet_vlen.c2
-rw-r--r--hl/test/test_table.c2
-rw-r--r--hl/tools/gif2h5/decompress.c4
-rw-r--r--hl/tools/gif2h5/gif2mem.c2
-rw-r--r--hl/tools/gif2h5/gifread.c44
-rw-r--r--hl/tools/gif2h5/h52giftest.sh.in2
-rw-r--r--hl/tools/gif2h5/writehdf.c4
-rw-r--r--hl/tools/h5watch/extend_dset.c4
-rw-r--r--m4/aclocal_cxx.m42
-rw-r--r--release_docs/HISTORY-1_0-1_8_0_rc3.txt80
-rw-r--r--release_docs/HISTORY-1_10.txt32
-rw-r--r--release_docs/HISTORY-1_10_0-1_12_0.txt2
-rw-r--r--release_docs/HISTORY-1_12.txt4
-rw-r--r--release_docs/HISTORY-1_8.txt50
-rw-r--r--release_docs/HISTORY-1_8_0-1_10_0.txt26
-rw-r--r--release_docs/INSTALL_Cygwin.txt4
-rw-r--r--release_docs/RELEASE.txt2
-rw-r--r--src/H5A.c4
-rw-r--r--src/H5AC.c4
-rw-r--r--src/H5ACdbg.c6
-rw-r--r--src/H5ACmpio.c4
-rw-r--r--src/H5Adense.c32
-rw-r--r--src/H5Adeprec.c2
-rw-r--r--src/H5C.c18
-rw-r--r--src/H5CS.c4
-rw-r--r--src/H5CX.c2
-rw-r--r--src/H5Cdbg.c4
-rw-r--r--src/H5Cimage.c12
-rw-r--r--src/H5Cmpio.c2
-rw-r--r--src/H5Cprefetched.c2
-rw-r--r--src/H5Dchunk.c8
-rw-r--r--src/H5Dint.c8
-rw-r--r--src/H5Dmpio.c8
-rw-r--r--src/H5Dvirtual.c12
-rw-r--r--src/H5E.c2
-rw-r--r--src/H5F.c2
-rw-r--r--src/H5FA.c2
-rw-r--r--src/H5FAcache.c2
-rw-r--r--src/H5FD.c2
-rw-r--r--src/H5FDcore.c2
-rw-r--r--src/H5FDhdfs.c8
-rw-r--r--src/H5FDlog.c2
-rw-r--r--src/H5FDmirror.c2
-rw-r--r--src/H5FDros3.c10
-rw-r--r--src/H5FSprivate.h2
-rw-r--r--src/H5FSsection.c6
-rw-r--r--src/H5Fefc.c2
-rw-r--r--src/H5Fint.c12
-rw-r--r--src/H5Fmount.c4
-rw-r--r--src/H5Fsuper.c4
-rw-r--r--src/H5Fsuper_cache.c4
-rw-r--r--src/H5Gcache.c4
-rw-r--r--src/H5Gcompact.c2
-rw-r--r--src/H5Gnode.c2
-rw-r--r--src/H5HFcache.c2
-rw-r--r--src/H5HFdbg.c4
-rw-r--r--src/H5HFdblock.c2
-rw-r--r--src/H5HFspace.c2
-rw-r--r--src/H5I.c2
-rw-r--r--src/H5MFaggr.c2
-rw-r--r--src/H5MFdbg.c2
-rw-r--r--src/H5Oainfo.c2
-rw-r--r--src/H5Oattr.c2
-rw-r--r--src/H5Oattribute.c10
-rw-r--r--src/H5Obogus.c4
-rw-r--r--src/H5Obtreek.c2
-rw-r--r--src/H5Ocache.c2
-rw-r--r--src/H5Ocache_image.c2
-rw-r--r--src/H5Ocont.c2
-rw-r--r--src/H5Ocopy.c2
-rw-r--r--src/H5Odrvinfo.c2
-rw-r--r--src/H5Odtype.c2
-rw-r--r--src/H5Oefl.c2
-rw-r--r--src/H5Ofill.c4
-rw-r--r--src/H5Ofsinfo.c4
-rw-r--r--src/H5Oginfo.c2
-rw-r--r--src/H5Olayout.c2
-rw-r--r--src/H5Olinfo.c2
-rw-r--r--src/H5Olink.c2
-rw-r--r--src/H5Omessage.c8
-rw-r--r--src/H5Omtime.c4
-rw-r--r--src/H5Oname.c2
-rw-r--r--src/H5Onull.c2
-rw-r--r--src/H5Opline.c4
-rw-r--r--src/H5Orefcount.c2
-rw-r--r--src/H5Osdspace.c2
-rw-r--r--src/H5Oshmesg.c2
-rw-r--r--src/H5Ostab.c2
-rw-r--r--src/H5Ounknown.c2
-rw-r--r--src/H5P.c8
-rw-r--r--src/H5PL.c14
-rw-r--r--src/H5PLpath.c2
-rw-r--r--src/H5Pdxpl.c2
-rw-r--r--src/H5Pencdec.c2
-rw-r--r--src/H5Pfapl.c8
-rw-r--r--src/H5Pint.c10
-rw-r--r--src/H5Pocpl.c14
-rw-r--r--src/H5S.c2
-rw-r--r--src/H5SL.c8
-rw-r--r--src/H5SM.c4
-rw-r--r--src/H5Shyper.c20
-rw-r--r--src/H5Smpio.c2
-rw-r--r--src/H5Spoint.c4
-rw-r--r--src/H5Sselect.c6
-rw-r--r--src/H5Stest.c10
-rw-r--r--src/H5T.c2
-rw-r--r--src/H5TS.c14
-rw-r--r--src/H5TSprivate.h6
-rw-r--r--src/H5Tbit.c10
-rw-r--r--src/H5Tconv.c42
-rw-r--r--src/H5Tfields.c2
-rw-r--r--src/H5Tnative.c2
-rw-r--r--src/H5Topaque.c2
-rw-r--r--src/H5VLcallback.c8
-rw-r--r--src/H5VLint.c4
-rw-r--r--src/H5VLnative.c4
-rw-r--r--src/H5VLnative_introspect.c2
-rw-r--r--src/H5VLpassthru.c2
-rw-r--r--src/H5VM.c2
-rw-r--r--src/H5WB.c2
-rw-r--r--src/H5Z.c2
-rw-r--r--src/H5Znbit.c2
-rw-r--r--src/H5Zscaleoffset.c4
-rw-r--r--src/H5Ztrans.c6
-rw-r--r--src/H5detect.c4
-rw-r--r--src/H5mpi.c2
-rw-r--r--src/H5private.h2
-rw-r--r--test/SWMR_POSIX_Order_UG.txt4
-rw-r--r--test/SWMR_UseCase_UG.txt6
-rw-r--r--test/accum.c4
-rw-r--r--test/atomic_writer.c2
-rw-r--r--test/big.c6
-rw-r--r--test/btree2.c130
-rw-r--r--test/cache.c66
-rw-r--r--test/cache_api.c8
-rw-r--r--test/cache_common.c10
-rw-r--r--test/cache_image.c12
-rw-r--r--test/cache_tagging.c46
-rw-r--r--test/chunk_info.c4
-rw-r--r--test/cmpd_dset.c2
-rw-r--r--test/cork.c6
-rw-r--r--test/direct_chunk.c14
-rw-r--r--test/dsets.c40
-rw-r--r--test/dt_arith.c22
-rw-r--r--test/dtypes.c12
-rw-r--r--test/earray.c18
-rw-r--r--test/enc_dec_plist_cross_platform.c2
-rw-r--r--test/enum.c2
-rw-r--r--test/external.c6
-rw-r--r--test/farray.c12
-rw-r--r--test/fheap.c8
-rw-r--r--test/file_image.c8
-rw-r--r--test/fillval.c4
-rw-r--r--test/filter_fail.c4
-rw-r--r--test/filter_plugin1_dsets.c2
-rw-r--r--test/filter_plugin3_dsets.c2
-rw-r--r--test/flushrefresh.c12
-rw-r--r--test/gen_bounds.c12
-rw-r--r--test/gen_old_group.c2
-rw-r--r--test/gen_plist.c2
-rw-r--r--test/genall5.c2
-rw-r--r--test/getname.c2
-rw-r--r--test/gheap.c2
-rw-r--r--test/h5test.c2
-rw-r--r--test/hdfs.c2
-rw-r--r--test/links.c8
-rw-r--r--test/mf.c44
-rw-r--r--test/mirror_vfd.c2
-rw-r--r--test/mount.c4
-rw-r--r--test/mtime.c2
-rw-r--r--test/objcopy.c12
-rw-r--r--test/ohdr.c6
-rw-r--r--test/page_buffer.c24
-rw-r--r--test/ros3.c4
-rw-r--r--test/s3comms.c22
-rw-r--r--test/set_extent.c14
-rw-r--r--test/stab.c6
-rw-r--r--test/swmr.c10
-rw-r--r--test/swmr_generator.c2
-rw-r--r--test/swmr_start_write.c2
-rw-r--r--test/tattr.c12
-rw-r--r--test/tcoords.c2
-rw-r--r--test/testmeta.c2
-rw-r--r--test/tfile.c40
-rw-r--r--test/tgenprop.c18
-rw-r--r--test/th5o.c2
-rw-r--r--test/th5s.c2
-rw-r--r--test/tid.c4
-rw-r--r--test/tmisc.c4
-rw-r--r--test/trefer.c2
-rw-r--r--test/tselect.c40
-rw-r--r--test/tsohm.c10
-rw-r--r--test/ttsafe_attr_vlen.c2
-rw-r--r--test/ttsafe_rec_rw_lock.c14
-rw-r--r--test/tvltypes.c2
-rw-r--r--test/twriteorder.c2
-rw-r--r--test/unlink.c6
-rw-r--r--test/use_common.c4
-rw-r--r--test/use_disable_mdc_flushes.c2
-rw-r--r--test/vds.c16
-rw-r--r--test/vds_env.c2
-rw-r--r--test/vfd.c20
-rw-r--r--testpar/t_2Gio.c48
-rw-r--r--testpar/t_bigio.c4
-rw-r--r--testpar/t_cache.c32
-rw-r--r--testpar/t_cache_image.c28
-rw-r--r--testpar/t_coll_chunk.c16
-rw-r--r--testpar/t_dset.c46
-rw-r--r--testpar/t_file.c2
-rw-r--r--testpar/t_mdset.c6
-rw-r--r--testpar/t_mpi.c4
-rw-r--r--testpar/t_ph5basic.c2
-rw-r--r--testpar/t_pread.c10
-rw-r--r--testpar/t_shapesame.c64
-rw-r--r--testpar/t_span_tree.c30
-rw-r--r--testpar/testphdf5.h4
-rw-r--r--tools/lib/h5diff.c2
-rw-r--r--tools/src/h5dump/h5dump_ddl.c2
-rw-r--r--tools/src/h5dump/h5dump_xml.c8
-rw-r--r--tools/src/h5import/h5import.c2
-rw-r--r--tools/src/h5perf/pio_engine.c4
-rw-r--r--tools/src/h5perf/sio_engine.c2
-rw-r--r--tools/src/h5perf/sio_perf.c2
-rw-r--r--tools/src/h5repack/h5repack.h2
-rw-r--r--tools/src/h5repack/h5repack_copy.c46
-rw-r--r--tools/src/h5repack/h5repack_filters.c4
-rw-r--r--tools/src/h5repack/h5repack_main.c4
-rw-r--r--tools/src/h5repack/h5repack_refs.c2
-rw-r--r--tools/src/h5repack/h5repack_verify.c2
-rw-r--r--tools/test/h5copy/h5copygentest.c2
-rw-r--r--tools/test/h5copy/testh5copy.sh.in2
-rw-r--r--tools/test/h5diff/h5diff_plugin.sh.in6
-rw-r--r--tools/test/h5diff/h5diffgentest.c12
-rw-r--r--tools/test/h5diff/testh5diff.sh.in8
-rw-r--r--tools/test/h5dump/h5dumpgentest.c30
-rw-r--r--tools/test/h5dump/testh5dump.sh.in2
-rw-r--r--tools/test/h5dump/testh5dumppbits.sh.in2
-rw-r--r--tools/test/h5format_convert/h5fc_gentest.c2
-rw-r--r--tools/test/h5format_convert/testh5fc.sh.in2
-rw-r--r--tools/test/h5jam/testh5jam.sh.in2
-rw-r--r--tools/test/h5ls/h5ls_plugin.sh.in2
-rw-r--r--tools/test/h5ls/testh5ls.sh.in4
-rw-r--r--tools/test/h5ls/testh5lsvds.sh.in4
-rw-r--r--tools/test/h5repack/dynlib_rpk.c2
-rw-r--r--tools/test/h5repack/dynlib_vrpk.c2
-rw-r--r--tools/test/h5repack/h5repack.sh.in6
-rw-r--r--tools/test/h5repack/h5repackgentest.c4
-rw-r--r--tools/test/h5repack/h5repacktst.c16
-rw-r--r--tools/test/h5repack/testfiles/README2
-rw-r--r--tools/test/h5stat/testh5stat.sh.in2
-rw-r--r--tools/test/misc/h5clear_gentest.c2
-rw-r--r--tools/test/misc/h5perf_gentest.c6
-rw-r--r--tools/test/perform/direct_write_perf.c2
-rw-r--r--tools/test/perform/zip_perf.c2
-rw-r--r--utils/mirror_vfd/mirror_server.c2
-rw-r--r--utils/mirror_vfd/mirror_server_stop.c4
340 files changed, 1342 insertions, 1327 deletions
diff --git a/.github/workflows/codespell.yml b/.github/workflows/codespell.yml
new file mode 100644
index 0000000..f8055e9
--- /dev/null
+++ b/.github/workflows/codespell.yml
@@ -0,0 +1,15 @@
+# GitHub Action to automate the identification of common misspellings in text files
+# https://github.com/codespell-project/codespell
+# https://github.com/codespell-project/actions-codespell
+name: codespell
+on: [push, pull_request]
+jobs:
+ codespell:
+ name: Check for spelling errors
+ runs-on: ubuntu-latest
+ steps:
+ - uses: actions/checkout@v2
+ - uses: codespell-project/actions-codespell@master
+ with:
+ skip: ./bin/trace,./hl/tools/h5watch/h5watch.c,./tools/test/h5jam/tellub.c
+ ignore_words_list: isnt,inout,nd,parms,parm,ba,offsetP,ser,ois,had,fiter,fo,clude,refere,minnum,offsetp,creat,ans:,eiter,lastr,ans,isn't,ifset,sur,trun,dne,tthe,hda,filname,te,htmp,minnum,ro,oce,Ot
diff --git a/bin/buildhdf5 b/bin/buildhdf5
index 113a278..786c35d 100755
--- a/bin/buildhdf5
+++ b/bin/buildhdf5
@@ -178,7 +178,7 @@ LOCATE_SZLIB()
esac
;; # end of case ncsa
unknown)
- # Unknow domain. Give a shot at the some standard places.
+ # Unknown domain. Give a shot at the some standard places.
szlibpaths="/usr/local"
;;
esac # end of case $mydomain
diff --git a/bin/cmakehdf5 b/bin/cmakehdf5
index da1aef8..e59c772 100755
--- a/bin/cmakehdf5
+++ b/bin/cmakehdf5
@@ -199,7 +199,7 @@ DUMP_LOGFILE()
# Show a start time stamp
TIMESTAMP
-# Initialize njobs if $AMKE is defined
+# Initialize njobs if $MAKE is defined
if [ -n "$MAKE" ]; then
# assume all arguments are for --jobs
njobs=`echo $MAKE | cut -s -d' ' -f2-`
@@ -365,7 +365,7 @@ STEP "Test the library and tools..." "ctest . -C Release $njobs" $testlog
# 7. Create an install image with this command:
STEP "Create an install image..." "cpack -C Release CPackConfig.cmake" $packlog
-# The implementation of installation is imcomplete (only works for linux).
+# The implementation of installation is incomplete (only works for linux).
# Screen it out for now till it is completed.
if false; then
# 8. Install with this command:
diff --git a/bin/genparser b/bin/genparser
index 462c2c2..9ea3152 100755
--- a/bin/genparser
+++ b/bin/genparser
@@ -27,7 +27,7 @@
# IMPORTANT OS X NOTE
#
# If you are using OS X, you will probably not have flex or bison
-# installed. In addtion, even if you do have bison installed, the bison
+# installed. In addition, even if you do have bison installed, the bison
# version you have installed may also have a bug that makes it unable to
# process our input files.
#
diff --git a/bin/h5cc.in b/bin/h5cc.in
index 6024b4d..6835b79 100644
--- a/bin/h5cc.in
+++ b/bin/h5cc.in
@@ -40,7 +40,7 @@ HL="@HL@"
## $CLINKER $H5BLD_CPPFLAGS $CPPFLAGS $H5BLD_CFLAGS $CFLAGS $LDFLAGS ##
## $LIBS $clibpath $link_objs $link_args $shared_link ##
## ##
-## These settings can be overriden by setting HDF5_CFLAGS, ##
+## These settings can be overridden by setting HDF5_CFLAGS, ##
## HDF5_CPPFLAGS, HDF5_LDFLAGS, or HDF5_LIBS in the environment. ##
## ##
############################################################################
@@ -144,7 +144,7 @@ usage() {
echo " shared libraries]"
echo " "
echo " You can also add or change paths and flags to the compile line using"
- echo " the following environment varibles or by assigning them to their counterparts"
+ echo " the following environment variables or by assigning them to their counterparts"
echo " in the 'Things You Can Modify to Override...'" section of $prog_name
echo " "
echo " Variable Current value to be replaced"
@@ -318,7 +318,7 @@ fi
if test "x$do_link" = "xyes"; then
shared_link=""
-# conditionnaly link with the hl library
+# conditionally link with the hl library
if test "X$HL" = "Xhl"; then
libraries=" $libraries -lhdf5_hl -lhdf5 "
else
diff --git a/bin/newer b/bin/newer
index e60ec45..5f0fdca 100755
--- a/bin/newer
+++ b/bin/newer
@@ -19,7 +19,7 @@
# Created Date: 2005/07/06
# Modification:
# Albert Cheng 2005/8/30
-# Changed from two arguments to mulitple arguments.
+# Changed from two arguments to multiple arguments.
if test $# -lt 2; then
exit 1
diff --git a/bin/warnhist b/bin/warnhist
index 742a6b5..0150138 100755
--- a/bin/warnhist
+++ b/bin/warnhist
@@ -78,7 +78,7 @@ sub do_help {
print "\t-S <file string list>\tDisplay warnings for files which contain a string, <file string list>\n";
print "\t\t<file string list> is a comma separated list, with no spaces\n";
print "\t\tFor example: 'H5Fint' or 'H5Fint,H5Gnode'\n";
- print "\t-l\tDisplay line nunbers for file/warning\n";
+ print "\t-l\tDisplay line numbers for file/warning\n";
print "\t-u\tLeave 'unique' types in warnings, instead of genericizing them\n";
print "\t-i <name list>\tIgnore named files, <name list>\n";
print "\t\t<name list> is a comma separated list, with no spaces\n";
diff --git a/c++/src/H5Attribute.cpp b/c++/src/H5Attribute.cpp
index 0e37c2b..fbb4ad6 100644
--- a/c++/src/H5Attribute.cpp
+++ b/c++/src/H5Attribute.cpp
@@ -562,7 +562,7 @@ Attribute::p_read_variable_len(const DataType &mem_type, H5std_string &strg) con
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/src/H5DataSet.cpp b/c++/src/H5DataSet.cpp
index 47031db..791cd67 100644
--- a/c++/src/H5DataSet.cpp
+++ b/c++/src/H5DataSet.cpp
@@ -770,7 +770,7 @@ DataSet::p_read_variable_len(const hid_t mem_type_id, const hid_t mem_space_id,
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
//--------------------------------------------------------------------------
diff --git a/c++/src/H5DataSpace.cpp b/c++/src/H5DataSpace.cpp
index 0485530..be64a17 100644
--- a/c++/src/H5DataSpace.cpp
+++ b/c++/src/H5DataSpace.cpp
@@ -631,7 +631,7 @@ DataSpace::getId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/src/H5DataType.cpp b/c++/src/H5DataType.cpp
index 3228dcb..07a1463 100644
--- a/c++/src/H5DataType.cpp
+++ b/c++/src/H5DataType.cpp
@@ -932,7 +932,7 @@ DataType::p_opentype(const H5Location &loc, const char *dtype_name) const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
@@ -988,7 +988,7 @@ DataType::close()
// - Replaced decRefCount with close() to let the C library
// handle the reference counting - BMR, Jun 1, 2006
// - Added the use of H5CPP_EXITED to terminate the HDF5 library
-// and elimiate previous memory leaks. See comments in the
+// and eliminate previous memory leaks. See comments in the
// header file "H5PredType.h" for details. - BMR, Mar 30, 2012
// - Major re-implementation of the global constants was done
// to avoid relying on the order of the creation and deletion
diff --git a/c++/src/H5DcreatProp.cpp b/c++/src/H5DcreatProp.cpp
index 4c71d5e..5d07943 100644
--- a/c++/src/H5DcreatProp.cpp
+++ b/c++/src/H5DcreatProp.cpp
@@ -553,7 +553,7 @@ DSetCreatPropList::setShuffle() const
///\exception H5::PropListIException
///\par Description
/// The values of space allocation time can be one of the
-/// followings:
+/// following:
/// \li \c H5D_ALLOC_TIME_DEFAULT
/// \li \c H5D_ALLOC_TIME_EARLY
/// \li \c H5D_ALLOC_TIME_LATE
diff --git a/c++/src/H5File.cpp b/c++/src/H5File.cpp
index 9a6f191..9735a3f 100644
--- a/c++/src/H5File.cpp
+++ b/c++/src/H5File.cpp
@@ -667,7 +667,7 @@ H5File::getLocId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// December 2000
diff --git a/c++/src/H5Group.cpp b/c++/src/H5Group.cpp
index fe791da..b017f4f 100644
--- a/c++/src/H5Group.cpp
+++ b/c++/src/H5Group.cpp
@@ -208,7 +208,7 @@ Group::getId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/src/H5Location.cpp b/c++/src/H5Location.cpp
index 065de0c..3436a5b 100644
--- a/c++/src/H5Location.cpp
+++ b/c++/src/H5Location.cpp
@@ -1747,7 +1747,7 @@ H5Location::getObjinfo(const H5std_string &name, hbool_t follow_link, H5G_stat_t
// Function: H5Location::getObjinfo
///\brief This is an overloaded member function, provided for convenience.
/// It differs from the above functions in that it doesn't have
-/// the paramemter \a follow_link.
+/// the parameter \a follow_link.
// Nov, 2005
//--------------------------------------------------------------------------
void
diff --git a/c++/src/H5Location.h b/c++/src/H5Location.h
index 27ddd54..0aec4d2 100644
--- a/c++/src/H5Location.h
+++ b/c++/src/H5Location.h
@@ -15,7 +15,7 @@
#ifndef H5Location_H
#define H5Location_H
-#include "H5Classes.h" // constains forward class declarations
+#include "H5Classes.h" // contains forward class declarations
namespace H5 {
diff --git a/c++/src/H5PredType.cpp b/c++/src/H5PredType.cpp
index cdd4590..fb44961 100644
--- a/c++/src/H5PredType.cpp
+++ b/c++/src/H5PredType.cpp
@@ -914,7 +914,7 @@ September 2015:
classes, such as PropList, PredType, DataSpace, etc... Previously,
these global constants were declared statically and the C++ library used
a constant, called PredType::AtExit, to detect when all the global
- contants are destroyed then close the C library (H5close). This method
+ constants are destroyed then close the C library (H5close). This method
relied on the order of the constants being created and destroyed and
that PredType constants be the last to be destroyed. In September
2015, it was recognized that the order in which the global constants were
diff --git a/c++/src/H5PropList.cpp b/c++/src/H5PropList.cpp
index 1300641..9c149bf 100644
--- a/c++/src/H5PropList.cpp
+++ b/c++/src/H5PropList.cpp
@@ -305,7 +305,7 @@ PropList::getId() const
///\exception H5::IdComponentException when the attempt to close the HDF5
/// object fails
// Description:
-// The underlaying reference counting in the C library ensures
+// The underlying reference counting in the C library ensures
// that the current valid id of this object is properly closed.
// Then the object's id is reset to the new id.
// Programmer Binh-Minh Ribler - 2000
diff --git a/c++/test/tobject.cpp b/c++/test/tobject.cpp
index 5892f39..ecf7f7e 100644
--- a/c++/test/tobject.cpp
+++ b/c++/test/tobject.cpp
@@ -732,7 +732,7 @@ test_object()
MESSAGE(5, ("Testing Object Functions\n"));
test_get_objname(); // Test get object name from groups/datasets
- test_existance(); // Test check for object existance
+ test_existance(); // Test check for object existence
test_get_objname_ontypes(); // Test get object name from types
test_get_objtype(); // Test get object type
test_open_object_header(); // Test object header functions (H5O)
diff --git a/c++/test/ttypes.cpp b/c++/test/ttypes.cpp
index 50f9282..ae98927 100644
--- a/c++/test/ttypes.cpp
+++ b/c++/test/ttypes.cpp
@@ -28,7 +28,7 @@ using namespace H5;
#include "h5cpputil.h" // C++ utilility header file
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
diff --git a/config/BlankForm b/config/BlankForm
index ce4bf53..14131b2 100644
--- a/config/BlankForm
+++ b/config/BlankForm
@@ -61,7 +61,7 @@ fi
#
# PROD_CFLAGS Flags to pass to the compiler to create a
# PROD_CPPFLAGS production version of the library. These
-# usualy exclude symbolic debugging switches
+# usually exclude symbolic debugging switches
# (like `-g') and include optimization switches
# (like `-O').
#
@@ -84,7 +84,7 @@ fi
# compiler or a descendent of gcc such as EGCS or PGCC.
#
# AM_CFLAGS Flags added directly into this variable will
-# be propogated to the compiler wrapper scripts (h5cc,
+# be propagated to the compiler wrapper scripts (h5cc,
# h5c++, et cetera) in addition to being used to compile
# the library.
#
diff --git a/config/cmake_ext_mod/HDFUseCXX.cmake b/config/cmake_ext_mod/HDFUseCXX.cmake
index fcd3ba6..6788773 100644
--- a/config/cmake_ext_mod/HDFUseCXX.cmake
+++ b/config/cmake_ext_mod/HDFUseCXX.cmake
@@ -17,7 +17,7 @@ ENABLE_LANGUAGE (CXX)
set (HDF_PREFIX "H5")
#-------------------------------------------------------------------------------
-# Fix CXX flags if we are compiling staticly on Windows using
+# Fix CXX flags if we are compiling statically on Windows using
# Windows_MT.cmake from config/cmake/UserMacros
#-------------------------------------------------------------------------------
if (BUILD_STATIC_CRT_LIBS)
diff --git a/configure.ac b/configure.ac
index d033599..3c21cfd 100644
--- a/configure.ac
+++ b/configure.ac
@@ -390,7 +390,7 @@ if test "X$CC_BASENAME" = "Xclang"; then
# 'address;undefined'. Which and which combinations of these are
# supported varies by compiler version, but unsupported options
# or combinations will result in configure errors reported in config.log.
- # Comma separated lists of sanitize options wil be entered intact in
+ # Comma separated lists of sanitize options will be entered intact in
# one -fsanitize=<list> flag. Space separated lists will be entered in
# separate -fsanitize=<item> flags.
# NOTE: No sanity checking done here!
@@ -3502,7 +3502,7 @@ fi
## ----------------------------------------------------------------------
## Decide whether the data accuracy has higher priority during data
-## conversions. If not, some hard conversions will still be prefered even
+## conversions. If not, some hard conversions will still be preferred even
## though the data may be wrong (for example, some compilers don't
## support denormalized floating values) to maximize speed.
##
@@ -3516,7 +3516,7 @@ AC_ARG_ENABLE([dconv-accuracy],
if test "$DATA_ACCURACY" = "yes"; then
AC_MSG_RESULT([yes])
AC_DEFINE([WANT_DATA_ACCURACY], [1],
- [Data accuracy is prefered to speed during data conversions])
+ [Data accuracy is preferred to speed during data conversions])
else
AC_MSG_RESULT([no])
fi
diff --git a/examples/h5_mount.c b/examples/h5_mount.c
index e984d5b..3f087f7 100644
--- a/examples/h5_mount.c
+++ b/examples/h5_mount.c
@@ -109,7 +109,7 @@ main(void)
}
/*
- * Close all identifers
+ * Close all identifiers
*/
H5Tclose(tid);
H5Dclose(did);
diff --git a/examples/h5_shared_mesg.c b/examples/h5_shared_mesg.c
index 0f623b9..f26b360 100644
--- a/examples/h5_shared_mesg.c
+++ b/examples/h5_shared_mesg.c
@@ -109,7 +109,7 @@ main(void)
* This keeps the indexes simple when only a few messages are shared,
* but allows them to scale for many messages. If many messages are
* deleted from the B-tree, it scales back down into a list.
- * A "resonable" setting for maximum list size and minimum btree size
+ * A "reasonable" setting for maximum list size and minimum btree size
* depends on what kinds of messages will be stored in the file.
* These numbers are the same for all indexes in a file.
* We'll guess at some numbers, though we could just as easily have kept
@@ -227,7 +227,7 @@ error:
/*-------------------------------------------------------------------------
* Function: create_standard_file
*
- * Purpose: A helper functon for the example. Creates an HDF5 file
+ * Purpose: A helper function for the example. Creates an HDF5 file
* with many repeated messages using the file creation
* property list FCPL.
*
diff --git a/examples/h5_vds-percival-unlim-maxmin.c b/examples/h5_vds-percival-unlim-maxmin.c
index 7e029a3..f459c20 100644
--- a/examples/h5_vds-percival-unlim-maxmin.c
+++ b/examples/h5_vds-percival-unlim-maxmin.c
@@ -152,7 +152,7 @@ main(void)
wdata[j] = 10 * (i + 1);
/*
- * Open the source files and datasets. Appen data to each dataset and
+ * Open the source files and datasets. Append data to each dataset and
* close all resources.
*/
diff --git a/examples/testh5cc.sh.in b/examples/testh5cc.sh.in
index ab07cb5..e90f94b 100644
--- a/examples/testh5cc.sh.in
+++ b/examples/testh5cc.sh.in
@@ -79,7 +79,7 @@ applib=libapp${H5TOOL}.a
# short hands
# Caution: if some *.h5 files must be cleaned here, list them by names.
# Don't use the wildcard form of *.h5 as it will wipe out even *.h5 generated
-# by otehr test programs. This will cause a racing condition error when
+# by other test programs. This will cause a racing condition error when
# parallel make (e.g., gmake -j 4) is used.
temp_SRC="$hdf5main $v16main $v18main $v110main $v112main $appmain $prog1 $prog2"
temp_OBJ=`echo $temp_SRC | sed -e "s/\.${suffix}/.o/g"`
diff --git a/fortran/examples/compound.f90 b/fortran/examples/compound.f90
index d9f8235..57624c6 100644
--- a/fortran/examples/compound.f90
+++ b/fortran/examples/compound.f90
@@ -41,7 +41,7 @@
INTEGER(HID_T) :: dt3_id ! Memory datatype identifier (for double precision field)
INTEGER(HID_T) :: dt4_id ! Memory datatype identifier (for real field)
INTEGER(HID_T) :: dt5_id ! Memory datatype identifier
- INTEGER(HID_T) :: plist_id ! Dataset trasfer property
+ INTEGER(HID_T) :: plist_id ! Dataset transfer property
INTEGER(SIZE_T) :: typesize
@@ -111,21 +111,21 @@
type_size = type_sizec + type_sizei + type_sized + type_sizer
CALL h5tcreate_f(H5T_COMPOUND_F, type_size, dtype_id, error)
!
- ! Insert memebers
+ ! Insert members
!
- ! CHARACTER*2 memeber
+ ! CHARACTER*2 member
!
offset = 0
CALL h5tinsert_f(dtype_id, "char_field", offset, dt5_id, error)
!
! INTEGER member
!
- offset = offset + type_sizec ! Offset of the second memeber is 2
+ offset = offset + type_sizec ! Offset of the second member is 2
CALL h5tinsert_f(dtype_id, "integer_field", offset, H5T_NATIVE_INTEGER, error)
!
! DOUBLE PRECISION member
!
- offset = offset + type_sizei ! Offset of the third memeber is 6
+ offset = offset + type_sizei ! Offset of the third member is 6
CALL h5tinsert_f(dtype_id, "double_field", offset, H5T_NATIVE_DOUBLE, error)
!
! REAL member
@@ -198,7 +198,7 @@
!
CALL h5dopen_f(file_id, dsetname, dset_id, error)
!
- ! Create memeory datatyoe to read character member of the compound datatype.
+ ! Create memory datatype to read character member of the compound datatype.
!
CALL h5tcopy_f(H5T_NATIVE_CHARACTER, dt2_id, error)
typesize = 2
@@ -208,7 +208,7 @@
offset = 0
CALL h5tinsert_f(dt1_id, "char_field", offset, dt2_id, error)
!
- ! Read part of the datatset and display it.
+ ! Read part of the dataset and display it.
!
CALL h5dread_f(dset_id, dt1_id, char_member_out, data_dims, error)
write(*,*) (char_member_out(i), i=1, dimsize)
diff --git a/fortran/examples/h5_cmprss.f90 b/fortran/examples/h5_cmprss.f90
index 61efc30..86d9cfa 100644
--- a/fortran/examples/h5_cmprss.f90
+++ b/fortran/examples/h5_cmprss.f90
@@ -51,7 +51,7 @@ PROGRAM h5_cmprss
!INTEGER :: szip_pixels_per_block
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
!
diff --git a/fortran/examples/h5_crtatt.f90 b/fortran/examples/h5_crtatt.f90
index 996a506..64d2bde 100644
--- a/fortran/examples/h5_crtatt.f90
+++ b/fortran/examples/h5_crtatt.f90
@@ -35,7 +35,7 @@ PROGRAM H5_CRTATT
INTEGER(HID_T) :: aspace_id ! Attribute Dataspace identifier
INTEGER(HID_T) :: atype_id ! Attribute Dataspace identifier
INTEGER(HSIZE_T), DIMENSION(1) :: adims = (/2/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
INTEGER(SIZE_T) :: attrlen ! Length of the attribute string
CHARACTER(LEN=80), DIMENSION(2) :: attr_data ! Attribute data
diff --git a/fortran/examples/h5_extend.f90 b/fortran/examples/h5_extend.f90
index 0ad6ef9..86cc3c6 100644
--- a/fortran/examples/h5_extend.f90
+++ b/fortran/examples/h5_extend.f90
@@ -86,7 +86,7 @@ PROGRAM H5_EXTEND
INTEGER, DIMENSION(1:3,1:10) :: rdata
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
@@ -110,7 +110,7 @@ PROGRAM H5_EXTEND
CALL h5pset_chunk_f(crp_list, RANK, dimsc, error)
!
- !Create a dataset with 3X3 dimensions using cparms creation propertie .
+ !Create a dataset with 3X3 dimensions using cparms creation properties.
!
CALL h5dcreate_f(file_id, dsetname, H5T_NATIVE_INTEGER, dataspace, &
dset_id, error, crp_list )
diff --git a/fortran/examples/refobjexample.f90 b/fortran/examples/refobjexample.f90
index d420192..873b420 100644
--- a/fortran/examples/refobjexample.f90
+++ b/fortran/examples/refobjexample.f90
@@ -26,7 +26,7 @@
IMPLICIT NONE
CHARACTER(LEN=10), PARAMETER :: filename = "FORTRAN.h5" ! File
CHARACTER(LEN=8), PARAMETER :: dsetnamei = "INTEGERS" ! Dataset with the integer data
- CHARACTER(LEN=17), PARAMETER :: dsetnamer = "OBJECT_REFERENCES" ! Dataset wtih object
+ CHARACTER(LEN=17), PARAMETER :: dsetnamer = "OBJECT_REFERENCES" ! Dataset with object
! references
CHARACTER(LEN=6), PARAMETER :: groupname1 = "GROUP1" ! Groups in the file
CHARACTER(LEN=6), PARAMETER :: groupname2 = "GROUP2" !
diff --git a/fortran/examples/testh5fc.sh.in b/fortran/examples/testh5fc.sh.in
index db687d8..71f41f8 100644
--- a/fortran/examples/testh5fc.sh.in
+++ b/fortran/examples/testh5fc.sh.in
@@ -61,7 +61,7 @@ applib=libapp${H5TOOL}.a
# short hands
# Caution: if some *.h5 files must be cleaned here, list them by names.
# Don't use the wildcard form of *.h5 as it will wipe out even *.h5 generated
-# by otehr test programs. This will cause a racing condition error when
+# by other test programs. This will cause a racing condition error when
# parallel make (e.g., gmake -j 4) is used.
temp_SRC="$hdf5main $appmain $prog1 $prog2"
temp_OBJ=`echo $temp_SRC | sed -e "s/\.${suffix}/.o/g"`
diff --git a/fortran/src/H5Af.c b/fortran/src/H5Af.c
index f78ade2..db1c040 100644
--- a/fortran/src/H5Af.c
+++ b/fortran/src/H5Af.c
@@ -272,7 +272,7 @@ done:
* PURPOSE
* Call H5Aopen to open an attribute
* INPUTS
- * obj_id - Identifer for object to which attribute is attached
+ * obj_id - Identifier for object to which attribute is attached
* attr_name - Attribute access property list
* attr_namelen - size of attr_name
* aapl_id - Link access property list
@@ -317,7 +317,7 @@ done:
* PURPOSE
* Call h5adelete_by_name to remove an attribute from a specified location
* INPUTS
- * loc_id - identifer for object to which attribute is attached
+ * loc_id - identifier for object to which attribute is attached
* obj_name - object identifier
* obj_namelen - name length
* attr_name - name of the attribute
@@ -424,7 +424,7 @@ done:
* INPUTS
*
*
- * loc_id - Identifer for object to which attribute is attached
+ * loc_id - Identifier for object to which attribute is attached
* obj_name - Name of object, relative to location,
* from which attribute is to be removed *TEST* check NULL
* idx_type - Type of index; Possible values are:
@@ -532,7 +532,7 @@ done:
* aapl_id - Attribute access property list
* lapl_id - Link access property list
* OUTPUTS
- * attr_id - attribute identifer
+ * attr_id - attribute identifier
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
diff --git a/fortran/src/H5Aff.F90 b/fortran/src/H5Aff.F90
index 8f56993..53f0a39 100644
--- a/fortran/src/H5Aff.F90
+++ b/fortran/src/H5Aff.F90
@@ -211,7 +211,7 @@ CONTAINS
!
! INPUTS
! obj_id - identifier of a group, dataset, or named
-! datatype atttribute to be attached to
+! datatype attribute to be attached to
! name - attribute name
! OUTPUTS
! attr_id - attribute identifier
@@ -487,7 +487,7 @@ CONTAINS
SUBROUTINE h5aget_name_by_idx_f(loc_id, obj_name, idx_type, order, &
n, name, hdferr, size, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: obj_name ! Name of object, relative to location,
! from which attribute is to be removed *TEST* check NULL
INTEGER, INTENT(IN) :: idx_type ! Type of index; Possible values are:
@@ -846,7 +846,7 @@ CONTAINS
! identifier and attribute name
!
! INPUTS
-! obj_id - Identifer for object to which attribute is attached
+! obj_id - Identifier for object to which attribute is attached
! attr_name - Name of attribute to open
! OUTPUTS
! attr_id - attribute identifier
@@ -934,7 +934,7 @@ CONTAINS
! SOURCE
SUBROUTINE H5Adelete_by_idx_f(loc_id, obj_name, idx_type, order, n, hdferr, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: obj_name ! Name of object, relative to location,
! from which attribute is to be removed
INTEGER, INTENT(IN) :: idx_type ! Type of index; Possible values are:
@@ -991,7 +991,7 @@ CONTAINS
! Removes an attribute from a specified location
!
! INPUTS
-! loc_id - Identifer for object to which attribute is attached
+! loc_id - Identifier for object to which attribute is attached
! obj_name - Name of attribute to open
! attr_name - Attribute access property list
! lapl_id - Link access property list
@@ -1005,7 +1005,7 @@ CONTAINS
! SOURCE
SUBROUTINE H5Adelete_by_name_f(loc_id, obj_name, attr_name, hdferr, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: obj_name ! Name of object, relative to location,
! from which attribute is to be removed
CHARACTER(LEN=*), INTENT(IN) :: attr_name ! Name of attribute to delete
@@ -1841,7 +1841,7 @@ CONTAINS
! dims parameter was added to make code portable;
! Aprile 4, 2001
!
-! Changed buf intent to INOUT to be consistant
+! Changed buf intent to INOUT to be consistent
! with how the C functions handles it. The pg
! compiler will return 0 if a buf value is not set.
! February, 2008
diff --git a/fortran/src/H5Df.c b/fortran/src/H5Df.c
index bcf792c..5c46cb3 100644
--- a/fortran/src/H5Df.c
+++ b/fortran/src/H5Df.c
@@ -145,7 +145,7 @@ DONE:
* Elena Pourmal
* Tuesday, May 14, 2002
* HISTORY
- * This function was added to accomodate h5dwrite_f with the
+ * This function was added to accommodate h5dwrite_f with the
* dims argument being of INTEGER(HSIZE_T) type
* SOURCE
*/
@@ -217,7 +217,7 @@ h5dwrite_ref_reg_c(hid_t_f *dset_id, hid_t_f *mem_type_id, hid_t_f *mem_space_id
* Elena Pourmal
* Wednesday, May 15, 2002
* HISTORY
- * This function was added to accomodate h5dread_f subroutine
+ * This function was added to accommodate h5dread_f subroutine
* with the dims parameter being of INTEGER(HSIZE_T_F) size.
* SOURCE
*/
@@ -783,7 +783,7 @@ h5dwrite_vl_string_c(hid_t_f *dset_id, hid_t_f *mem_type_id, hid_t_f *mem_space_
return ret_value;
}
/*
- * Move data from temorary buffer
+ * Move data from temporary buffer
*/
tmp_p = tmp;
for (i = 0; i < num_elem; i++) {
diff --git a/fortran/src/H5Dff.F90 b/fortran/src/H5Dff.F90
index 75af58e..655cf1c 100644
--- a/fortran/src/H5Dff.F90
+++ b/fortran/src/H5Dff.F90
@@ -476,7 +476,7 @@ CONTAINS
!
! Changed name from the now obsolete h5dextend_f
! to h5dset_extent_f. Provided interface to old name
-! for backward compatability. -MSB- March 14, 2008
+! for backward compatibility. -MSB- March 14, 2008
!
! SOURCE
SUBROUTINE h5dset_extent_f(dataset_id, size, hdferr)
@@ -1195,7 +1195,7 @@ CONTAINS
IMPLICIT NONE
INTEGER(HID_T), INTENT(IN) :: dset_id ! Dataset identifier
INTEGER(HID_T), INTENT(IN) :: mem_type_id ! Memory datatype identifier
- INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the bufffer buf
+ INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the buffer buf
TYPE(hobj_ref_t_f), DIMENSION(dims(1)), INTENT(IN), TARGET :: buf ! Data buffer
INTEGER, INTENT(OUT) :: hdferr ! Error code
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: mem_space_id ! Memory dataspace identfier
@@ -1227,7 +1227,7 @@ CONTAINS
IMPLICIT NONE
INTEGER(HID_T), INTENT(IN) :: dset_id ! Dataset identifier
INTEGER(HID_T), INTENT(IN) :: mem_type_id ! Memory datatype identifier
- INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the bufffer buf
+ INTEGER(HSIZE_T), DIMENSION(*), INTENT(IN) :: dims ! size of the buffer buf
TYPE(hdset_reg_ref_t_f), DIMENSION(dims(1)), INTENT(IN), TARGET :: buf ! Data buffer
INTEGER, INTENT(OUT) :: hdferr ! Error code
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: mem_space_id ! Memory dataspace identfier
diff --git a/fortran/src/H5Fff.F90 b/fortran/src/H5Fff.F90
index a273431..ecb40b7 100644
--- a/fortran/src/H5Fff.F90
+++ b/fortran/src/H5Fff.F90
@@ -97,7 +97,7 @@ CONTAINS
INTEGER(HID_T), INTENT(OUT) :: file_id ! File identifier
INTEGER, INTENT(OUT) :: hdferr ! Error code
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: creation_prp
- ! File creation propertly
+ ! File creation property
! list identifier
INTEGER(HID_T), OPTIONAL, INTENT(IN) :: access_prp
! File access property list
diff --git a/fortran/src/H5Gf.c b/fortran/src/H5Gf.c
index f445fee..0201761 100644
--- a/fortran/src/H5Gf.c
+++ b/fortran/src/H5Gf.c
@@ -390,13 +390,13 @@ DONE:
* Call H5Glink2 to link the specified type
* INPUTS
* cur_loc_id - identifier of file or group
- * cur_name - name of the existing object for hard link releative
+ * cur_name - name of the existing object for hard link relative
* to cur_loc_id location,
* anything for the soft link
* current_namelen - current name length
* link_type - link type
* new_loc_id - location identifier
- * new_name - new name for the object releative to the new_loc_id
+ * new_name - new name for the object relative to the new_loc_id
* location
* new_namelen - new_name length
* RETURNS
@@ -879,7 +879,7 @@ done:
* February 15, 2008
* HISTORY
*
- * - Added 'mounted' paramater
+ * - Added 'mounted' parameter
* M. Scot Breitenfeld
* July 16, 2008
* SOURCE
@@ -1014,7 +1014,7 @@ done:
* February 18, 2008
* HISTORY
*
- * - Added 'mounted' paramater
+ * - Added 'mounted' parameter
* M. Scot Breitenfeld
* July 16, 2008
* SOURCE
diff --git a/fortran/src/H5Gff.F90 b/fortran/src/H5Gff.F90
index 9e7665f..bfca595 100644
--- a/fortran/src/H5Gff.F90
+++ b/fortran/src/H5Gff.F90
@@ -73,7 +73,7 @@ CONTAINS
! called C functions (it is needed for Windows
! port). March 5, 2001
!
-! Added additional optional paramaters in 1.8
+! Added additional optional parameters in 1.8
! MSB - February 27, 2008
!
! SOURCE
@@ -491,8 +491,8 @@ CONTAINS
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: current_namelen ! Lenghth of the current_name string
- INTEGER :: new_namelen ! Lenghth of the new_name string
+ INTEGER :: current_namelen ! Length of the current_name string
+ INTEGER :: new_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5glink_c(loc_id, link_type, current_name, &
@@ -524,7 +524,7 @@ CONTAINS
! PURPOSE
! Creates a link of the specified type from new_name
! to current_name. current_name and new_name are interpreted
-! releative to current and new location identifiers.
+! relative to current and new location identifiers.
!
! INPUTS
! cur_loc_id - location identifier
@@ -559,8 +559,8 @@ CONTAINS
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: cur_namelen ! Lenghth of the current_name string
- INTEGER :: new_namelen ! Lenghth of the new_name string
+ INTEGER :: cur_namelen ! Length of the current_name string
+ INTEGER :: new_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5glink2_c(cur_loc_id, cur_name, cur_namelen, &
@@ -617,7 +617,7 @@ CONTAINS
CHARACTER(LEN=*), INTENT(IN) :: name ! Name of an object
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the name character string
+ INTEGER :: namelen ! Length of the name character string
INTERFACE
INTEGER FUNCTION h5gunlink_c(loc_id, name, namelen) BIND(C,NAME='h5gunlink_c')
@@ -666,8 +666,8 @@ CONTAINS
CHARACTER(LEN=*), INTENT(IN) :: new_name ! New name of an object
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the current_name string
- INTEGER :: new_namelen ! Lenghth of the new_name string
+ INTEGER :: namelen ! Length of the current_name string
+ INTEGER :: new_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5gmove_c(loc_id, name, namelen, new_name, new_namelen) BIND(C,NAME='h5gmove_c')
@@ -715,7 +715,7 @@ CONTAINS
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
INTEGER :: src_namelen ! Length of the current_name string
- INTEGER :: dst_namelen ! Lenghth of the new_name string
+ INTEGER :: dst_namelen ! Length of the new_name string
INTERFACE
INTEGER FUNCTION h5gmove2_c(src_loc_id, src_name, src_namelen, &
@@ -776,7 +776,7 @@ CONTAINS
! points to
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the current_name string
+ INTEGER :: namelen ! Length of the current_name string
INTERFACE
INTEGER FUNCTION h5gget_linkval_c(loc_id, name, namelen, size, buffer) BIND(C,NAME='h5gget_linkval_c')
@@ -827,8 +827,8 @@ CONTAINS
CHARACTER(LEN=*), INTENT(IN) :: comment ! New name of an object
INTEGER, INTENT(OUT) :: hdferr ! Error code
!*****
- INTEGER :: namelen ! Lenghth of the current_name string
- INTEGER :: commentlen ! Lenghth of the comment string
+ INTEGER :: namelen ! Length of the current_name string
+ INTEGER :: commentlen ! Length of the comment string
INTERFACE
INTEGER FUNCTION h5gset_comment_c(loc_id, name, namelen, &
@@ -1024,7 +1024,7 @@ CONTAINS
!
! HISTORY
!
-! - Added 'mounted' paramater
+! - Added 'mounted' parameter
! M. Scot Breitenfeld
! July 16, 2008
!
@@ -1104,7 +1104,7 @@ CONTAINS
! February 18, 2008
!
! HISTORY
-! Added 'mounted' paramater
+! Added 'mounted' parameter
! M. Scot Breitenfeld
! July 16, 2008
!
@@ -1206,7 +1206,7 @@ CONTAINS
! February 18, 2008
!
! HISTORY
-! Added 'mounted' paramater
+! Added 'mounted' parameter
! M. Scot Breitenfeld
! July 16, 2008
! SOURCE
diff --git a/fortran/src/H5Lff.F90 b/fortran/src/H5Lff.F90
index 3bfcf65..a568824 100644
--- a/fortran/src/H5Lff.F90
+++ b/fortran/src/H5Lff.F90
@@ -469,7 +469,7 @@ CONTAINS
! SOURCE
SUBROUTINE h5ldelete_by_idx_f(loc_id, group_name, index_field, order, n, hdferr, lapl_id)
IMPLICIT NONE
- INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifer for object to which attribute is attached
+ INTEGER(HID_T), INTENT(IN) :: loc_id ! Identifier for object to which attribute is attached
CHARACTER(LEN=*), INTENT(IN) :: group_name ! Name of object, relative to location,
! from which attribute is to be removed
INTEGER, INTENT(IN) :: index_field ! Type of index; Possible values are:
@@ -1315,7 +1315,7 @@ CONTAINS
! Inputs:
! group_id - Identifier specifying subject group
! index_type - Type of index which determines the order:
-! H5_INDEX_NAME_F - Alpha-numeric index on name
+! H5_INDEX_NAME_F - Alphanumeric index on name
! H5_INDEX_CRT_ORDER_F - Index on creation order
! order - Order within index:
! H5_ITER_INC_F - Increasing order
@@ -1391,7 +1391,7 @@ CONTAINS
! loc_id - File or group identifier specifying location of subject group
! group_name - Name of subject group
! index_type - Type of index which determines the order:
-! H5_INDEX_NAME_F - Alpha-numeric index on name
+! H5_INDEX_NAME_F - Alphanumeric index on name
! H5_INDEX_CRT_ORDER_F - Index on creation order
! order - Order within index:
! H5_ITER_INC_F - Increasing order
diff --git a/fortran/src/H5Pf.c b/fortran/src/H5Pf.c
index 1f74955..ab09ab9 100644
--- a/fortran/src/H5Pf.c
+++ b/fortran/src/H5Pf.c
@@ -145,7 +145,7 @@ done:
* plist1_id - property list identifier
* plist2_id - property list identifier
* OUTPUTS
- * c_flag - flag to indicate that lists are eqaul
+ * c_flag - flag to indicate that lists are equal
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -936,7 +936,7 @@ h5pset_fapl_stdio_c(hid_t_f *prp_id)
* NAME
* h5pget_fapl_stdio_c
* PURPOSE
- * Call H5Pget_fapl_stdio to determine whther the low level file driver
+ * Call H5Pget_fapl_stdio to determine whether the low level file driver
* uses the functions declared in the stdio.h
* INPUTS
* prp_id - property list identifier
@@ -1013,7 +1013,7 @@ h5pset_fapl_sec2_c(hid_t_f *prp_id)
* NAME
* h5pget_fapl_sec2_c
* PURPOSE
- * Call H5Pget_fapl_stdio to determine whther the low level file driver
+ * Call H5Pget_fapl_stdio to determine whether the low level file driver
* uses the functions declared in the unistd.h
* INPUTS
* prp_id - property list identifier
@@ -2186,7 +2186,7 @@ h5pset_fclose_degree_c(hid_t_f *fapl_id, int_f *degree)
* PURPOSE
* Call H5Pset_buffer to set size of conversion buffer
* INPUTS
- * prp_id - t`dataset trasfer property list identifier
+ * prp_id - t`dataset transfer property list identifier
* size - size of the buffer
* OUTPUTS
* NONE
@@ -2221,7 +2221,7 @@ h5pset_buffer_c(hid_t_f *prp_id, hsize_t_f *size)
* PURPOSE
* Call H5Pget_buffer to get size of conversion buffer
* INPUTS
- * prp_id - t`dataset trasfer property list identifier
+ * prp_id - t`dataset transfer property list identifier
* OUTPUTS
* size - size of conversion buffer
* RETURNS
@@ -3005,11 +3005,11 @@ h5pget_nprops_c(hid_t_f *plist, size_t_f *nprops)
* h5pget_class_parent_c
* PURPOSE
* Call H5Pget_class_parent to get the parent class of
- * a genereic property class
+ * a generic property class
* INPUTS
* prp_id - property list to query
* OUTPUTS
- * parent_id - parent classs identifier
+ * parent_id - parent class identifier
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -3505,7 +3505,7 @@ h5pset_family_offset_c(hid_t_f *prp_id, hsize_t_f *offset)
* NAME
* h5pset_fapl_multi_c
* PURPOSE
- * Call H5Pset_fapl_multi to set multi file dirver
+ * Call H5Pset_fapl_multi to set multi file driver
* INPUTS
* prp_id - file_creation property list identifier
* mem_map - memory mapping array
@@ -3570,7 +3570,7 @@ h5pset_fapl_multi_c(hid_t_f *prp_id, int_f *memb_map, hid_t_f *memb_fapl, _fcd m
tmp_p = tmp_p + c_lenmax;
}
/*
- * Take care of othe arguments
+ * Take care of other arguments
*/
tmp_max_addr = (long double)(HADDR_MAX);
c_prp_id = (hid_t)*prp_id;
@@ -3603,7 +3603,7 @@ DONE:
* NAME
* h5pset_fapl_multi_sc
* PURPOSE
- * Call H5Pset_fapl_multi to set multi file dirver
+ * Call H5Pset_fapl_multi to set multi file driver
* INPUTS
* prp_id - file_creation property list identifier
* RETURNS
@@ -3641,7 +3641,7 @@ h5pset_fapl_multi_sc(hid_t_f *prp_id, int_f *flag)
* NAME
* h5pget_fapl_multi_c
* PURPOSE
- * Call H5Pget_fapl_multi to set multi file dirver
+ * Call H5Pget_fapl_multi to set multi file driver
* INPUTS
* prp_id - file_creation property list identifier
* lenmax - length of the name a sdeclared in Fortran
@@ -4607,7 +4607,7 @@ h5pget_copy_object_c(hid_t_f *ocp_plist_id, int_f *copy_options)
* INPUTS
*
* prp_id - property list identifier to query
- * expression_len - buffer size transorm expression
+ * expression_len - buffer size transform expression
*
* Output:
* expression - buffer to hold transform expression
@@ -4670,7 +4670,7 @@ done:
*
* prp_id - property list identifier to query
* expression - buffer to hold transform expression
- * expression_len - buffer size transorm expression
+ * expression_len - buffer size transform expression
*
* Output:
*
@@ -5478,7 +5478,7 @@ h5pget_fapl_mpio_c(hid_t_f *prp_id, int_f *comm, int_f *info)
* h5pset_dxpl_mpio_c
* PURPOSE
* Call H5Pset_dxpl_mpio to set transfer mode of the dataset
- * trasfer property list
+ * transfer property list
* INPUTS
* prp_id - property list identifier
* data_xfer_mode - transfer mode
@@ -5530,7 +5530,7 @@ h5pset_dxpl_mpio_c(hid_t_f *prp_id, int_f *data_xfer_mode)
* h5pget_dxpl_mpio_c
* PURPOSE
* Call H5Pget_dxpl_mpio to get transfer mode of the dataset
- * trasfer property list
+ * transfer property list
* INPUTS
* prp_id - property list identifier
* data_xfer_mode - buffer to retrieve transfer mode
diff --git a/fortran/src/H5Pff.F90 b/fortran/src/H5Pff.F90
index a7f2366..4831a50 100644
--- a/fortran/src/H5Pff.F90
+++ b/fortran/src/H5Pff.F90
@@ -500,7 +500,7 @@ CONTAINS
! a chunked layout dataset.
!
! INPUTS
-! prp_id - datatset creation property list identifier
+! prp_id - dataset creation property list identifier
! ndims - number of dimensions for each chunk
! dims - array with dimension sizes for each chunk
! OUTPUTS
@@ -1880,7 +1880,7 @@ CONTAINS
! INPUTS
!
! prp_id - file access property list identifier
-! gc_reference - flag for stting garbage collection on
+! gc_reference - flag for string garbage collection on
! and off (1 or 0)
! OUTPUTS
!
@@ -1931,7 +1931,7 @@ CONTAINS
! prp_id - file access property list identifier
! OUTPUTS
!
-! gc_reference - flag for stting garbage collection on
+! gc_reference - flag for string garbage collection on
! and off (1 or 0)
! hdferr - error code
! Success: 0
@@ -2659,7 +2659,7 @@ CONTAINS
! h5pequal_f
!
! PURPOSE
-! Checks if two property lists are eqaul
+! Checks if two property lists are equal
!
! INPUTS
!
@@ -3059,7 +3059,7 @@ CONTAINS
! INPUTS
!
! plist_id - file access property list identifier
-! size - metatdata block size
+! size - metadata block size
! OUTPUTS
!
! hdferr - error code
@@ -3103,7 +3103,7 @@ CONTAINS
! plist_id - file access property list identifier
! OUTPUTS
!
-! size - metatdata block size
+! size - metadata block size
! hdferr - error code
! Success: 0
! Failure: -1
@@ -3596,7 +3596,7 @@ CONTAINS
! h5pget_class_parent_f
!
! PURPOSE
-! Retrieves the parent class of a genric property class.
+! Retrieves the parent class of a generic property class.
!
! INPUTS
!
@@ -5391,7 +5391,7 @@ CONTAINS
!
! HISTORY
!
-! Should hdferr return just 0 or 1 and add another arguement for the size?
+! Should hdferr return just 0 or 1 and add another argument for the size?
! Fortran90 Interface:
SUBROUTINE h5pget_data_transform_f(plist_id, expression, hdferr, size)
IMPLICIT NONE
diff --git a/fortran/src/H5Sf.c b/fortran/src/H5Sf.c
index 341acfd..ea42a31 100644
--- a/fortran/src/H5Sf.c
+++ b/fortran/src/H5Sf.c
@@ -388,7 +388,7 @@ done:
* PURPOSE
* Call H5Sget_select_elem_pointlist
* get a list of element points in the
- * current dataspace selectin.
+ * current dataspace selection.
* Starting with the startpoint-th point in the
* list of points, numpoints points are put into the user's
* buffer. If the user's buffer fills up before numpoints
@@ -759,7 +759,7 @@ done:
* space_id - identifier of the dataspace
* rank - dataspace rank
* current_size - array with the new dimension sizes
- * maximum_size - aray with maximum sizes of dimensions
+ * maximum_size - array with maximum sizes of dimensions
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -804,7 +804,7 @@ done:
* space_id - identifier of the dataspace
* OUTPUTS
* dims - array with the dimension sizes
- * maxdims - aray with maximum sizes of dimensions
+ * maxdims - array with maximum sizes of dimensions
* RETURNS
* number of dataspace dimensions (rank) on success, -1 on failure
* AUTHOR
@@ -852,7 +852,7 @@ done:
* NAME
* h5sis_simple_c
* PURPOSE
- * Call H5Sis_simple to detrmine if the dataspace
+ * Call H5Sis_simple to determine if the dataspace
* is simple.
* INPUTS
* space_id - identifier of the dataspace
diff --git a/fortran/src/H5Sff.F90 b/fortran/src/H5Sff.F90
index 62d665a..76b0dea 100644
--- a/fortran/src/H5Sff.F90
+++ b/fortran/src/H5Sff.F90
@@ -447,7 +447,7 @@ CONTAINS
! INPUTS
! space_id - dataspace identifier
! startpoint - element point to start with
-! num_points - number of elemnt points to get
+! num_points - number of element points to get
! OUTPUTS
! buf - buffer with element points selected
! hdferr - Returns 0 if successful and -1 if fails
diff --git a/fortran/src/H5Tf.c b/fortran/src/H5Tf.c
index 1516cc7..f3016f1 100644
--- a/fortran/src/H5Tf.c
+++ b/fortran/src/H5Tf.c
@@ -1560,7 +1560,7 @@ h5tget_member_type_c(hid_t_f *type_id, int_f *field_idx, hid_t_f *datatype)
* Call H5Tcreate to create a datatype
* INPUTS
* cls - class type
- * size - size of the class memeber
+ * size - size of the class member
* RETURNS
* 0 on success, -1 on failure
* AUTHOR
@@ -2033,7 +2033,7 @@ h5tvlen_create_c(hid_t_f *type_id, hid_t_f *vltype_id)
* NAME
* h5tis_variable_str_c
* PURPOSE
- * Call H5Tis_variable_str to detrmine if the datatype
+ * Call H5Tis_variable_str to determine if the datatype
* is a variable string.
* INPUTS
* type_id - identifier of the dataspace
@@ -2069,7 +2069,7 @@ h5tis_variable_str_c(hid_t_f *type_id, int_f *flag)
* NAME
* h5tget_member_class_c
* PURPOSE
- * Call H5Tget_member_class to detrmine ithe class of the compound
+ * Call H5Tget_member_class to determine ithe class of the compound
* datatype member
* INPUTS
* type_id - identifier of the dataspace
diff --git a/fortran/src/H5Tff.F90 b/fortran/src/H5Tff.F90
index 0eab39e..7582dab 100644
--- a/fortran/src/H5Tff.F90
+++ b/fortran/src/H5Tff.F90
@@ -86,7 +86,7 @@ CONTAINS
! called C functions (it is needed for Windows
! port). March 7, 2001
!
-! Added optional parameter 'tapl_id' for compatability
+! Added optional parameter 'tapl_id' for compatibility
! with H5Topen2. April 9, 2009.
!
! SOURCE
@@ -212,7 +212,7 @@ CONTAINS
! h5tcopy_f
!
! PURPOSE
-! Creates a copy of exisiting datatype.
+! Creates a copy of existing datatype.
!
! INPUTS
! type_id - datatype identifier
diff --git a/fortran/src/H5_buildiface.F90 b/fortran/src/H5_buildiface.F90
index fb3a002..090b6db 100644
--- a/fortran/src/H5_buildiface.F90
+++ b/fortran/src/H5_buildiface.F90
@@ -14,7 +14,7 @@
!
! NOTES
! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF
-! depending on availablity.It generates code that makes use of
+! depending on availability.It generates code that makes use of
! STORAGE_SIZE/SIZEOF in H5_gen.F90. STORAGE_SIZE is standard
! compliant and should always be chosen over SIZEOF.
!
@@ -437,7 +437,7 @@ PROGRAM H5_buildiface
! dims parameter was added to make code portable;
! Aprile 4, 2001
!
-! Changed buf intent to INOUT to be consistant
+! Changed buf intent to INOUT to be consistent
! with how the C functions handles it. The pg
! compiler will return 0 if a buf value is not set.
! February, 2008
@@ -564,7 +564,7 @@ PROGRAM H5_buildiface
! dims parameter was added to make code portable;
! Aprile 4, 2001
!
-! Changed buf intent to INOUT to be consistant
+! Changed buf intent to INOUT to be consistent
! with how the C functions handles it. The pg
! compiler will return 0 if a buf value is not set.
! February, 2008
@@ -677,7 +677,7 @@ PROGRAM H5_buildiface
! Optional parameters:
! mem_space_id - memory dataspace identifier
! file_space_id - file dataspace identifier
-! xfer_prp - trasfer property list identifier
+! xfer_prp - transfer property list identifier
!
! AUTHOR
! Elena Pourmal
diff --git a/fortran/src/H5_ff.F90 b/fortran/src/H5_ff.F90
index ebf888c..39ad84f 100644
--- a/fortran/src/H5_ff.F90
+++ b/fortran/src/H5_ff.F90
@@ -700,7 +700,7 @@ CONTAINS
!
! Outputs:
! majnum - major version of the library
-! minum - minor version of the library
+! minimum - minor version of the library
! relnum - release version of the library
! error - Returns 0 if successful and -1 if fails
!
@@ -735,7 +735,7 @@ CONTAINS
!
! Inputs:
! majnum - major version of the library
-! minum - minor version of the library
+! minimum - minor version of the library
! relnum - release version of the library
!
! Outputs:
diff --git a/fortran/src/H5f90global.F90 b/fortran/src/H5f90global.F90
index 0ed8a17..d202547 100644
--- a/fortran/src/H5f90global.F90
+++ b/fortran/src/H5f90global.F90
@@ -853,7 +853,7 @@ CONTAINS
END SUBROUTINE H5_Fortran_string_f2c
-! Copy Fortran string to C charater array, assuming the C array is one-char
+! Copy Fortran string to C character array, assuming the C array is one-char
! longer for the terminating null char.
! fstring : the Fortran input string
! cstring : the C output string (with memory already allocated)
@@ -881,7 +881,7 @@ CONTAINS
!!$ cstring(j) = C_NULL_CHAR
!!$end subroutine MPIR_Fortran_string_f2c
!!$
-!!$! Copy C charater array to Fortran string
+!!$! Copy C character array to Fortran string
!!$subroutine MPIR_Fortran_string_c2f(cstring, fstring)
!!$ implicit none
!!$ character(kind=c_char), intent(in) :: cstring(:)
diff --git a/fortran/src/H5f90proto.h b/fortran/src/H5f90proto.h
index fbbf7c2..4aa33f6 100644
--- a/fortran/src/H5f90proto.h
+++ b/fortran/src/H5f90proto.h
@@ -540,7 +540,7 @@ H5_FCDLL int_f h5pget_dxpl_mpio_c(hid_t_f *prp_id, int_f *data_xfer_mode);
H5_FCDLL int_f h5pset_dxpl_mpio_c(hid_t_f *prp_id, int_f *data_xfer_mode);
#endif
/*
- * Functions frome H5Rf.c
+ * Functions from H5Rf.c
*/
H5_FCDLL int_f h5rcreate_region_c(int_f *ref, hid_t_f *loc_id, _fcd name, int_f *namelen, hid_t_f *space_id);
H5_FCDLL int_f h5rcreate_ptr_c(void *ref, hid_t_f *loc_id, _fcd name, int_f *namelen, int_f *ref_type,
diff --git a/fortran/src/h5fc.in b/fortran/src/h5fc.in
index a56d38d..661fde5 100644
--- a/fortran/src/h5fc.in
+++ b/fortran/src/h5fc.in
@@ -38,7 +38,7 @@ HL="@HL@"
## $FLINKER $FCFLAGS $H5BLD_FCFLAGS $F9XSUFFIXFLAG $LDFLAGS $LIBS ##
## $fmodules $link_objs $link_args $shared_link ##
## ##
-## These settings can be overriden by setting HDF5_FCFLAGS, ##
+## These settings can be overridden by setting HDF5_FCFLAGS, ##
## HDF5_LDFLAGS, or HDF5_LIBS in the environment. ##
## ##
############################################################################
@@ -136,7 +136,7 @@ usage() {
echo " shared libraries]"
echo " "
echo " You can also add or change paths and flags to the compile line using"
- echo " the following environment varibles or by assigning them to their counterparts"
+ echo " the following environment variables or by assigning them to their counterparts"
echo " in the 'Things You Can Modify to Override...'" section of $prog_name
echo " "
echo " Variable Current value to be replaced"
@@ -298,7 +298,7 @@ fi
if test "x$do_link" = "xyes"; then
shared_link=""
-# conditionnaly link with the hl library
+# conditionally link with the hl library
if test "X$HL" = "Xhl"; then
libraries=" $libraries -lhdf5hl_fortran -lhdf5_hl -lhdf5_fortran -lhdf5 "
else
diff --git a/fortran/test/H5_test_buildiface.F90 b/fortran/test/H5_test_buildiface.F90
index 94beab1..302fa38 100644
--- a/fortran/test/H5_test_buildiface.F90
+++ b/fortran/test/H5_test_buildiface.F90
@@ -14,7 +14,7 @@
!
! NOTES
! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF
-! depending on availablity.It generates code that makes use of
+! depending on availability.It generates code that makes use of
! STORAGE_SIZE/SIZEOF in H5fortran_detect.f90. STORAGE_SIZE is standard
! compliant and should always be chosen over SIZEOF.
!
diff --git a/fortran/test/fflush1.F90 b/fortran/test/fflush1.F90
index c192d93..5684f81 100644
--- a/fortran/test/fflush1.F90
+++ b/fortran/test/fflush1.F90
@@ -88,7 +88,7 @@
data_dims(2) = NY
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
CALL check("h5open_f",error,total_error)
diff --git a/fortran/test/fflush2.F90 b/fortran/test/fflush2.F90
index dc3bc85..a6e16c1 100644
--- a/fortran/test/fflush2.F90
+++ b/fortran/test/fflush2.F90
@@ -81,7 +81,7 @@ PROGRAM FFLUSH2EXAMPLE
data_dims(2) = NY
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
CALL h5open_f(error)
CALL check("h5open_f",error,total_error)
@@ -163,7 +163,7 @@ PROGRAM FFLUSH2EXAMPLE
CALL check("h5fclose_f",error,total_error)
!
- !Close FORTRAN predifined datatypes
+ !Close FORTRAN predefined datatypes
!
CALL h5_cleanup_f(filename, H5P_DEFAULT_F, error)
CALL h5close_f(error)
diff --git a/fortran/test/tH5A.F90 b/fortran/test/tH5A.F90
index d5ce9a2..2e76dad 100644
--- a/fortran/test/tH5A.F90
+++ b/fortran/test/tH5A.F90
@@ -81,7 +81,7 @@ CONTAINS
INTEGER(HID_T) :: atype5_id !Integer Attribute Datatype identifier
INTEGER(HSIZE_T), DIMENSION(1) :: adims = (/2/) ! Attribute dimension
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
INTEGER(SIZE_T) :: attrlen ! Length of the attribute string
INTEGER(HID_T) :: attr_space !Returned String Attribute Space identifier
diff --git a/fortran/test/tH5A_1_8.F90 b/fortran/test/tH5A_1_8.F90
index 4e02c58..eb781a5 100644
--- a/fortran/test/tH5A_1_8.F90
+++ b/fortran/test/tH5A_1_8.F90
@@ -64,7 +64,7 @@ SUBROUTINE attribute_test_1_8(cleanup, total_error)
INTEGER :: ret_total_error
! ********************
-! test_attr equivelent
+! test_attr equivalent
! ********************
! WRITE(*,*) "TESTING ATTRIBUTES"
@@ -840,7 +840,7 @@ SUBROUTINE test_attr_info_by_idx(new_format, fcpl, fapl, total_error)
my_dataset = dset3
END SELECT
- ! Check for query on non-existant attribute
+ ! Check for query on non-existent attribute
n = 0
@@ -1111,7 +1111,7 @@ SUBROUTINE test_attr_shared_rename( fcpl, fapl, total_error)
INTEGER :: test_shared
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
! Initialize "big" attribute data
@@ -1497,7 +1497,7 @@ SUBROUTINE test_attr_delete_by_idx(new_format, fcpl, fapl, total_error)
END SELECT
- ! Check for deleting non-existant attribute
+ ! Check for deleting non-existent attribute
!EP CALL H5Adelete_by_idx_f(my_dataset, '.', idx_type, order, 0_HSIZE_T,error, lapl_id=H5P_DEFAULT_F)
CALL H5Adelete_by_idx_f(my_dataset, '.', idx_type, order, hzero,error, lapl_id=H5P_DEFAULT_F)
CALL verify("H5Adelete_by_idx_f",error,minusone,total_error)
@@ -1773,7 +1773,7 @@ SUBROUTINE test_attr_shared_delete(fcpl, fapl, total_error)
INTEGER :: test_shared
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
! Output message about test being performed
diff --git a/fortran/test/tH5D.F90 b/fortran/test/tH5D.F90
index 006d62a..5f8484e 100644
--- a/fortran/test/tH5D.F90
+++ b/fortran/test/tH5D.F90
@@ -187,7 +187,7 @@ CONTAINS
DO i = 1, 4
DO j = 1, 6
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- WRITE(*, *) "dataset test error occured"
+ WRITE(*, *) "dataset test error occurred"
WRITE(*,*) "data read is not the same as the data written"
END IF
END DO
@@ -196,7 +196,7 @@ CONTAINS
! Check if no change to null_dset_data
!
IF (null_dset_data .NE. 1) THEN
- WRITE(*, *) "null dataset test error occured"
+ WRITE(*, *) "null dataset test error occurred"
END IF
!
! End access to the dataset and release resources used by it.
@@ -258,7 +258,7 @@ CONTAINS
INTEGER(HID_T) :: dset_id ! Dataset identifier
INTEGER(HID_T) :: dataspace ! Dataspace identifier
INTEGER(HID_T) :: memspace ! memory Dataspace identifier
- INTEGER(HID_T) :: crp_list ! dataset creatation property identifier
+ INTEGER(HID_T) :: crp_list ! dataset creation property identifier
!
!dataset dimensions at creation time
@@ -313,7 +313,7 @@ CONTAINS
END DO
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
! CALL h5init_types_f(error)
! CALL check("h5init_types_f",error,total_error)
@@ -347,7 +347,7 @@ CONTAINS
CALL check("h5pset_chunk_f",error,total_error)
!
- !Create a dataset with 3X3 dimensions using cparms creation propertie .
+ !Create a dataset with 3X3 dimensions using cparms creation properties.
!
CALL h5dcreate_f(file_id, dsetname, H5T_NATIVE_INTEGER, dataspace, dset_id, error, crp_list )
CALL check("h5dcreate_f",error,total_error)
@@ -426,7 +426,7 @@ CONTAINS
CALL h5sget_simple_extent_ndims_f(dataspace, rankr, error)
CALL check("h5sget_simple_extent_ndims_f",error,total_error)
IF (rankr .NE. RANK) THEN
- WRITE(*,*) "dataset rank error occured"
+ WRITE(*,*) "dataset rank error occurred"
STOP
END IF
@@ -436,7 +436,7 @@ CONTAINS
CALL h5sget_simple_extent_dims_f(dataspace, dimsr, maxdimsr, error)
CALL check("h5sget_simple_extent_dims_f",error,total_error)
IF ((dimsr(1) .NE. dims1(1)) .OR. (dimsr(2) .NE. dims1(2))) THEN
- WRITE(*,*) "dataset dimensions error occured"
+ WRITE(*,*) "dataset dimensions error occurred"
STOP
END IF
@@ -466,7 +466,7 @@ CONTAINS
DO ih = 1, dims1(1)
DO jh = 1, dims1(2)
IF (data_out(ih,jh) .NE. data_in(ih, jh)) THEN
- WRITE(*, *) "extend dataset test error occured"
+ WRITE(*, *) "extend dataset test error occurred"
WRITE(*, *) "read value is not the same as the written values"
END IF
END DO
diff --git a/fortran/test/tH5F.F90 b/fortran/test/tH5F.F90
index 06dc6de..3affed0 100644
--- a/fortran/test/tH5F.F90
+++ b/fortran/test/tH5F.F90
@@ -554,7 +554,7 @@ CONTAINS
do i = 1, NX
do j = 1, NY
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- write(*, *) "reopen test error occured"
+ write(*, *) "reopen test error occurred"
END IF
end do
end do
diff --git a/fortran/test/tH5G_1_8.F90 b/fortran/test/tH5G_1_8.F90
index 534a2d4..222ba9d 100644
--- a/fortran/test/tH5G_1_8.F90
+++ b/fortran/test/tH5G_1_8.F90
@@ -654,7 +654,7 @@ SUBROUTINE group_info(cleanup, fapl, total_error)
INTEGER(HID_T) :: file, scalar, grp, d1
CHARACTER(LEN=12), PARAMETER :: filename ='TestLinks.h5'
INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension
- INTEGER :: arank = 1 ! Attribure rank
+ INTEGER :: arank = 1 ! Attribute rank
INTEGER :: error
INTEGER :: cset ! Indicates the character set used for the link’s name.
diff --git a/fortran/test/tH5MISC_1_8.F90 b/fortran/test/tH5MISC_1_8.F90
index c294f21..e35ef1f 100644
--- a/fortran/test/tH5MISC_1_8.F90
+++ b/fortran/test/tH5MISC_1_8.F90
@@ -160,7 +160,7 @@ SUBROUTINE test_genprop_basic_class(total_error)
CALL verify("H5Pequal_f", flag, .TRUE., total_error)
- ! Make certain false postives aren't being returned
+ ! Make certain false positives aren't being returned
CALL H5Pequal_f(cid2, H5P_FILE_CREATE_F, flag, error)
CALL check("H5Pequal_f", error, total_error)
CALL verify("H5Pequal_f", flag, .FALSE., total_error)
diff --git a/fortran/test/tH5P.F90 b/fortran/test/tH5P.F90
index 1d1208c..75e4e72 100644
--- a/fortran/test/tH5P.F90
+++ b/fortran/test/tH5P.F90
@@ -349,8 +349,8 @@ SUBROUTINE multi_file_test(cleanup, total_error)
DO i = 1, 4
DO j = 1, 6
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- WRITE(*, *) "dataset test error occured"
- WRITE(*,*) "data read is not the same as the data writen"
+ WRITE(*, *) "dataset test error occurred"
+ WRITE(*,*) "data read is not the same as the data written"
END IF
END DO
END DO
@@ -575,7 +575,7 @@ SUBROUTINE test_chunk_cache(cleanup, total_error)
CALL H5Pclose_f(dapl2,error)
CALL check("H5Pclose_f", error, total_error)
- ! Similary, test use of H5Dcreate2 with H5P_DEFAULT
+ ! Similarly, test use of H5Dcreate2 with H5P_DEFAULT
CALL H5Dclose_f(dsid, error)
CALL check("H5Dclose_f", error, total_error)
diff --git a/fortran/test/tH5P_F03.F90 b/fortran/test/tH5P_F03.F90
index c9b23f1..744f3ca 100644
--- a/fortran/test/tH5P_F03.F90
+++ b/fortran/test/tH5P_F03.F90
@@ -1006,28 +1006,28 @@ SUBROUTINE test_vds(total_error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(nsize.NE.LEN(SRC_FILE_LEN_EXACT))THEN
- PRINT*,"virtual filenname size is incorrect"
+ PRINT*,"virtual filename size is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that is very small
CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_TINY, error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_TINY.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_TINY)))THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that small by one
CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_SMALL, error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_SMALL.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_SMALL)))THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that is exact
CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_EXACT, error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_EXACT.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)))THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that bigger by one
@@ -1035,7 +1035,7 @@ SUBROUTINE test_vds(total_error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_LARGE(1:LEN(SRC_FILE_LEN_EXACT)).NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)).AND. &
SRC_FILE_LEN_LARGE(LEN(SRC_FILE_LEN_EXACT):).NE.'')THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! check passing a buffer that is very big
@@ -1043,7 +1043,7 @@ SUBROUTINE test_vds(total_error)
CALL check("H5Pget_virtual_filename_f", error, total_error)
IF(SRC_FILE_LEN_HUGE(1:LEN(SRC_FILE_LEN_EXACT)).NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)).AND. &
SRC_FILE_LEN_HUGE(LEN(SRC_FILE_LEN_EXACT):).NE.'')THEN
- PRINT*,"virtual filenname returned is incorrect"
+ PRINT*,"virtual filename returned is incorrect"
total_error = total_error + 1
ENDIF
! Get source dataset name
diff --git a/fortran/test/tH5R.F90 b/fortran/test/tH5R.F90
index 5144d83..4e62b20 100644
--- a/fortran/test/tH5R.F90
+++ b/fortran/test/tH5R.F90
@@ -142,7 +142,7 @@ SUBROUTINE refobjtest(cleanup, total_error)
CALL check("h5gclose_f",error,total_error)
!
- ! Craete references to two groups, integer dataset and shared datatype
+ ! Create references to two groups, integer dataset and shared datatype
! and write it to the dataset in the file
!
CALL h5rcreate_f(file_id, groupname1, ref(1), error)
diff --git a/fortran/test/tH5S.F90 b/fortran/test/tH5S.F90
index 08f0b59..302db51 100644
--- a/fortran/test/tH5S.F90
+++ b/fortran/test/tH5S.F90
@@ -151,7 +151,7 @@ CONTAINS
CALL h5sget_simple_extent_dims_f(space2_id, dimsout, maxdimsout, error)
CALL check("h5sget_simple_extent_dims_f", error, total_error)
IF ((dimsout(1) .NE. dims1(1)) .OR. (dimsout(2) .NE. dims1(2)) ) THEN
- write(*,*)"error occured, copied dims not same"
+ write(*,*)"error occurred, copied dims not same"
END IF
!
@@ -159,14 +159,14 @@ CONTAINS
!
CALL h5sget_simple_extent_ndims_f(space2_id, rank2, error)
CALL check("h5sget_simple_extent_ndims_f", error, total_error)
- IF (rank2 .NE. rank1) write(*,*)"error occured, copied ranks not same"
+ IF (rank2 .NE. rank1) write(*,*)"error occurred, copied ranks not same"
!
!get the copied space's number of elements.
!
CALL h5sget_simple_extent_npoints_f(space2_id, npoints, error)
CALL check("h5sget_simple_extent_npoints_f", error, total_error)
- IF (npoints .NE. 24) write(*,*)"error occured, number of elements not correct"
+ IF (npoints .NE. 24) write(*,*)"error occurred, number of elements not correct"
!
@@ -194,7 +194,7 @@ CONTAINS
CALL h5sget_simple_extent_dims_f(space2_id, dimsout, maxdimsout, error)
CALL check("h5sget_simple_extent_dims_f", error, total_error)
IF ((dimsout(1) .NE. dims2(1)) .OR. (dimsout(2) .NE. dims2(2)) ) THEN
- write(*,*)"error occured, copied dims not same"
+ write(*,*)"error occurred, copied dims not same"
END IF
!
@@ -235,8 +235,8 @@ CONTAINS
do i = 1, 4
do j = 1, 6
IF (data1_out(i,j) .NE. data1_in(i, j)) THEN
- write(*, *) "dataset test error occured"
- write(*,*) "data read is not the same as the data writen"
+ write(*, *) "dataset test error occurred"
+ write(*,*) "data read is not the same as the data written"
END IF
end do
end do
@@ -256,8 +256,8 @@ CONTAINS
do i = 1, 6
do j = 1, 6
IF (data2_out(i,j) .NE. data2_in(i, j)) THEN
- write(*, *) "dataset test error occured"
- write(*,*) "data read is not the same as the data writen"
+ write(*, *) "dataset test error occurred"
+ write(*,*) "data read is not the same as the data written"
END IF
end do
end do
diff --git a/fortran/test/tH5Sselect.F90 b/fortran/test/tH5Sselect.F90
index 178ac32..41cd16f 100644
--- a/fortran/test/tH5Sselect.F90
+++ b/fortran/test/tH5Sselect.F90
@@ -149,7 +149,7 @@ CONTAINS
!
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
! CALL h5init_types_f(error)
! CALL check("h5init_types_f", error, total_error)
@@ -430,7 +430,7 @@ CONTAINS
end do
!
- !Initialize FORTRAN predifined datatypes
+ !Initialize FORTRAN predefined datatypes
!
! CALL h5init_types_f(error)
! CALL check("h5init_types_f", error, total_error)
@@ -912,7 +912,7 @@ CONTAINS
!
CALL h5sget_select_hyper_nblocks_f(dataspace, num_blocks, error)
CALL check("h5sget_select_hyper_nblocks_f", error, total_error)
- IF (num_blocks .NE. 4) write (*,*) "error occured with num_blocks"
+ IF (num_blocks .NE. 4) write (*,*) "error occurred with num_blocks"
!write(*,*) num_blocks
!result of num_blocks is 4
@@ -945,11 +945,11 @@ CONTAINS
CALL h5sget_select_bounds_f(dataspace, startout, endout, error)
CALL check("h5sget_select_bounds_f", error, total_error)
IF ( (startout(1) .ne. 1) .or. (startout(2) .ne. 1) ) THEN
- write(*,*) "error occured to select_bounds's start position"
+ write(*,*) "error occurred to select_bounds's start position"
END IF
IF ( (endout(1) .ne. 5) .or. (endout(2) .ne. 5) ) THEN
- write(*,*) "error occured to select_bounds's end position"
+ write(*,*) "error occurred to select_bounds's end position"
END IF
!write(*,*) (startout(i), i = 1, RANK)
!result of startout is 0, 0
@@ -978,7 +978,7 @@ CONTAINS
!
CALL h5sget_select_elem_npoints_f(dataspace, num_points, error)
CALL check("h5sget_select_elem_npoints_f", error, total_error)
- IF (num_points .NE. 10) write(*,*) "error occured with num_points"
+ IF (num_points .NE. 10) write(*,*) "error occurred with num_points"
!write(*,*) num_points
! result of num_points is 10
diff --git a/fortran/test/tH5T.F90 b/fortran/test/tH5T.F90
index 8b10136..d24ac89 100644
--- a/fortran/test/tH5T.F90
+++ b/fortran/test/tH5T.F90
@@ -72,7 +72,7 @@ CONTAINS
INTEGER(HID_T) :: dt4_id ! Memory datatype identifier (for real field)
INTEGER(HID_T) :: dt5_id ! Memory datatype identifier
INTEGER(HID_T) :: membtype_id ! Datatype identifier
- INTEGER(HID_T) :: plist_id ! Dataset trasfer property
+ INTEGER(HID_T) :: plist_id ! Dataset transfer property
INTEGER(HSIZE_T), DIMENSION(1) :: dims = (/dimsize/) ! Dataset dimensions
@@ -98,7 +98,7 @@ CONTAINS
INTEGER :: class ! Datatype class
INTEGER :: num_members ! Number of members in the compound datatype
CHARACTER(LEN=256) :: member_name
- INTEGER :: len ! Lenght of the name of the compound datatype member
+ INTEGER :: len ! Length of the name of the compound datatype member
INTEGER :: member_index ! index of the field
INTEGER(HSIZE_T), DIMENSION(3) :: array_dims=(/2,3,4/)
INTEGER :: array_dims_range = 3
@@ -219,9 +219,9 @@ CONTAINS
CALL h5tcreate_f(H5T_COMPOUND_F, type_size, dtype_id, error)
CALL check("h5tcreate_f", error, total_error)
!
- ! Insert memebers
+ ! Insert members
!
- ! CHARACTER*2 memeber
+ ! CHARACTER*2 member
!
offset = 0
CALL h5tinsert_f(dtype_id, "char_field", offset, dt5_id, error)
@@ -229,13 +229,13 @@ CONTAINS
!
! INTEGER member
!
- offset = offset + type_sizec ! Offset of the second memeber is 2
+ offset = offset + type_sizec ! Offset of the second member is 2
CALL h5tinsert_f(dtype_id, "integer_field", offset, H5T_NATIVE_INTEGER, error)
CALL check("h5tinsert_f", error, total_error)
!
! DOUBLE PRECISION member
!
- offset = offset + type_sizei ! Offset of the third memeber is 6
+ offset = offset + type_sizei ! Offset of the third member is 6
CALL h5tinsert_f(dtype_id, "double_field", offset, H5T_NATIVE_DOUBLE, error)
CALL check("h5tinsert_f", error, total_error)
!
@@ -416,7 +416,7 @@ CONTAINS
CALL h5tequal_f(membtype_id, H5T_NATIVE_INTEGER, flag, error)
CALL check("h5tequal_f", error, total_error)
if(.not. flag) then
- write(*,*) "Wrong member type returned for integer memebr"
+ write(*,*) "Wrong member type returned for integer member"
total_error = total_error + 1
endif
CALL h5tget_member_class_f(dtype_id, i-1, class, error)
@@ -435,7 +435,7 @@ CONTAINS
CALL h5tequal_f(membtype_id, H5T_NATIVE_DOUBLE, flag, error)
CALL check("h5tequal_f", error, total_error)
if(.not. flag) then
- write(*,*) "Wrong member type returned for double precision memebr"
+ write(*,*) "Wrong member type returned for double precision member"
total_error = total_error + 1
endif
CALL h5tget_member_class_f(dtype_id, i-1, class, error)
@@ -454,7 +454,7 @@ CONTAINS
CALL h5tequal_f(membtype_id, H5T_NATIVE_REAL, flag, error)
CALL check("h5tequal_f", error, total_error)
if(.not. flag) then
- write(*,*) "Wrong member type returned for real memebr"
+ write(*,*) "Wrong member type returned for real member"
total_error = total_error + 1
endif
CALL h5tget_member_class_f(dtype_id, i-1, class, error)
diff --git a/fortran/test/tH5VL.F90 b/fortran/test/tH5VL.F90
index 2269fec..eb0470c 100644
--- a/fortran/test/tH5VL.F90
+++ b/fortran/test/tH5VL.F90
@@ -147,7 +147,7 @@ CONTAINS
CALL check("h5dvlen_get_max_len_f", error, total_error)
if(max_len .ne. data_dims(1)) then
total_error = total_error + 1
- write(*,*) "Wrong number of elemets returned by h5dvlen_get_max_len_f"
+ write(*,*) "Wrong number of elements returned by h5dvlen_get_max_len_f"
endif
!
! Read the dataset.
@@ -317,7 +317,7 @@ CONTAINS
CALL check("h5dvlen_get_max_len_f", error, total_error)
if(max_len .ne. data_dims(1)) then
total_error = total_error + 1
- write(*,*) "Wrong number of elemets returned by h5dvlen_get_max_len_f"
+ write(*,*) "Wrong number of elements returned by h5dvlen_get_max_len_f"
endif
!
! Read the dataset.
diff --git a/fortran/test/tH5Z.F90 b/fortran/test/tH5Z.F90
index 8c39fea..799067a 100644
--- a/fortran/test/tH5Z.F90
+++ b/fortran/test/tH5Z.F90
@@ -374,7 +374,7 @@ CONTAINS
do i = 1, N
do j = 1, M
IF (data_out(i,j) .NE. dset_data(i, j)) THEN
- write(*, *) "dataset test error occured"
+ write(*, *) "dataset test error occurred"
write(*,*) "data read is not the same as the data written"
num_errors = num_errors + 1
IF (num_errors .GE. 512) THEN
diff --git a/hl/c++/test/ptableTest.cpp b/hl/c++/test/ptableTest.cpp
index 9616c1b..651f057 100644
--- a/hl/c++/test/ptableTest.cpp
+++ b/hl/c++/test/ptableTest.cpp
@@ -223,7 +223,7 @@ TestGetNext()
for (record = 1; record < 6; record++)
wrapper.AppendPacket(&record);
- /* Ensure that we can interate through the records and get the right ones */
+ /* Ensure that we can iterate through the records and get the right ones */
for (i = 1; i < 6; i++) {
wrapper.GetNextPacket(&record);
if (record != i)
@@ -237,7 +237,7 @@ TestGetNext()
if (error < 0)
goto error;
- /* Ensure that we can interate through the records and get the right ones */
+ /* Ensure that we can iterate through the records and get the right ones */
for (i = 1; i < 6; i++) {
error = wrapper.GetNextPacket(&record);
if (record != i || error < 0)
diff --git a/hl/examples/ex_image1.c b/hl/examples/ex_image1.c
index ead1715..1294ab3 100644
--- a/hl/examples/ex_image1.c
+++ b/hl/examples/ex_image1.c
@@ -56,10 +56,10 @@ main(void)
H5IMmake_image_8bit(file_id, "image1", (hsize_t)WIDTH, (hsize_t)HEIGHT, buf);
/* make a palette */
- H5IMmake_palette(file_id, "pallete", pal_dims, pal);
+ H5IMmake_palette(file_id, "palette", pal_dims, pal);
/* attach the palette to the image */
- H5IMlink_palette(file_id, "image1", "pallete");
+ H5IMlink_palette(file_id, "image1", "palette");
/* close the file. */
H5Fclose(file_id);
diff --git a/hl/fortran/src/H5DSff.F90 b/hl/fortran/src/H5DSff.F90
index b241e9a..bbd9918 100644
--- a/hl/fortran/src/H5DSff.F90
+++ b/hl/fortran/src/H5DSff.F90
@@ -216,7 +216,7 @@ CONTAINS
END SUBROUTINE H5DSis_attached_f
!
-! H5DSiterate_scales: Impliment in F2003
+! H5DSiterate_scales: Implement in F2003
!
!-------------------------------------------------------------------------
diff --git a/hl/fortran/src/H5HL_buildiface.F90 b/hl/fortran/src/H5HL_buildiface.F90
index 4f34575..ca66ebf 100644
--- a/hl/fortran/src/H5HL_buildiface.F90
+++ b/hl/fortran/src/H5HL_buildiface.F90
@@ -14,7 +14,7 @@
!
! NOTES
! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF
-! depending on availablity.It generates code that makes use of
+! depending on availability.It generates code that makes use of
! STORAGE_SIZE/SIZEOF in H5fortran_detect.f90. STORAGE_SIZE is standard
! compliant and should always be chosen over SIZEOF.
!
diff --git a/hl/fortran/src/H5LTff.F90 b/hl/fortran/src/H5LTff.F90
index ba4b770..a4ab247 100644
--- a/hl/fortran/src/H5LTff.F90
+++ b/hl/fortran/src/H5LTff.F90
@@ -88,7 +88,7 @@ MODULE H5LT_CONST
TYPE(C_PTR), VALUE :: buf ! data buffer
CHARACTER(KIND=C_CHAR), DIMENSION(*), INTENT(in) :: dtype ! flag indicating the datatype of the
! the buffer:
- ! R=Real, D=DOUBLE, I=Interger, C=Character
+ ! R=Real, D=DOUBLE, I=Integer, C=Character
INTEGER(size_t) :: SizeOf_buf ! Sizeof the buf datatype
END FUNCTION h5ltset_attribute_c
END INTERFACE
@@ -107,7 +107,7 @@ MODULE H5LT_CONST
TYPE(C_PTR), VALUE :: buf ! data buffer
CHARACTER(KIND=C_CHAR), DIMENSION(*), INTENT(in) :: dtype ! flag indicating the datatype of the
! the buffer:
- ! R=Real, D=DOUBLE, I=Interger
+ ! R=Real, D=DOUBLE, I=Integer
INTEGER(size_t), INTENT(in) :: SizeOf_buf ! Sizeof the buf data type
END FUNCTION h5ltget_attribute_c
END INTERFACE
diff --git a/hl/fortran/src/H5TBfc.c b/hl/fortran/src/H5TBfc.c
index 23e3377..91c420b 100644
--- a/hl/fortran/src/H5TBfc.c
+++ b/hl/fortran/src/H5TBfc.c
@@ -83,7 +83,7 @@ h5tbmake_table_c(size_t_f *namelen1, _fcd name1, hid_t_f *loc_id, size_t_f *name
(tmp = (char *)HD5f2cstring(field_names, (size_t) * (max_char_size_field_names) * (size_t)num_elem)))
HGOTO_DONE(FAIL)
/*
- * move data from temorary buffer
+ * move data from temporary buffer
*/
tmp_p = tmp;
for (i = 0; i < num_elem; i++) {
@@ -188,7 +188,7 @@ h5tbmake_table_ptr_c(size_t_f *namelen1, _fcd name1, hid_t_f *loc_id, size_t_f *
(tmp = (char *)HD5f2cstring(field_names, (size_t) * (max_char_size_field_names) * (size_t)num_elem)))
HGOTO_DONE(FAIL)
/*
- * move data from temorary buffer
+ * move data from temporary buffer
*/
tmp_p = tmp;
for (i = 0; i < num_elem; i++) {
diff --git a/hl/fortran/src/H5TBff.F90 b/hl/fortran/src/H5TBff.F90
index 35a88b8..82f34a8 100644
--- a/hl/fortran/src/H5TBff.F90
+++ b/hl/fortran/src/H5TBff.F90
@@ -961,7 +961,7 @@ CONTAINS
INTEGER(size_t), DIMENSION(1:nfields), INTENT(inout) :: field_offsets ! field offsets
INTEGER(size_t), INTENT(inout):: type_size ! type size
INTEGER(size_t) :: namelen ! name length
- INTEGER(size_t) :: maxlen ! maxiumum length of input field names
+ INTEGER(size_t) :: maxlen ! maximum length of input field names
INTEGER(size_t), DIMENSION(1:nfields) :: namelen2 ! name lengths
INTEGER(size_t) :: c_maxlen_out ! maximum character length of a field array element
END FUNCTION h5tbget_field_info_c
diff --git a/hl/fortran/test/tsttable.F90 b/hl/fortran/test/tsttable.F90
index a77068c..e7bd88e 100644
--- a/hl/fortran/test/tsttable.F90
+++ b/hl/fortran/test/tsttable.F90
@@ -172,11 +172,11 @@ SUBROUTINE test_table1()
!
offset = 0
field_offset(1) = offset
- offset = offset + type_sizec ! Offset of the second memeber is 2
+ offset = offset + type_sizec ! Offset of the second member is 2
field_offset(2) = offset
- offset = offset + type_sizei ! Offset of the second memeber is 6
+ offset = offset + type_sizei ! Offset of the second member is 6
field_offset(3) = offset
- offset = offset + type_sized ! Offset of the second memeber is 14
+ offset = offset + type_sized ! Offset of the second member is 14
field_offset(4) = offset
!-------------------------------------------------------------------------
@@ -571,7 +571,7 @@ SUBROUTINE test_table2()
INTEGER(HSIZE_T), PARAMETER :: nrecords = 8 ! nrecords
CHARACTER(len=8), PARAMETER :: filename = "f2tab.h5" ! File name
- CHARACTER(LEN=5), PARAMETER :: table_name = "tabel" ! table name
+ CHARACTER(LEN=5), PARAMETER :: table_name = "table" ! table name
CHARACTER(LEN=10), PARAMETER :: table_name_fill = "tabel_fill" ! table name
! Define field information
diff --git a/hl/src/H5DO.c b/hl/src/H5DO.c
index 3bdb4e2..1703669 100644
--- a/hl/src/H5DO.c
+++ b/hl/src/H5DO.c
@@ -182,7 +182,7 @@ H5DOappend(hid_t dset_id, hid_t dxpl_id, unsigned axis, size_t extension, hid_t
if (FAIL == H5Sselect_hyperslab(new_space_id, H5S_SELECT_SET, start, stride, count, block))
goto done;
- /* The # of elemnts in the new extended dataspace */
+ /* The # of elements in the new extended dataspace */
if ((snelmts = H5Sget_select_npoints(new_space_id)) < 0)
goto done;
nelmts = (hsize_t)snelmts;
diff --git a/hl/src/H5DS.c b/hl/src/H5DS.c
index c947d16..387419b 100644
--- a/hl/src/H5DS.c
+++ b/hl/src/H5DS.c
@@ -88,7 +88,7 @@ H5DSset_scale(hid_t dsid, const char *dimname)
return FAIL;
/*-------------------------------------------------------------------------
- * check if the dataset is a dataset wich has references to dimension scales
+ * check if the dataset is a dataset which has references to dimension scales
*-------------------------------------------------------------------------
*/
@@ -1122,7 +1122,7 @@ H5DSdetach_scale(hid_t did, hid_t dsid, unsigned int idx)
goto out;
} /* nelmts */
- /* Free refrences */
+ /* Free references */
if (is_new_ref) {
if (H5Treclaim(tid, sid, H5P_DEFAULT, ndsbuf) < 0)
goto out;
@@ -1189,7 +1189,7 @@ out:
dsbuf = NULL;
}
if (buf) {
- /* Failure occured before H5Treclaim was called;
+ /* Failure occurred before H5Treclaim was called;
free the pointers allocated when we read data in */
for (i = 0; i < rank; i++) {
if (buf[i].p)
@@ -1633,7 +1633,7 @@ H5DSiterate_scales(hid_t did, unsigned int dim, int *ds_idx, H5DS_iterate_t visi
if (H5DSwith_new_ref(did, &is_new_ref) < 0)
return FAIL;
- /* get the number of scales assotiated with this DIM */
+ /* get the number of scales associated with this DIM */
if ((nscales = H5DSget_num_scales(did, dim)) < 0)
return FAIL;
diff --git a/hl/test/test_ds.c b/hl/test/test_ds.c
index c66ef49..12b6840 100644
--- a/hl/test/test_ds.c
+++ b/hl/test/test_ds.c
@@ -3021,7 +3021,7 @@ test_simple(void)
}
/*-------------------------------------------------------------------------
- * dettach for DIM0
+ * detach for DIM0
*-------------------------------------------------------------------------
*/
@@ -3319,11 +3319,11 @@ test_simple(void)
goto out;
dim = 0;
- /* iterate trough the 1st dimension of "dset_a" and verify that its DS is valid */
+ /* iterate through the 1st dimension of "dset_a" and verify that its DS is valid */
if (H5DSiterate_scales(did, dim, NULL, verify_scale, NULL) < 0)
goto out;
- /* iterate trough the 2nd dimension of "dset_a" and verify that its DS is valid
+ /* iterate through the 2nd dimension of "dset_a" and verify that its DS is valid
start at DS index 2 */
dim = 1;
scale_idx = 2;
@@ -3344,11 +3344,11 @@ test_simple(void)
goto out;
dim = 0;
- /* iterate trough the 1st dimension of "dset_a" and read the DS */
+ /* iterate through the 1st dimension of "dset_a" and read the DS */
if (H5DSiterate_scales(did, dim, NULL, read_scale, s1_wbuf) < 0)
goto out;
- /* iterate trough the 2nd dimension of "dset_a" and read the DS
+ /* iterate through the 2nd dimension of "dset_a" and read the DS
start at DS index 2 */
dim = 1;
scale_idx = 2;
@@ -3381,9 +3381,9 @@ test_simple(void)
goto out;
{
int match_size; /* does this scale size matches the dataset DIM size */
- int idx = 0; /* scale index to start iterating, on return, index where iterator stoped */
+ int idx = 0; /* scale index to start iterating, on return, index where iterator stopped */
- /* iterate trough all the dimensions */
+ /* iterate through all the dimensions */
for (dim = 0; dim < (unsigned)rank; dim++) {
if ((match_size = H5DSiterate_scales(did, dim, &idx, match_dim_scale, NULL)) < 0)
goto out;
@@ -3475,9 +3475,9 @@ test_simple(void)
goto out;
{
int match_size; /* does this scale size matches the dataset DIM size */
- int idx; /* scale index to start iterating, on return, index where iterator stoped */
+ int idx; /* scale index to start iterating, on return, index where iterator stopped */
- /* iterate trough all the dimensions */
+ /* iterate through all the dimensions */
for (dim = 0; dim < (unsigned)rank; dim++) {
/* always start at 1st scale */
idx = 0;
@@ -4025,7 +4025,7 @@ test_errors(void)
PASSED();
/*-------------------------------------------------------------------------
- * try to attach a scale to an image, pallete or table
+ * try to attach a scale to an image, palette or table
*-------------------------------------------------------------------------
*/
@@ -4036,7 +4036,7 @@ test_errors(void)
goto out;
/* make a palette */
- if (H5IMmake_palette(fid, "pallete", pal_dims, NULL) < 0)
+ if (H5IMmake_palette(fid, "palette", pal_dims, NULL) < 0)
goto out;
/* open the previous written "ds_b" */
@@ -4248,7 +4248,7 @@ test_iterators(void)
if ((did = H5Dopen2(fid, "dset_a", H5P_DEFAULT)) < 0)
goto out;
- /* try to iterate trough the 1st dimension of "dset_a", return error */
+ /* try to iterate through the 1st dimension of "dset_a", return error */
if (H5DSiterate_scales(did, 0, NULL, verify_scale, NULL) < 0)
goto out;
@@ -4269,7 +4269,7 @@ test_iterators(void)
if ((did = H5Dopen2(fid, "dset_a", H5P_DEFAULT)) < 0)
goto out;
- /* try to iterate trough the 3rd dimension of "dset_a", return error */
+ /* try to iterate through the 3rd dimension of "dset_a", return error */
if (H5DSiterate_scales(did, 3, NULL, verify_scale, NULL) == SUCCEED)
goto out;
@@ -4306,7 +4306,7 @@ test_iterators(void)
goto out;
}
- /* iterate trough the 1st dimension of "dset_a" */
+ /* iterate through the 1st dimension of "dset_a" */
if (H5DSiterate_scales(did, 0, NULL, op_continue, NULL) < 0)
goto out;
diff --git a/hl/test/test_file_image.c b/hl/test/test_file_image.c
index 36d4db3..4448e41 100644
--- a/hl/test/test_file_image.c
+++ b/hl/test/test_file_image.c
@@ -244,7 +244,7 @@ test_file_image(size_t open_images, size_t nflags, const unsigned *flags)
*/
HDmemset((uint8_t *)tmp_ptr + SUPER_STATUS_FLAGS_OFF_V0_V1, (int)0,
(size_t)SUPER_STATUS_FLAGS_SIZE_V0_V1);
- /* Does the comparision */
+ /* Does the comparison */
if (HDmemcmp(tmp_ptr, buf_ptr[i], (size_t)buf_size[i]) != 0)
FAIL_PUTS_ERROR("comparison of TMP vfd and user buffer failed");
/* Free the temporary buffer */
diff --git a/hl/test/test_h5do_compat.c b/hl/test/test_h5do_compat.c
index c554ae7..6f19d4b 100644
--- a/hl/test/test_h5do_compat.c
+++ b/hl/test/test_h5do_compat.c
@@ -31,7 +31,7 @@
*
* Purpose: Test the basic functionality of H5DOwrite_chunk
*
- * Return: Success: An identifer for the dataset used in the tests
+ * Return: Success: An identifier for the dataset used in the tests
* Failure: H5I_INVALID_HID
*
*-------------------------------------------------------------------------
diff --git a/hl/test/test_image.c b/hl/test/test_image.c
index e54875e..4997df6 100644
--- a/hl/test/test_image.c
+++ b/hl/test/test_image.c
@@ -219,7 +219,7 @@ test_simple(void)
*-------------------------------------------------------------------------
*/
- HL_TESTING2("pallete functions");
+ HL_TESTING2("palette functions");
if (H5IMget_npalettes(fid, IMAGE1_NAME, &npals) < 0)
goto out;
@@ -422,7 +422,7 @@ test_data(void)
/*-------------------------------------------------------------------------
* palette #4. blue-red
- * make a palette whith blue to red colors
+ * make a palette with blue to red colors
*-------------------------------------------------------------------------
*/
for (i = 0, n = 0; i < 256 * 3; i += 3, n++) {
diff --git a/hl/test/test_ld.c b/hl/test/test_ld.c
index dc5755c..150d04c 100644
--- a/hl/test/test_ld.c
+++ b/hl/test/test_ld.c
@@ -189,7 +189,7 @@ typedef struct set_t {
* This will fail on heiwa and amani when VALID_FIELDS1 is "field1,field3,field4"
* because of alignment problems:
* amani and heiwa - 8 byte alignment
- * jam - 4 byte alignemnt
+ * jam - 4 byte alignment
* This will need to be fixed in the library for H5Tget_native_type().
*/
/* VALID_FIELDS1 "field1,field2.a,field3,field4" */
@@ -213,8 +213,8 @@ typedef struct test_valid_fields2 {
#define TEST_BUF_SIZE 100
int * iibuf; /* buffer for storing retrieved elements */
int * ibuf; /* buffer for storing retrieved elements (integer) */
-set_t * cbuf; /* buffer for storing retrieved elemnets (compound) */
-set_t * ccbuf; /* buffer for storing retrieved elemnets (compound) */
+set_t * cbuf; /* buffer for storing retrieved elements (compound) */
+set_t * ccbuf; /* buffer for storing retrieved elements (compound) */
test_valid_fields1 *vbuf1; /* buffer for storing retrieved elements (FIELDS1) */
test_valid_fields2 *vbuf2; /* buffer for storing retrieved elements (FIELDS2) */
diff --git a/hl/test/test_packet_vlen.c b/hl/test/test_packet_vlen.c
index c33479e..aa88a6c 100644
--- a/hl/test/test_packet_vlen.c
+++ b/hl/test/test_packet_vlen.c
@@ -677,7 +677,7 @@ error: /* An error has occurred. Clean up and exit. */
*
* Description:
* - Added a fixed-length packet table to the file for variety
- * - Use the helper funtion verify_ptlengthtype to test H5PTis_varlen
+ * - Use the helper function verify_ptlengthtype to test H5PTis_varlen
* on each packet table.
*
* 2016/01/27 -BMR
diff --git a/hl/test/test_table.c b/hl/test/test_table.c
index 6199e27..90a930c 100644
--- a/hl/test/test_table.c
+++ b/hl/test/test_table.c
@@ -1464,7 +1464,7 @@ test_table(hid_t fid, int do_write)
HL_TESTING2("getting field info");
- /* alocate */
+ /* allocate */
names_out = (char **)HDmalloc(sizeof(char *) * (size_t)NFIELDS);
for (i = 0; i < NFIELDS; i++) {
names_out[i] = (char *)HDmalloc(sizeof(char) * 255);
diff --git a/hl/tools/gif2h5/decompress.c b/hl/tools/gif2h5/decompress.c
index 64d9643..ca33f86 100644
--- a/hl/tools/gif2h5/decompress.c
+++ b/hl/tools/gif2h5/decompress.c
@@ -178,7 +178,7 @@ Decompress(GIFIMAGEDESC *GifImageDesc, GIFHEAD *GifHead)
DataMask = (1 << ((GifHead->PackedField & 0x07) + 1)) - 1;
Raster = GifImageDesc->GIFImage;
- /* Check for image seperator */
+ /* Check for image separator */
/* Now read in values from the image descriptor */
IWidth = GifImageDesc->ImageWidth;
@@ -193,7 +193,7 @@ Decompress(GIFIMAGEDESC *GifImageDesc, GIFHEAD *GifHead)
*/
/*
- * Start reading the raster data. First we get the intial code size and
+ * Start reading the raster data. First we get the initial code size and
* compute decompressor constant values, based on this code size.
*/
diff --git a/hl/tools/gif2h5/gif2mem.c b/hl/tools/gif2h5/gif2mem.c
index c66250b..2c2225e 100644
--- a/hl/tools/gif2h5/gif2mem.c
+++ b/hl/tools/gif2h5/gif2mem.c
@@ -40,7 +40,7 @@ Gif2Mem(GIFBYTE *MemGif, GIFTOMEM *GifMemoryStruct)
{
/*
* The gif structure outline for passing data to memory is given in gif.h.
- * These pointers are redunant, should take them out in ver. 2
+ * These pointers are redundant, should take them out in ver. 2
*/
GIFHEAD * gifHead; /* GIF Header structure */
GIFIMAGEDESC ** gifImageDesc; /* Logical Image Descriptor struct */
diff --git a/hl/tools/gif2h5/gifread.c b/hl/tools/gif2h5/gifread.c
index dd9e5be..3a6bcae 100644
--- a/hl/tools/gif2h5/gifread.c
+++ b/hl/tools/gif2h5/gifread.c
@@ -50,15 +50,15 @@ GetByte(const GIFBYTE *MemGif)
* Global Color Table (if any) from a GIF image file. The information
* is stored in a GIFHEAD structure.
*
- * Returns: -1 if a FILE stream error occured during the read,
- * otherwise 0 if no error occured.
+ * Returns: -1 if a FILE stream error occurred during the read,
+ * otherwise 0 if no error occurred.
*/
int
ReadGifHeader(GIFHEAD * GifHead, /* Pointer to GIF header structure */
GIFBYTE **MemGif2) /* GIF image file input FILE stream */
{
GIFWORD i; /* Loop counter */
- GIFWORD tableSize; /* Number of entires in the Global Color Table */
+ GIFWORD tableSize; /* Number of entries in the Global Color Table */
GifHead->TableSize = 0;
for (i = 0; i < 6; i++) {
@@ -103,7 +103,7 @@ ReadGifHeader(GIFHEAD * GifHead, /* Pointer to GIF header structure */
return -1;
#endif /* 0 */
- return 0; /* No FILE stream error occured */
+ return 0; /* No FILE stream error occurred */
}
/*
@@ -116,8 +116,8 @@ ReadGifHeader(GIFHEAD * GifHead, /* Pointer to GIF header structure */
** Note that the ImageSeparator field value in the GIFIMAGEDESC
** structure is assigned by the function calling ReadGifImageDesc().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifImageDesc(GIFIMAGEDESC *GifImageDesc, /* Pointer to GIF image descriptor structure */
@@ -196,7 +196,7 @@ ReadGifImageDesc(GIFIMAGEDESC *GifImageDesc, /* Pointer to GIF image descriptor
*TempPtr++ = *(*MemGif2)++;
} while (ch1);
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -205,8 +205,8 @@ ReadGifImageDesc(GIFIMAGEDESC *GifImageDesc, /* Pointer to GIF image descriptor
** Note that the Introducer and Label field values in the GIFGRAPHICCONTROL
** structure are assigned by the function calling ReadGifGraphicControl().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifGraphicControl(GIFGRAPHICCONTROL *GifGraphicControl, /* Pointer to GC Extension structure */
@@ -219,7 +219,7 @@ ReadGifGraphicControl(GIFGRAPHICCONTROL *GifGraphicControl, /* Pointer to GC Ext
GifGraphicControl->GCEDump[i] = *(*MemGif2)++;
}
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -228,8 +228,8 @@ ReadGifGraphicControl(GIFGRAPHICCONTROL *GifGraphicControl, /* Pointer to GC Ext
** Note that the Introducer and Label field values in the GIFLPLAINTEXT
** structure are assigned by the function calling ReadGifPlainText().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifPlainText(GIFPLAINTEXT *GifPlainText, /* Pointer to Plain Text Extension structure */
@@ -256,7 +256,7 @@ ReadGifPlainText(GIFPLAINTEXT *GifPlainText, /* Pointer to Plain Text Extension
return(-1);
*/
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -265,8 +265,8 @@ ReadGifPlainText(GIFPLAINTEXT *GifPlainText, /* Pointer to Plain Text Extension
** Note that the Introducer and Label field values in the GIFAPPLICATION
** structure are assigned by the function calling ReadGifApplication().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifApplication(GIFAPPLICATION *GifApplication, /* Pointer to Application Extension structure */
@@ -292,7 +292,7 @@ ReadGifApplication(GIFAPPLICATION *GifApplication, /* Pointer to Application Ext
return(-1);
*/
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -301,8 +301,8 @@ ReadGifApplication(GIFAPPLICATION *GifApplication, /* Pointer to Application Ext
** Note that the Introducer and Label field values in the GIFCOMMENT
** structure are assigned by the function calling ReadGifComment().
**
-** Returns: -1 if a FILE stream error occured during the read,
-** otherwise 0 if no error occured.
+** Returns: -1 if a FILE stream error occurred during the read,
+** otherwise 0 if no error occurred.
*/
int
ReadGifComment(GIFCOMMENT *GifComment, /* Pointer to GIF Comment Extension structure */
@@ -316,7 +316,7 @@ ReadGifComment(GIFCOMMENT *GifComment, /* Pointer to GIF Comment Extension struc
GifComment->Terminator = 0;
- return (0); /* No FILE stream error occured */
+ return (0); /* No FILE stream error occurred */
}
/*
@@ -326,8 +326,8 @@ ReadGifComment(GIFCOMMENT *GifComment, /* Pointer to GIF Comment Extension struc
** A GIF "sub-block" is a single count byte followed by 1 to 255
** additional data bytes.
**
-** Returns: A NULL pointer if a memory allocation error occured,
-** otherwise a valid pointer if no error occured.
+** Returns: A NULL pointer if a memory allocation error occurred,
+** otherwise a valid pointer if no error occurred.
*/
static GIFBYTE *
ReadDataSubBlocks(GIFBYTE **MemGif2, /* GIF image file input FILE stream */
@@ -364,7 +364,7 @@ ReadDataSubBlocks(GIFBYTE **MemGif2, /* GIF image file input FILE stream
if ((dataSize = *(*MemGif2)++) == 0)
break; /* Block Terminator encountered */
- /* Increase the buffer size to accomodate the next sub-block */
+ /* Increase the buffer size to accommodate the next sub-block */
if (!(ptr1 = ptr2 = (GIFBYTE *)realloc(ptr2, bufSize + dataSize + 1)))
return ((GIFBYTE *)NULL);
diff --git a/hl/tools/gif2h5/h52giftest.sh.in b/hl/tools/gif2h5/h52giftest.sh.in
index 2f84ef6..4f4057f 100644
--- a/hl/tools/gif2h5/h52giftest.sh.in
+++ b/hl/tools/gif2h5/h52giftest.sh.in
@@ -46,7 +46,7 @@ TOOLTEST()
# Verify the test runs with failure (return code is not 0)
# Use for testing if tool can handle error conditions like
-# illegal input, bad arguments, exeeding limits, ...
+# illegal input, bad arguments, exceeding limits, ...
TOOLTESTFAIL()
{
# for now, discard any error messages generated.
diff --git a/hl/tools/gif2h5/writehdf.c b/hl/tools/gif2h5/writehdf.c
index a444cf7..b49e2a6 100644
--- a/hl/tools/gif2h5/writehdf.c
+++ b/hl/tools/gif2h5/writehdf.c
@@ -27,7 +27,7 @@
* Programmer: Unknown
*
* Modifications: pvn
- * Use the HDF5 IMAGE API to write the HDF5 image and pallete
+ * Use the HDF5 IMAGE API to write the HDF5 image and palette
*
* Date: January, 31, 2006
*
@@ -68,7 +68,7 @@ WriteHDF(GIFTOMEM GifMemoryStruct, char *HDFName)
#endif /* UNUSED */
if ((file_id = H5Fcreate(HDFName, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT)) < 0) {
- /* error occured opening the HDF File for write */
+ /* error occurred opening the HDF File for write */
fprintf(stderr, "HDF file could not be opened for writing\n");
fprintf(stderr,
"NOTE: GIF file must be present in the same directory as the binary on UNIX systems.\n");
diff --git a/hl/tools/h5watch/extend_dset.c b/hl/tools/h5watch/extend_dset.c
index 07a4053..7756506 100644
--- a/hl/tools/h5watch/extend_dset.c
+++ b/hl/tools/h5watch/extend_dset.c
@@ -83,7 +83,7 @@ extend_dset_two(const char *file, char *dname, int action1, int action2)
size_t dtype_size; /* size of the dataset's datatype */
unsigned num_elmts; /* number of elements in the dataset */
int * ibuf = NULL; /* buffer for storing retrieved elements (integer) */
- set_t * cbuf = NULL; /* buffer for storing retrieved elemnets (compound) */
+ set_t * cbuf = NULL; /* buffer for storing retrieved elements (compound) */
/* Allocate memory */
if (NULL == (ibuf = (int *)HDcalloc(TEST_BUF_SIZE, sizeof(int))))
@@ -237,7 +237,7 @@ extend_dset_one(const char *file, char *dname, int action)
hsize_t count[1]; /* dimension sizes of appended data */
size_t dtype_size; /* size of the dataset's datatype */
int * ibuf = NULL; /* buffer for storing retrieved elements (integer) */
- set_t * cbuf = NULL; /* buffer for storing retrieved elemnets (compound) */
+ set_t * cbuf = NULL; /* buffer for storing retrieved elements (compound) */
/* Allocate memory */
if (NULL == (ibuf = (int *)HDcalloc(TEST_BUF_SIZE, sizeof(int))))
diff --git a/m4/aclocal_cxx.m4 b/m4/aclocal_cxx.m4
index 2779066..188f7ca 100644
--- a/m4/aclocal_cxx.m4
+++ b/m4/aclocal_cxx.m4
@@ -20,7 +20,7 @@ dnl Contains Macros for HDF5 C++
dnl *********************************
dnl
dnl Special characteristics that have no autoconf counterpart but that
-dnl we need as part of the C++ support. To distinquish these, they
+dnl we need as part of the C++ support. To distinguish these, they
dnl have a [PAC] prefix.
dnl Checking if C++ needs old style header files in includes
diff --git a/release_docs/HISTORY-1_0-1_8_0_rc3.txt b/release_docs/HISTORY-1_0-1_8_0_rc3.txt
index f23a2c5..071e646 100644
--- a/release_docs/HISTORY-1_0-1_8_0_rc3.txt
+++ b/release_docs/HISTORY-1_0-1_8_0_rc3.txt
@@ -103,9 +103,9 @@ New Features
-JML 2006/6/21
- If both shared and static libraries are installed, now both will be
tested during 'make install'. -MAM 2006/06/21
- - Added support to explicity enable stream_vfd or shared libraries
+ - Added support to explicitly enable stream_vfd or shared libraries
when using parallel via the '--enable-stream_vfd' and
- '--enable-shared' options, respectively. If not explicity defined,
+ '--enable-shared' options, respectively. If not explicitly defined,
These settings default to enabled when parallel is not used,
and disabled when parallel is used. -MAM 2006/06/17
- Remove the flexible parallel code and the --enable-fphdf5
@@ -169,7 +169,7 @@ New Features
variable $FC, not $F9X. F9X still works, but is depreciated.
The output of make may be different. This should be only a
cosmetic effect.
- make depened (or make dep) is no longer recognized, since automake
+ make depend (or make dep) is no longer recognized, since automake
handles dependency tracking.
Some new configure options exist. --enable-dependency-tracking
and --disable-dependency-tracking are used to control automake's
@@ -306,11 +306,11 @@ New Features
- Added managements of collective IO supports for chunking storage
inside parallel HDF5
1) Implemented One IO with collective mode for all chunks in the
- application by building one MPI derived datatype accross all
+ application by building one MPI derived datatype across all
chunks.
2) Implemented the decision-making support to do collective IO inside
MPI-IO per chunk.
- 3) Added the decision-making support to do one IO accross all chunks
+ 3) Added the decision-making support to do one IO across all chunks
or to do multiple IOs with each IO per chunk.
4) Added the support to handle the case some processes won't do any IOs in
collectively.
@@ -319,7 +319,7 @@ New Features
contributions to IOs, a special macro is added to change
collective IO mode to independent IO mode inside HDF5 library.
- Currently we find that MPICH at Linux and vender MPI-IO package at NCSA
+ Currently we find that MPICH at Linux and vendor MPI-IO package at NCSA
Altix cannot handle this case.
"hdf5_mpi_special_collective_io_works=${hdf5_mpi_special_collective_io_works='no'}"
@@ -469,7 +469,7 @@ New Features
object's object header. QAK 2003/10/06
- Added new H5Fget_freespace() routine to query the free space in a
given file. QAK 2003/10/06
- - Added backward compatability with v1.6 for new Error API. SLU -
+ - Added backward compatibility with v1.6 for new Error API. SLU -
2003/09/24
- Changed 'objno' field in H5G_stat_t structure from 'unsigned long[2]'
to 'haddr_t'. QAK - 2003/08/08
@@ -496,7 +496,7 @@ New Features
- Add options of using atomicity and file-sync to test_mpio_1wMr.
AKC - 2003/11/13
- Added parallel test, test_mpio_1wMr, which tests if the
- underlaying parallel I/O system is conforming to the POSIX
+ underlying parallel I/O system is conforming to the POSIX
write/read requirement. AKC - 2003/11/12
Fortran Library:
@@ -606,7 +606,7 @@ New Features
and H5LTdtype_to_text. H5LTtext_to_dtype creates a HDF5 data type
given a text description; H5LTdtype_to_text converts a data type
to text description. Only DDL definition is supported as text
- desciption now. SLU - 2006/05/17
+ description now. SLU - 2006/05/17
- Added Packet Table API for creating tables with less overhead than
H5TB API. Added C++ wrapper for Packet Tables. See documentation.
JML - 2004/03/28
@@ -639,7 +639,7 @@ Support for new platforms, languages and compilers.
- Added support for HPUX11.23 for both 32 and 64-bit; HDF5 C++
- Added support for 64-bit Windows with Visual Studio .NET and 2005.
SJW - 2007/06/25
- - Added suport for HPUX11.23 for both 32 and 64-bit; HDF5 C++
+ - Added support for HPUX11.23 for both 32 and 64-bit; HDF5 C++
shared library is not supported with +DD64 flag.
EIP - 2006/06/22
- Added support for VAX floating numbers for Alpha Open VMS 7.3.2
@@ -655,7 +655,7 @@ Support for new platforms, languages and compilers.
- HDF5 Fortran APIs are supported on Mac OSX with IBM XL Fortran
compiler version 8.1. This is a default compiler.
- HDF5 Fortran APIs are supported on MAC OSX with Absoft F95 compiler
- version 8.2; set F9X environment varibale to f95, for example
+ version 8.2; set F9X environment variable to f95, for example
setenv F9X f95
Use --disable-shared --enable-static configure flags when Absoft
compiler is used.
@@ -700,7 +700,7 @@ Bug Fixes since HDF5-1.6.0 release
- Re-open the file
- Delete one of the attributes on the object
- Add a smaller attribute to the object
- - Delete the smaller atttribute on the object
+ - Delete the smaller attribute on the object
- Add a larger attribute on the object
After this, the number of header messages stored for the object
@@ -855,7 +855,7 @@ Bug Fixes since HDF5-1.6.0 release
- Fixed dtypes "sw long double -> double" failure in QSC class
machines. AKC - 2004/4/16
- Fixed problem with fletcher32 filter when converting data of different
- endianess. PVN - 2004/03/10
+ endianness. PVN - 2004/03/10
- Fixed problem with H5Tget_native_type() not handling opaque fields
correctly. QAK - 2004/01/31
- Fixed several errors in B-tree deletion code which could cause a
@@ -1271,7 +1271,7 @@ Known Problems
sometimes. This can be verified with t_mpi test under testpar.
* On IRIX6.5, when C compiler version >7.4, the complicate MPI derived data type
code will work. However, the user should be aware to enlarge MPI_TYPE_MAX environment
- variable to some certian value in order to use collective irregular selection code.
+ variable to some certain value in order to use collective irregular selection code.
For example, the current parallel HDF5 test needs to enlarge MPI_TYPE_MAX to
200,000 to make the test pass.
* The dataset created or rewritten with the v1.6.3 library or after can't
@@ -2098,7 +2098,7 @@ Bug Fixes since HDF5-1.6.5 Release
- Re-open the file.
- Delete one of the attributes on the object.
- Add a smaller attribute to the object.
- - Delete the smaller atttribute on the object.
+ - Delete the smaller attribute on the object.
- Add a larger attribute on the object.
After this, the number of header messages stored for the object
@@ -2399,11 +2399,11 @@ Known Problems
For 32-bit:
You may do this, under csh:
- setenv LD_LIBRARYN32_PATH .......(existing pathes):[full path of HDF5
+ setenv LD_LIBRARYN32_PATH .......(existing paths):[full path of HDF5
directory/test/.libs]:[full path of HDF5 directory/src/.libs]
For 64-bit:
- setenv LD_LIBRARY64_PATH ......(existing pathes):[full path of HDF5
+ setenv LD_LIBRARY64_PATH ......(existing paths):[full path of HDF5
directory/test/.libs]:[full path of HDF5 directory/src/.libs]
NOTE: This problem ONLY affects the testing of the HDF5 library when you
@@ -2723,7 +2723,7 @@ Bug Fixes since HDF5-1.6.4 Release
-------
- Fixed collective IO in chunking-storage. HDF5 may have called the
wrong routine when the shape of the dataspace in the file and in
- the bufferred chunk were different. This bug was fixed to make sure
+ the buffered chunk were different. This bug was fixed to make sure
the correct routine is called. KY - 2005/10/19
- Fixed core dump when closing root groups opened through two different
file handles that operate on the same actual file. QAK - 2005/10/02
@@ -2744,7 +2744,7 @@ Bug Fixes since HDF5-1.6.4 Release
together all have the same "file close degree". QAK - 2005/07/19
- More bug fixes on holding open files that are mounted and have
IDs open. QAK - 2005/07/14
- - Dataset sieve buffer cache was inadvertantly disabled; it has been
+ - Dataset sieve buffer cache was inadvertently disabled; it has been
re-enabled. QAK - 2005/07/08
- Don't unmount child files until the parent file actually closes.
(Previously, if an object was holding open a file, the child files
@@ -3024,7 +3024,7 @@ Known Problems
to intel forum.
* When testing parallel HDF5 with the C compiler version MIPSpro 7.4.3 at IRIX
- 6.5, set enviroment variable MPI_TYPE_MAX to be a bigger number, for example
+ 6.5, set environment variable MPI_TYPE_MAX to be a bigger number, for example
120000, in order to pass the complicated collective IO tests inside parallel
HDF5 library. This is not a problem inside parallel HDF5 library. You can
always set a bigger number in your system.
@@ -4448,7 +4448,7 @@ New Features
verbosity via the '-v' option. The default is less verbose
than before. AKC - 2004/01/23
- Added parallel test, test_mpio_1wMr, which tests if the
- underlaying parallel I/O system is conforming to the POSIX
+ underlying parallel I/O system is conforming to the POSIX
write/read requirement. This version includes options of using
atomicity and file-sync. AKC - 2003/11/27
@@ -4821,7 +4821,7 @@ Known Problems
Testing random sw long double -> double conversions
If -ieee is used, the converted doubles spread over the range 0.0 to 1**-308.
If -ieee is not used, the converted double values are mostly 0.0 but
- occassionaly as 1**-308. This has been reported to the system staff.
+ occasionally as 1**-308. This has been reported to the system staff.
All other tests have passed.
* On AIX 5.1 when 64-bit parallel C Library is built with zlib configured in,
@@ -5187,7 +5187,7 @@ Linux 2.4 IA64 Intel y n y n y y y
Known Problems
==============
* Fortran subroutine h5pget_driver_f doesn't return a correct driver information.
- The fix willl be available in the 1.6.2 release.
+ The fix will be available in the 1.6.2 release.
* There are two h5dump xml tests(h5dump --xml thlink.h5 and h5dump --xml tmany.h5)
failed on windows xp with .NET for debug and debug dll. Release and Release
dll work fine.
@@ -5368,7 +5368,7 @@ Library:
* The first version of szip compression support were implemented.
User should have static szlib library installed. Using function
H5Pset_szip to pass the szip parameters to the HDF5 library.
- More detailed decription of the process will be followed.
+ More detailed description of the process will be followed.
KY-2003/04/01
* Added Fletcher32 checksum as a filter in pipeline. It only works in
chunked dataset. SLU - 2003/2/11
@@ -5416,15 +5416,15 @@ Library:
be in different locations in the same file. The old functions H5Glink
and H5Gmove remain valid. SLU - 2002/04/26
* Fill-value's behaviors for contiguous dataset have been redefined.
- Basicly, dataset won't allocate space until it's necessary. Full details
+ Basically, dataset won't allocate space until it's necessary. Full details
are available at http://hdf.ncsa.uiuc.edu/RFC/Fill_Value, at this moment.
SLU - 2002/04/11
* Added new routine "H5Dfill" to fill a selection with a particular value
in memory. QAK - 2002/04/09
* Improved performance of "regular" hyperslab I/O when using MPI-IO and the
- datatype conversion is unneccessary. QAK - 2002/04/02
+ datatype conversion is unnecessary. QAK - 2002/04/02
* Improved performance of single hyperslab I/O when datatype conversion is
- unneccessary. QAK - 2002/04/02
+ unnecessary. QAK - 2002/04/02
* Added new "H5Sget_select_type" API function to determine which type of
selection is defined for a dataspace ("all", "none", "hyperslab" or
"point"). QAK - 2002/02/07
@@ -5438,9 +5438,9 @@ Library:
compatibility is enabled.
* Added a new file access property, file close degree, to control file
close behavior. It has four values, H5F_CLOSE_WEAK, H5F_CLOSE_SEMI,
- H5F_CLOSE_STRONG, and H5F_CLOSE_DEFAULT. Two correspont functions
+ H5F_CLOSE_STRONG, and H5F_CLOSE_DEFAULT. Two corresponding functions
H5Pset_fclose_degree and H5Pget_fclose_degree are also provided. Two
- new functions H5Fget_obj_count and H5Fget_obj_ids are offerted to assist
+ new functions H5Fget_obj_count and H5Fget_obj_ids are offered to assist
this new feature. For full details, please refer to the reference
manual under the description of H5Fcreate, H5Fopen, H5Fclose and the
functions mentioned above.
@@ -5458,7 +5458,7 @@ Library:
* Improved regular hyperslab I/O by about a factor of 6 or so.
* Modified the Pablo build procedure to permit building of the instrumented
library to link either with the Trace libraries as before or with the
- Pablo Performance Caputure Facility.
+ Pablo Performance Capture Facility.
* Added new F90 APIs for generic properties, new filters, and
time/space allocation properties.
* C++ API:
@@ -5536,7 +5536,7 @@ Library
* Fixed data corruption problem which could occur when fill values were
written to a contiguously stored dataset in parallel. QAK - 2002/08/27
* Fixed VL memory leak when data is overwritten. The heap objects holding
- old data are freed. If the fill value writting time is set to
+ old data are freed. If the fill value writing time is set to
H5D_FILL_TIME_NEVER, the library prohibits user to create VL type dataset.
The library free all the heap objects storing VL type if there is nested
VL type(a VL type contains another VL type). SLU - 2002/07/10
@@ -6175,7 +6175,7 @@ Known Problems
compiler optimization causes errors in many HDF5 Library tests. Use -g -h zero flags
to build HDF5 Library.
- * On Cray SV1 10.0.1. 0 datatype convertion test fails. Please check HDF FTP site
+ * On Cray SV1 10.0.1.0 datatype conversion test fails. Please check HDF FTP site
if patch is available. We will try to provide one in the nearest future.
* For configuration, building and testing with Intel and PGI compilers see
@@ -6932,7 +6932,7 @@ Known Problems
HDF5_MPI_OPT_TYPES to a non-zero value such as 1.
* On IA64 systems one has to use -DIA64 compilation flag to compile
- h4toh5 and h5toh4 utilites. After configuration step manually modify
+ h4toh5 and h5toh4 utilities. After configuration step manually modify
Makefile in the tools/h4toh4 and tools/h5toh4 directories to add
-DIA64 to the compilation flags.
@@ -6998,7 +6998,7 @@ New Features
and other version information in the source code.
* Parallel HDF5 now runs on the HP V2500 and HP N4000 machines.
* F90 API:
- - Added aditional parameter "dims" to the h5dread_f/h5dwrite_f and
+ - Added additional parameter "dims" to the h5dread_f/h5dwrite_f and
h5aread_f/h5awrite_f subroutines. This parameter is a 1-D array
of size 7 and contains the sizes of the data buffer dimensions.
This change enables portability between Windows and UNIX platforms.
@@ -7012,8 +7012,8 @@ New Features
* F90 static library is available on Windows platforms.
See INSTALL_Windows_withF90.txt for details.
* F90 APIs are available on HPUX 11.00 and 10.20 and IBM SP platforms.
- * H5 <-> GIF convertor has been added. This is available under
- tools/gifconv. The convertor supports the ability to create animated
+ * H5 <-> GIF converter has been added. This is available under
+ tools/gifconv. The converter supports the ability to create animated
gifs as well.
* Verified correct operation of library on Solaris 2.8 in both 64-bit and
32-bit compilation modes. See INSTALL document for instructions on
@@ -7064,7 +7064,7 @@ Bug Fixes since HDF5-1.4.1 Release
when --enable-hdf5v1_2 configure flag is enabled.
* Added --enable-linux-lfs flag to allow more control over whether to
enable or disable large file support on Linux.
- * Fixed various bugs releated to SDS dimensional scale conversions in H4->H5
+ * Fixed various bugs related to SDS dimensional scale conversions in H4->H5
converter.
* Fixed a bug to correctly convert HDF4 objects with fill value into HDF5.
* Fixed a bug of H5pubconf.h causing repeated definitions if it is included
@@ -7345,7 +7345,7 @@ Bug Fixes since HDF5-1.4.0 Release
causing errors during I/O
* The RCSID string in H5public.h was causing the C++ compiling problem
because when it was included multiple times, C++ did not like
- multiple definitions of the same static variable. All occurance of
+ multiple definitions of the same static variable. All occurrence of
RCSID definition are removed since we have not used it consistently
before.
@@ -8111,7 +8111,7 @@ Known Problems
o SunOS 5.6 with C WorkShop Compilers 4.2: hyperslab selections will
fail if library is compiled using optimization of any level.
o TFLOPS: dsets test fails if compiled with optimization turned on.
-o J90: tools fail to dispay data for the datasets with a compound datatype.
+o J90: tools fail to display data for the datasets with a compound datatype.
Platforms Tested
================
@@ -8379,7 +8379,7 @@ C. Changes Since Version 1.0.1
* New query functions for selections.
* I/O operations bypass the stripmining loop and go directly to
- storage for certain contiguous selections in the absense of type
+ storage for certain contiguous selections in the absence of type
conversions. In other cases the stripmining buffers are used more
effectively.
diff --git a/release_docs/HISTORY-1_10.txt b/release_docs/HISTORY-1_10.txt
index 127b821..4796ebb 100644
--- a/release_docs/HISTORY-1_10.txt
+++ b/release_docs/HISTORY-1_10.txt
@@ -161,7 +161,7 @@ New Features
YES NO Default to STATIC
NO YES Default to SHARED
NO NO Default to SHARED
- The defaults can be overriden by setting the config option
+ The defaults can be overridden by setting the config option
HDF5_INSTALL_MOD_FORTRAN to one of NO, SHARED, or STATIC
(ADB - 2020/07/09, HDFFV-11116)
@@ -610,7 +610,7 @@ Bug Fixes since HDF5-1.10.6 release
- h5diff fixed a command line parsing error.
h5diff would ignore the argument to -d (delta) if it is smaller than DBL_EPSILON.
- The macro H5_DBL_ABS_EQUAL was removed and a direct value comparision was used.
+ The macro H5_DBL_ABS_EQUAL was removed and a direct value comparison was used.
(ADB - 2020/07/20, HDFFV-10897)
@@ -618,7 +618,7 @@ Bug Fixes since HDF5-1.10.6 release
h5diff would ignore all objects with a supplied path if the exclude-path argument is used.
Adding the exclude-attribute argument will only eclude attributes, with the supplied path,
- from comparision.
+ from comparison.
(ADB - 2020/07/20, HDFFV-5935)
@@ -1434,7 +1434,7 @@ Known Problems
and thus does not suffer from the 32 bit size limitation described here.
OpenMPI releases prior to v3.1.3 appear to have other datatype issues however,
e.g. within a single parallel test (testphdf5) the subtests (cdsetr, eidsetr)
- report data verfication errors before eventually aborting.
+ report data verification errors before eventually aborting.
The most recent versions of OpenMPI (v3.1.3 or newer) have evidently
resolved these isses and parallel HDF5 testing does not currently report
errors though occasional hangs have been observed.
@@ -1636,7 +1636,7 @@ New Features
(DER - 2019/02/03, HDFFV-10696)
- - Add ability to minimze dataset object headers.
+ - Add ability to minimize dataset object headers.
Creation of many, very small datasets resulted in extensive file bloat
due to extra space in the dataset object headers -- this space is
@@ -2137,7 +2137,7 @@ Known Problems
and thus does not suffer from the 32 bit size limitation described here.
OpenMPI releases prior to v3.1.3 appear to have other datatype issues however,
e.g. within a single parallel test (testphdf5) the subtests (cdsetr, eidsetr)
- report data verfication errors before eventually aborting.
+ report data verification errors before eventually aborting.
The most recent versions of OpenMPI (v3.1.3 or newer) have evidently
resolved these isses and parallel HDF5 testing does not currently report
errors though occasional hangs have been observed.
@@ -2865,7 +2865,7 @@ Bug Fixes since HDF5-1.10.2 release
When the function is called with the 'allgather' parameter set
to a non-true value, the function will receive data from all MPI
- ranks and gather it to the single rank specied by the 'root'
+ ranks and gather it to the single rank specified by the 'root'
parameter. However, the bug in the function caused memory for
the received data to be allocated on all MPI ranks, not just the
singular rank specified as the receiver. In some circumstances,
@@ -3431,11 +3431,11 @@ New Features
file driver in HDF5 by previously having called H5Pset_fapl_mpio().
HDF5 parallel file operations which do not employ multiple ranks
- e.g. specifiying MPI_COMM_SELF (whose MPI_Comm_size == 1)
+ e.g. specifying MPI_COMM_SELF (whose MPI_Comm_size == 1)
as opposed to MPI_COMM_WORLD, will not be affected by this
optimization. Conversely, parallel file operations on subgroups
of MPI_COMM_WORLD are allowed to be run in parallel with each
- subgroup operating as an independant collection of processes.
+ subgroup operating as an independent collection of processes.
(RAW - 2017/10/10, HDFFV-10294)
@@ -3574,7 +3574,7 @@ New Features
(ADB - 2017/12/20)
- The H5I_REFERENCE value in the H5I_type_t enum (defined in H5Ipublic.h)
- has been marked as deprectated.
+ has been marked as deprecated.
JNI code which refers to this value will be removed in a future
major version of the library. The code will remain unchanged in the
@@ -3895,7 +3895,7 @@ Bug Fixes since HDF5-1.10.1 release
Use of this VFD feature flag was not in line with the documentation in
the public H5FDpublic.h file. In particular, it was being used as a
- proxy for determining if SWMR I/O is allowed. This is unecessary as we
+ proxy for determining if SWMR I/O is allowed. This is unnecessary as we
already have a feature flag for this (H5FD_SUPPORTS_SWMR_IO).
(DER - 2017/05/31, HDFFV-10214)
@@ -4796,7 +4796,7 @@ Bug Fixes since HDF5-1.10.0-patch1 release
files to ensure that the library fails gracefully in these cases:
* Writing to a corrupted file that has an object message which is
- incorrectly marked as sharable on disk results in a buffer overflow /
+ incorrectly marked as shareable on disk results in a buffer overflow /
invalid write instead of a clean error message.
* Decoding data from a corrupted file with a dataset encoded with the
@@ -5258,7 +5258,7 @@ This release supports the following features:
identically to the 1.8 library with the --with-default-api-version=v18
configure flag. This allows existing code to be compiled with the 1.10
library without requiring immediate changes to the application source
- code. For addtional configuration options and other details, see
+ code. For additional configuration options and other details, see
"API Compatibility Macros in HDF5" at
https://www.hdfgroup.org/HDF5/doc/RM/APICompatMacros.html.
@@ -5342,7 +5342,7 @@ This release supports the following features:
Library
--------
- - Concurrent Access to HDF5 Files - Single Writer/ Multple Reader (SWMR)
+ - Concurrent Access to HDF5 Files - Single Writer/ Multiple Reader (SWMR)
The Single Writer/ Multiple Reader or SWMR feature enables users to
read data concurrently while writing it. Communications between the
@@ -5970,7 +5970,7 @@ This release supports the following features:
identically to the 1.8 library with the --with-default-api-version=v18
configure flag. This allows existing code to be compiled with the 1.10
library without requiring immediate changes to the application source
- code. For addtional configuration options and other details, see
+ code. For additional configuration options and other details, see
"API Compatibility Macros in HDF5" at
https://www.hdfgroup.org/HDF5/doc/RM/APICompatMacros.html.
@@ -6054,7 +6054,7 @@ This release supports the following features:
Library
--------
- - Concurrent Access to HDF5 Files - Single Writer/ Multple Reader (SWMR)
+ - Concurrent Access to HDF5 Files - Single Writer/ Multiple Reader (SWMR)
The Single Writer/ Multiple Reader or SWMR feature enables users to
read data concurrently while writing it. Communications between the
diff --git a/release_docs/HISTORY-1_10_0-1_12_0.txt b/release_docs/HISTORY-1_10_0-1_12_0.txt
index 60516fc..a83e58d 100644
--- a/release_docs/HISTORY-1_10_0-1_12_0.txt
+++ b/release_docs/HISTORY-1_10_0-1_12_0.txt
@@ -810,7 +810,7 @@ Bug Fixes since HDF5-1.10.2 release
When the function is called with the 'allgather' parameter set
to a non-true value, the function will receive data from all MPI
- ranks and gather it to the single rank specied by the 'root'
+ ranks and gather it to the single rank specified by the 'root'
parameter. However, the bug in the function caused memory for
the received data to be allocated on all MPI ranks, not just the
singular rank specified as the receiver. In some circumstances,
diff --git a/release_docs/HISTORY-1_12.txt b/release_docs/HISTORY-1_12.txt
index 641ada2..e461465 100644
--- a/release_docs/HISTORY-1_12.txt
+++ b/release_docs/HISTORY-1_12.txt
@@ -654,7 +654,7 @@ Bug Fixes since HDF5-1.12.0 release
dimension sizes.
The problem was fixed by ensuring that decoding the dimension sizes
- and max values will not go beyong the end of the buffer.
+ and max values will not go beyond the end of the buffer.
(BMR - 2021/05/12, HDFFV-11223)
@@ -995,7 +995,7 @@ Bug Fixes since HDF5-1.12.0 release
- h5diff fixed a command line parsing error.
h5diff would ignore the argument to -d (delta) if it is smaller than DBL_EPSILON.
- The macro H5_DBL_ABS_EQUAL was removed and a direct value comparision was used.
+ The macro H5_DBL_ABS_EQUAL was removed and a direct value comparison was used.
(ADB - 2020/07/20, HDFFV-10897)
diff --git a/release_docs/HISTORY-1_8.txt b/release_docs/HISTORY-1_8.txt
index da3d849..1e31010 100644
--- a/release_docs/HISTORY-1_8.txt
+++ b/release_docs/HISTORY-1_8.txt
@@ -1213,7 +1213,7 @@ New Features
H5PLreplace - Replace the path at the specified index.
H5PLinsert - Insert a plugin path at the specified index, moving
other paths after the index.
- H5PLremove - Remove the plugin path at the specifed index and
+ H5PLremove - Remove the plugin path at the specified index and
compacting the list.
H5PLget - Query the plugin path at the specified index.
H5PLsize - Query the size of the current list of plugin paths.
@@ -1251,7 +1251,7 @@ New Features
// Returns a copy of the creation property list of a datatype.
DataType::getCreatePlist() - C API H5Tget_create_plist()
- // Opens an object within a location, regarless its type.
+ // Opens an object within a location, regardless its type.
H5Location::openObjId() - C API H5Oopen()
H5Location::openObjId() - C API H5Oopen()
@@ -1730,7 +1730,7 @@ Bug Fixes since HDF5-1.8.17
files to ensure that the library fails gracefully in these cases:
* Writing to a corrupted file that has an object message which is
- incorrectly marked as sharable on disk results in a buffer overflow /
+ incorrectly marked as shareable on disk results in a buffer overflow /
invalid write instead of a clean error message.
* Decoding data from a corrupted file with a dataset encoded with the
@@ -4622,7 +4622,7 @@ Bug Fixes since HDF5-1.8.13
Removed the module name from the package list of Fortran modules because
that module was never generated. This was only an issue for Fortran
- compliers that are not F2003 compatible.
+ compilers that are not F2003 compatible.
(ADB - 2014/10/16, HDFFV-8932)
@@ -4632,7 +4632,7 @@ Bug Fixes since HDF5-1.8.13
change was made which required the renaming of several public symbols in
H5Ppublic.h.
- The shared libary interface version number ("soname") has been increased
+ The shared library interface version number ("soname") has been increased
on account of these changes. For a full list of the changed symbols, see
the interface compatibility report, which is available as a link off of
the 'HDF5 Software Changes from Release to Release' document:
@@ -5891,7 +5891,7 @@ Known Problems
* The data conversion test dt_arith.c fails in "long double" to integer
conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -6746,7 +6746,7 @@ Bug Fixes since HDF5-1.8.10
- Fixed a bug in the core VFD that causes failures when opening files
> 2 GB.
Fixes HDFFV-8124 and HDFFV-8158 (DER 2013/01/07)
- - Fixed a bug where unintialized memory was read during variable-length
+ - Fixed a bug where uninitialized memory was read during variable-length
type conversion. This caused segfaults in netCDF. Fixes HDFFV-8159.
(DER 2013/03/30)
- Removed the H5Pset/get_dxpl_multi functions from the library. The
@@ -7094,7 +7094,7 @@ Known Problems
* The data conversion test dt_arith.c fails in "long double" to integer
conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -7718,8 +7718,8 @@ Known Problems
(MSC - 2012/04/18)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -8317,8 +8317,8 @@ Known Problems
(MSC - 2012/04/18)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -8696,7 +8696,7 @@ Bug Fixes since HDF5-1.8.8
------
- h5ltget_attribute_string_f: The h5ltget_attribute_string_f used to return
the C NULL character in the returned character buffer. The returned
- charactor buffer now does not return the C NULL character; the buffer
+ character buffer now does not return the C NULL character; the buffer
is blank-padded if needed. (MSB - 2012/03/23)
@@ -8799,7 +8799,7 @@ The following platforms and compilers have been tested for this release.
Mac OS X Snow Leopard 10.6.8 i686-apple-darwin10-gcc-4.2.1 (GCC) 4.2.1 (gcc)
Darwin Kernel Version 10.8.0 i686-apple-darwin10-g++-4.2.1 (GCC) 4.2.1 (g++)
Intel 64-bit (Apple Inc. build 5666) (dot 3)
- (fred) GNU Fortan (GCC) 4.6.1 (gfortran)
+ (fred) GNU Fortran (GCC) 4.6.1 (gfortran)
Intel C (icc), Fortran (ifort), C++ (icpc)
12.1.0.038 Build 20110811
@@ -8976,8 +8976,8 @@ Known Problems
(MSC - 2012/04/18)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -9394,7 +9394,7 @@ Bug Fixes since HDF5-1.8.7
Tools
-----
- h5diff: fixed segfault over non-comparable attribute with different
- dimention or rank, along with '-c' option to display details.
+ dimension or rank, along with '-c' option to display details.
(JKM - 2011/10/24 - HDFFV-7770)
- Fixed h5diff to display all the comparable objects and attributes
regardless of detecting non-comparables. (JKM - 2011/09/16 - HDFFV-7693)
@@ -9433,7 +9433,7 @@ Bug Fixes since HDF5-1.8.7
(-d, -p and --use-system-epsilon) are used together.
(JKM - 2011/07/07 - HDFFV-7600)
- Fixed h5dump so that it displays the first line of each element in correct
- position for multiple dimention array types. Before this fix,
+ position for multiple dimension array types. Before this fix,
the first line of each element in an array was
displayed after the last line of previous element without
moving to the next line (+indentation).
@@ -10024,7 +10024,7 @@ Bug Fixes since HDF5-1.8.6
ibm-aix. (AKC - 2011/4/14)
- Shared C libraries are no longer disabled on Mac when Fortran
is enabled. Shared Fortran libraries are still not supported on Mac,
- so configure will disable them by default, but this is overrideable
+ so configure will disable them by default, but this is overridable
with the new --enable-unsupported configure option. The configure
summary has been updated to reflect the fact that the shared-ness of
the C++/Fortran wrapper libraries may not align with the C library.
@@ -10666,7 +10666,7 @@ Bug Fixes since HDF5-1.8.5
- Added a new set of unit tests that are run during 'make check' to verify
the behavior of the metadata accumulator. (MAM - 2010/10/15)
- Modified library to always cache symbol table information. Libraries
- from version 1.6.3 and earler have a bug which causes them to require
+ from version 1.6.3 and earlier have a bug which causes them to require
this information for some operations. (NAF - 2010/09/21 - Bug 1864)
- Fixed a bug where the library could generate an assertion/core dump when
a file that had been created with H5Pset_libver_bounds(fapl,
@@ -11826,7 +11826,7 @@ Bug Fixes since HDF5-1.8.3
created files exhibiting this problem should now be readable.
NAF - 2009/10/27 - 1483
- Fixed error in library's internal caching mechanisms which could cause
- an assertion failure (and attendent core dump) when encountering an
+ an assertion failure (and attendant core dump) when encountering an
unusually formatted file. (QAK - 2009/10/13)
- Fixed incorrect return value for H5Pget_preserve. AKC - 2009/10/08 - 1628
- Fixed an assertion failure that occurred when H5Ocopy was called on a
@@ -12350,7 +12350,7 @@ Bug Fixes since HDF5-1.8.2
can only be successfully opened with write access. NAF - 2009/03/23 - 1189
- Removed the long_long #define and replaced all instances with
"long long". This caused problems with third party products. All
- currently supported compliers support the type. ADB - 2009/03/05
+ currently supported compilers support the type. ADB - 2009/03/05
- Fixed various bugs that could prevent the fill value from being written
in certain rare cases. NAF - 2009/02/26 - 1469
- Fixed a bug that prevented more than one dataset chunk from being cached
@@ -13488,7 +13488,7 @@ Bug Fixes since HDF5-1.8.0
- The h5tset(get)_fields subroutines were missing the parameter to
specify a sign position; fixed. (EIP - 2008/05/23)
- Many APIs were fixed to work with the 8-byte integers in Fortran vs.
- 4-byte integers in C. This change is trasparent to user applications.
+ 4-byte integers in C. This change is transparent to user applications.
C++ API
------
@@ -13882,7 +13882,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF5 make files and test scipts to execute your filter script
+ You get the HDF5 make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
@@ -14419,7 +14419,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF5 make files and test scipts to execute your filter script
+ You get the HDF5 make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
diff --git a/release_docs/HISTORY-1_8_0-1_10_0.txt b/release_docs/HISTORY-1_8_0-1_10_0.txt
index 616aa24..ffdcb56 100644
--- a/release_docs/HISTORY-1_8_0-1_10_0.txt
+++ b/release_docs/HISTORY-1_8_0-1_10_0.txt
@@ -211,7 +211,7 @@ New Features
Parallel Library:
-----------------
- - Add H5Pget_mpio_no_collective_cause() function that retrive reasons
+ - Add H5Pget_mpio_no_collective_cause() function that retrieves reasons
why the collective I/O was broken during read/write IO access.
(JKM - 2012/08/30 HDFFV-8143)
- Special Collective IO (IO when some processes do not contribute to the
@@ -331,10 +331,10 @@ New Features
- h5repack: Improved performance for big chunked datasets (size > 128MB)
when used with layout (-l) or compression (-f) option.
It would perform much better prior to the improvement,
- especially for cases that chunk dimentions looks like
+ especially for cases that chunk dimensions looks like
"1024x5x1" (compare to "1x5x1024"). When bigger numbers
are toward front and smaller number is toward back in chunk
- dimentions. HDFFV-7862 (JKM - 2012/03/01)
+ dimensions. HDFFV-7862 (JKM - 2012/03/01)
- h5dump: Added new option --no-compact-subset. This option will not
interpret the '[' character as starting the compact form of
subsetting. This is useful when the "h5dump error: unable to
@@ -564,7 +564,7 @@ Bug Fixes since HDF5-1.8.0 release
- Fixed rare corruption bugs that could occur when using the new object
header format. (NAF - 2012/3/15 - HDFFV-7879)
- Creating a dataset in a read-only file caused seg fault when the file
- is closed. It's fixed. The attemp to create a dataset will fail
+ is closed. It's fixed. The attempt to create a dataset will fail
with the error stack indicating the file is read-only. (SLU -
2012/1/25. Issue 7756)
- Fixed a seg fault that could occur when shrinking a dataset with chunks
@@ -700,7 +700,7 @@ Bug Fixes since HDF5-1.8.0 release
(NAF - 2009/03/23 - 1189)
- Removed the long_long #define and replaced all instances with
"long long". This caused problems with third party products. All
- currently supported compliers support the type. (ADB - 2009/03/05)
+ currently supported compilers support the type. (ADB - 2009/03/05)
- Fixed various bugs that could prevent the fill value from being written
in certain rare cases. (NAF - 2009/02/26 - 1469)
- Fixed a bug that prevented more than one dataset chunk from being cached
@@ -995,7 +995,7 @@ Bug Fixes since HDF5-1.8.0 release
- h5ls: fixed segfault when access region reference data in an
attribute. HDFFV-7838 (JKM 2011/12/29)
- h5diff: fixed segfault over non-comparable attribute with different
- dimention or rank, along with '-c' option to display details.
+ dimension or rank, along with '-c' option to display details.
HDFFV-7770 (JKM 2011/10/24)
- Fixed h5diff to display all the comparable object and attribute
regardless of non-comparables. HDFFV-7693 (JKM 2011/09/16)
@@ -1025,7 +1025,7 @@ Bug Fixes since HDF5-1.8.0 release
when mutually exclusive options (-d, -p and --use-system-epsilon)
are used together. HDFFV-7600 (JKM 2011/07/07)
- Fixed h5dump to display the first line of each element into correct
- position for multiple dimention array type.
+ position for multiple dimension array type.
Before this fix, the first line of each element in array were
displayed after the last line of previous element without
moving to the next line (+indentation).
@@ -1077,7 +1077,7 @@ Bug Fixes since HDF5-1.8.0 release
as source file. (XCAO 2011/3/8)
- Fixed h5dump for skipping some values for long array type dataset on
Windows. This issue only occurred on Windows due to the different
- return behavior from _vsnprintf() funtion. Bug#2161 (JKM 2011/3/3)
+ return behavior from _vsnprintf() function. Bug#2161 (JKM 2011/3/3)
- Fixed h5dump for skipping array indices every certain number
when the array type dataset is relatively big. The certain number
varies according to the size of array. Bug#2092 (JKM 2011/2/15).
@@ -1202,7 +1202,7 @@ Bug Fixes since HDF5-1.8.0 release
------
- Lite: The h5ltget_attribute_string_f used to return the C NULL character in the
- returned character buffer. The returned Fortran charactor buffer now does
+ returned character buffer. The returned Fortran character buffer now does
not return the C NULL character. (MSB - 2012/3/23)
- Lite: The h5ltget_dataset_info_f function (gets information about a dataset)
was not correctly returning the dimension array. (PVN - 2009/3/23)
@@ -1262,7 +1262,7 @@ Bug Fixes since HDF5-1.8.0 release
The strings command of Mac inspects library files and older
versions of strings may not know newer library format, resulting
in errors. Fixed by sending the library file as stdin to the strings
- coommand to avoid this problem. (AKC - 2013/03/08 HDFFV-8305)
+ command to avoid this problem. (AKC - 2013/03/08 HDFFV-8305)
- Fixed a typo in the ERROR macro in test/testhdf5.h. It segmentation
faulted when used before. (AKC - 2013/02/12 HDFFV-8267)
@@ -1535,8 +1535,8 @@ Known Problems
of GCC. (Issue 8017. SLU - 2012/6/12)
* The data conversion test dt_arith.c fails in "long double" to integer
- conversion on Ubuntu 11.10 (3.0.0.13 kernal) with GCC 4.6.1 if the library
- is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernal
+ conversion on Ubuntu 11.10 (3.0.0.13 kernel) with GCC 4.6.1 if the library
+ is built with optimization -O3 or -O2. The older GCC (4.5) or newer kernel
(3.2.2 on Fedora) doesn't have the problem. Users should lower down the
optimization level (-O1 or -O0) by defining CFLAGS in the command line of
"configure" like:
@@ -1719,7 +1719,7 @@ Known Problems
exit $RETURN_VALUE
- You get the HDF make files and test scipts to execute your filter script
+ You get the HDF make files and test scripts to execute your filter script
by setting the environment variable "RUNSERIAL" to the full path of the
script prior to running configure for parallel builds. Remember to
"unsetenv RUNSERIAL" before running configure for a serial build.
diff --git a/release_docs/INSTALL_Cygwin.txt b/release_docs/INSTALL_Cygwin.txt
index edf3b92..8dc5bc7 100644
--- a/release_docs/INSTALL_Cygwin.txt
+++ b/release_docs/INSTALL_Cygwin.txt
@@ -2,7 +2,7 @@
HDF5 Build and Install Instructions for Cygwin
************************************************************************
-This document is a instruction on how to build, test and install HDF5 libary on
+This document is an instruction on how to build, test and install HDF5 library on
Cygwin. See detailed information in hdf5/INSTALL.
NOTE: hdf5 can be built with CMake, see the INSTALL_CMake.txt file for more guidance.
@@ -66,7 +66,7 @@ Preconditions:
2.2.2 Szip
The HDF5 library has a predefined compression filter that uses
the extended-Rice lossless compression algorithm for chunked
- datatsets. For more information on Szip compression, license terms,
+ datasets. For more information on Szip compression, license terms,
and obtaining the Szip source code, see:
https://portal.hdfgroup.org/display/HDF5/Szip+Compression+in+HDF+Products
diff --git a/release_docs/RELEASE.txt b/release_docs/RELEASE.txt
index 3d6ff48..deab326 100644
--- a/release_docs/RELEASE.txt
+++ b/release_docs/RELEASE.txt
@@ -350,7 +350,7 @@ Bug Fixes since HDF5-1.12.1 release
function. This improves the I/O performance. In the case of the parallel
library, it also avoids breaking to independent I/O, which makes it
possible to apply a filter when writing or reading data to or from
- teh HDF5 file.
+ the HDF5 file.
(JWSB - 2021/09/13)
diff --git a/src/H5A.c b/src/H5A.c
index 3d20f83..736d3cb 100644
--- a/src/H5A.c
+++ b/src/H5A.c
@@ -1279,7 +1279,7 @@ done:
if all attributes were processed.
DESCRIPTION
- This function interates over the attributes of dataset or group
+ This function iterates over the attributes of dataset or group
specified with 'loc_id' & 'obj_name'. For each attribute of the object,
the 'op_data' and some additional information (specified below) are passed
to the 'op' function. The iteration begins with the '*idx'
@@ -1357,7 +1357,7 @@ done:
if all attributes were processed.
DESCRIPTION
- This function interates over the attributes of dataset or group
+ This function iterates over the attributes of dataset or group
specified with 'loc_id' & 'obj_name'. For each attribute of the object,
the 'op_data' and some additional information (specified below) are passed
to the 'op' function. The iteration begins with the '*idx'
diff --git a/src/H5AC.c b/src/H5AC.c
index 87ea0b7..071e066 100644
--- a/src/H5AC.c
+++ b/src/H5AC.c
@@ -1229,7 +1229,7 @@ done:
* metadata cache flush.
*
* Initially, this means setting up the slist prior to the
- * flush. We do this in a seperate call because
+ * flush. We do this in a separate call because
* H5F__flush_phase2() make repeated calls to H5AC_flush().
* Handling this detail in separate calls allows us to avoid
* the overhead of setting up and taking down the skip list
@@ -1281,7 +1281,7 @@ done:
* flush.
*
* Initially, this means taking down the slist after the
- * flush. We do this in a seperate call because
+ * flush. We do this in a separate call because
* H5F__flush_phase2() make repeated calls to H5AC_flush().
* Handling this detail in separate calls allows us to avoid
* the overhead of setting up and taking down the skip list
diff --git a/src/H5ACdbg.c b/src/H5ACdbg.c
index 2d24adb..656c731 100644
--- a/src/H5ACdbg.c
+++ b/src/H5ACdbg.c
@@ -197,11 +197,11 @@ done:
* If either the parent or the child is not in the metadata
* cache, the function sets *fd_exists_ptr to FALSE.
*
- * If both are in the cache, the childs list of parents is
+ * If both are in the cache, the child's list of parents is
* searched for the proposed parent. If the proposed parent
- * is found in the childs parent list, the function sets
+ * is found in the child's parent list, the function sets
* *fd_exists_ptr to TRUE. In all other non-error cases,
- * the function sets *fd_exists_ptr FALSE.
+ * the function sets *fd_exists_ptr to FALSE.
*
* Return: SUCCEED on success/FAIL on failure. Note that
* *fd_exists_ptr is undefined on failure.
diff --git a/src/H5ACmpio.c b/src/H5ACmpio.c
index dcc1ab2..b2b32f2 100644
--- a/src/H5ACmpio.c
+++ b/src/H5ACmpio.c
@@ -1140,7 +1140,7 @@ done:
*
* While the list of candidate cache entries is prepared
* elsewhere, this function is the main routine for distributing
- * and applying the list. It must be run simultaniously on
+ * and applying the list. It must be run simultaneously on
* all processes that have the relevant file open. To ensure
* proper synchronization, there is a barrier at the beginning
* of this function.
@@ -1319,7 +1319,7 @@ done:
*
* This function is the main routine for handling this
* notification procedure. It must be called
- * simultaniously on all processes that have the relevant
+ * simultaneously on all processes that have the relevant
* file open. To this end, it is called only during a
* sync point, with a barrier prior to the call.
*
diff --git a/src/H5Adense.c b/src/H5Adense.c
index 1575f3c..3298305 100644
--- a/src/H5Adense.c
+++ b/src/H5Adense.c
@@ -337,7 +337,7 @@ H5A__dense_open(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
H5HF_t * fheap = NULL; /* Fractal heap handle */
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
htri_t attr_exists; /* Attribute exists in v2 B-tree */
H5A_t * ret_value = NULL; /* Return value */
@@ -356,7 +356,7 @@ H5A__dense_open(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, NULL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -428,7 +428,7 @@ H5A__dense_insert(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
H5WB_t * wb = NULL; /* Wrapped buffer for attribute data */
uint8_t attr_buf[H5A_ATTR_BUF_SIZE]; /* Buffer for serializing message */
unsigned mesg_flags = 0; /* Flags for storing message */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -442,7 +442,7 @@ H5A__dense_insert(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
htri_t shared_mesg; /* Should this message be stored in the Shared Message table? */
@@ -726,7 +726,7 @@ H5A__dense_write(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
H5HF_t * fheap = NULL; /* Fractal heap handle */
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -742,7 +742,7 @@ H5A__dense_write(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -864,7 +864,7 @@ H5A__dense_rename(H5F_t *f, const H5O_ainfo_t *ainfo, const char *old_name, cons
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
H5B2_t * bt2_corder = NULL; /* v2 B-tree handle for creation order ndex */
H5A_t * attr_copy = NULL; /* Copy of attribute to rename */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
htri_t shared_mesg; /* Should this message be stored in the Shared Message table? */
htri_t attr_exists; /* Attribute exists in v2 B-tree */
herr_t ret_value = SUCCEED; /* Return value */
@@ -881,7 +881,7 @@ H5A__dense_rename(H5F_t *f, const H5O_ainfo_t *ainfo, const char *old_name, cons
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1178,7 +1178,7 @@ H5A__dense_iterate(H5F_t *f, hid_t loc_id, const H5O_ainfo_t *ainfo, H5_index_t
/* Check on iteration order */
if (order == H5_ITER_NATIVE && H5F_addr_defined(bt2_addr)) {
H5A_bt2_ud_it_t udata; /* User data for iterator callback */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
/* Open the fractal heap */
if (NULL == (fheap = H5HF_open(f, ainfo->fheap_addr)))
@@ -1188,7 +1188,7 @@ H5A__dense_iterate(H5F_t *f, hid_t loc_id, const H5O_ainfo_t *ainfo, H5_index_t
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1335,7 +1335,7 @@ H5A__dense_remove(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
H5A_t * attr_copy = NULL; /* Copy of attribute to remove */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -1353,7 +1353,7 @@ H5A__dense_remove(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1582,7 +1582,7 @@ H5A__dense_remove_by_idx(H5F_t *f, const H5O_ainfo_t *ainfo, H5_index_t idx_type
/* If there is an index defined for the field, use it */
if (H5F_addr_defined(bt2_addr)) {
H5A_bt2_ud_rmbi_t udata; /* User data for v2 B-tree record removal */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
/* Open the fractal heap */
if (NULL == (fheap = H5HF_open(f, ainfo->fheap_addr)))
@@ -1592,7 +1592,7 @@ H5A__dense_remove_by_idx(H5F_t *f, const H5O_ainfo_t *ainfo, H5_index_t idx_type
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
@@ -1672,7 +1672,7 @@ H5A__dense_exists(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
H5HF_t * fheap = NULL; /* Fractal heap handle */
H5HF_t * shared_fheap = NULL; /* Fractal heap handle for shared header messages */
H5B2_t * bt2_name = NULL; /* v2 B-tree handle for name index */
- htri_t attr_sharable; /* Flag indicating attributes are sharable */
+ htri_t attr_sharable; /* Flag indicating attributes are shareable */
htri_t ret_value = TRUE; /* Return value */
FUNC_ENTER_PACKAGE
@@ -1690,7 +1690,7 @@ H5A__dense_exists(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name)
if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared")
- /* Get handle for shared message heap, if attributes are sharable */
+ /* Get handle for shared message heap, if attributes are shareable */
if (attr_sharable) {
haddr_t shared_fheap_addr; /* Address of fractal heap to use */
diff --git a/src/H5Adeprec.c b/src/H5Adeprec.c
index e43706f..0153e31 100644
--- a/src/H5Adeprec.c
+++ b/src/H5Adeprec.c
@@ -351,7 +351,7 @@ done:
last operator if it was non-zero, or zero if all attributes were processed.
DESCRIPTION
- This function interates over the attributes of dataset or group
+ This function iterates over the attributes of dataset or group
specified with 'loc_id'. For each attribute of the object, the
'op_data' and some additional information (specified below) are passed
to the 'op' function. The iteration begins with the '*attr_number'
diff --git a/src/H5C.c b/src/H5C.c
index 0fc4416..490db81 100644
--- a/src/H5C.c
+++ b/src/H5C.c
@@ -771,7 +771,7 @@ H5C_prep_for_file_close(H5F_t *f)
*
* 2) Since the FSM settle routines are only invoked once during
* file close, invoking them now will prevent their invocation
- * during a flush, and thus avoid any resulting entrie dirties,
+ * during a flush, and thus avoid any resulting entry dirties,
* deletions, insertion, or moves during the flush.
*/
if (H5C__serialize_cache(f) < 0)
@@ -2564,7 +2564,7 @@ H5C_protect(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *udata, unsign
*
* Also, if the min_clean_size requirement is not met, we
* should also call H5C__make_space_in_cache() to bring us
- * into complience.
+ * into compliance.
*/
if (cache_ptr->index_size >= cache_ptr->max_cache_size)
@@ -4861,7 +4861,7 @@ H5C__autoadjust__ageout__evict_aged_out_entries(H5F_t *f, hbool_t write_permitte
/* for now at least, don't bother to maintain the minimum clean size,
* as the cache should now be less than its maximum size. Due to
- * the vaguries of the cache size reduction algorthim, we may not
+ * the vaguries of the cache size reduction algorithm, we may not
* reduce the size of the cache.
*
* If we do, we will calculate a new minimum clean size, which will
@@ -5873,7 +5873,7 @@ H5C__flush_invalidate_ring(H5F_t *f, H5C_ring_t ring, unsigned flags)
* If either of these happen, and one of the target
* or proxy entries happens to be the next entry in
* the hash bucket, we could either find ourselves
- * either scanning a non-existant entry, scanning
+ * either scanning a non-existent entry, scanning
* through a different bucket, or skipping an entry.
*
* Neither of these are good, so restart the
@@ -8086,7 +8086,7 @@ done:
* To cork the object
* Return error if the object is already corked
* H5C__UNCORK:
- * To uncork the obejct
+ * To uncork the object
* Return error if the object is not corked
* H5C__GET_CORKED:
* To retrieve the cork status of an object in
@@ -8517,7 +8517,7 @@ H5C__serialize_cache(H5F_t *f)
* each entry in the cache to zero before we start the serialization.
* This allows us to detect the case in which any entry is serialized
* more than once (a performance issues), and more importantly, the
- * case is which any flush depencency parent is serializes more than
+ * case is which any flush dependency parent is serializes more than
* once (a correctness issue).
*/
{
@@ -8674,8 +8674,8 @@ H5C__serialize_ring(H5F_t *f, H5C_ring_t ring)
* are serialized correctly, it doesn't matter if we have to go back
* and serialize an entry a second time.
*
- * These possible actions result in the following modfications to
- * tha basic algorithm:
+ * These possible actions result in the following modifications to
+ * the basic algorithm:
*
* 1) In the event of an entry expunge, eviction or removal, we must
* restart the scan as it is possible that the next entry in our
@@ -8725,7 +8725,7 @@ H5C__serialize_ring(H5F_t *f, H5C_ring_t ring)
* condition appears.
*
* Observe that either eviction or removal of entries as a result of
- * a serialization is not a problem as long as the flush depencency
+ * a serialization is not a problem as long as the flush dependency
* tree does not change beyond the removal of a leaf.
*/
while (!done) {
diff --git a/src/H5CS.c b/src/H5CS.c
index 6510c25..e871a75 100644
--- a/src/H5CS.c
+++ b/src/H5CS.c
@@ -47,7 +47,7 @@ typedef struct H5CS_t {
* each thread individually. The association of stacks to threads will
* be handled by the pthread library.
*
- * In order for this macro to work, H5CS_get_my_stack() must be preceeded
+ * In order for this macro to work, H5CS_get_my_stack() must be preceded
* by "H5CS_t *fstack =".
*/
static H5CS_t *H5CS__get_stack(void);
@@ -127,7 +127,7 @@ H5CS__get_stack(void)
herr_t
H5CS_print_stack(const H5CS_t *fstack, FILE *stream)
{
- const int indent = 2; /* Indention level */
+ const int indent = 2; /* Indentation level */
int i; /* Local index ariable */
/* Don't push this function on the function stack... :-) */
diff --git a/src/H5CX.c b/src/H5CX.c
index a9dc8fb..3392eea 100644
--- a/src/H5CX.c
+++ b/src/H5CX.c
@@ -50,7 +50,7 @@
* each thread individually. The association of contexts to threads will
* be handled by the pthread library.
*
- * In order for this macro to work, H5CX_get_my_context() must be preceeded
+ * In order for this macro to work, H5CX_get_my_context() must be preceded
* by "H5CX_node_t *ctx =".
*/
#define H5CX_get_my_context() H5CX__get_context()
diff --git a/src/H5Cdbg.c b/src/H5Cdbg.c
index 104f1af..de9019c 100644
--- a/src/H5Cdbg.c
+++ b/src/H5Cdbg.c
@@ -844,9 +844,9 @@ H5C_stats__reset(H5C_t H5_ATTR_UNUSED *cache_ptr)
* If either the parent or the child is not in the metadata
* cache, the function sets *fd_exists_ptr to FALSE.
*
- * If both are in the cache, the childs list of parents is
+ * If both are in the cache, the child's list of parents is
* searched for the proposed parent. If the proposed parent
- * is found in the childs parent list, the function sets
+ * is found in the child's parent list, the function sets
* *fd_exists_ptr to TRUE. In all other non-error cases,
* the function sets *fd_exists_ptr FALSE.
*
diff --git a/src/H5Cimage.c b/src/H5Cimage.c
index 43ad760..98c1291 100644
--- a/src/H5Cimage.c
+++ b/src/H5Cimage.c
@@ -591,7 +591,7 @@ H5C__deserialize_prefetched_entry(H5F_t *f, H5C_t *cache_ptr, H5C_cache_entry_t
*
* Note that at present, dirty can't be set to true with prefetched
* entries. However this may change, so include this functionality
- * against that posibility.
+ * against that possibility.
*
* Also, note that it is possible for a prefetched entry to be dirty --
* hence the value assigned to ds_entry_ptr->is_dirty below.
@@ -1132,7 +1132,7 @@ done:
* image superblock extension message must be deleted from
* the superblock extension and the image block freed
*
- * Contrawise, if the file is openened R/O, the metadata
+ * Contrawise, if the file is opened R/O, the metadata
* cache image superblock extension message and image block
* must be left as is. Further, any dirty entries in the
* cache image block must be marked as clean to avoid
@@ -1824,7 +1824,7 @@ done:
* Purpose: Decode the metadata cache image entry from the supplied
* buffer into the supplied instance of H5C_image_entry_t.
* This includes allocating a buffer for the entry image,
- * loading it, and seting ie_ptr->image_ptr to point to
+ * loading it, and setting ie_ptr->image_ptr to point to
* the buffer.
*
* Advances the buffer pointer to the first byte
@@ -2341,7 +2341,7 @@ done:
* also be a flush dependency child.
*
* Finally, note that for purposes of the cache image, flush
- * dependency height ends when a flush dependecy relation
+ * dependency height ends when a flush dependency relation
* passes off the cache image.
*
* On exit, the flush dependency height of each entry in the
@@ -2503,7 +2503,7 @@ H5C__prep_for_file_close__compute_fd_heights(const H5C_t *cache_ptr)
entry_ptr = entry_ptr->il_next;
} /* while (entry_ptr != NULL) */
- /* At present, no extenal parent or child flush dependency links
+ /* At present, no external parent or child flush dependency links
* should exist -- hence the following assertions. This will change
* if we support ageout of entries in the cache image.
*/
@@ -3384,7 +3384,7 @@ done:
* creating if specified.
*
* In general, the size and location of the cache image block
- * will be unknow at the time that the cache image superblock
+ * will be unknown at the time that the cache image superblock
* message is created. A subsequent call to this routine will
* be used to write the correct data.
*
diff --git a/src/H5Cmpio.c b/src/H5Cmpio.c
index 66c6601..fa09b20 100644
--- a/src/H5Cmpio.c
+++ b/src/H5Cmpio.c
@@ -1095,7 +1095,7 @@ done:
*
* Note that this function is a modified version of
* H5C_flush_cache() -- any changes there may need to be
- * reflected here and vise versa.
+ * reflected here and vice versa.
*
* Return: Non-negative on success/Negative on failure.
*
diff --git a/src/H5Cprefetched.c b/src/H5Cprefetched.c
index 07fca2b..f89c233 100644
--- a/src/H5Cprefetched.c
+++ b/src/H5Cprefetched.c
@@ -302,7 +302,7 @@ H5C__prefetched_entry_free_icr(void *_thing)
HDassert(entry_ptr->fd_parent_count == 0);
if (entry_ptr->image_ptr != NULL)
- HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "prefetched entry image buffer still attatched?")
+ HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "prefetched entry image buffer still attached?")
entry_ptr = H5FL_FREE(H5C_cache_entry_t, entry_ptr);
diff --git a/src/H5Dchunk.c b/src/H5Dchunk.c
index 8884cb2..f0fc3a4 100644
--- a/src/H5Dchunk.c
+++ b/src/H5Dchunk.c
@@ -959,7 +959,7 @@ H5D__chunk_init(H5F_t *f, const H5D_t *const dset, hid_t dapl_id)
if (!(scaled_power2up = H5VM_power2up(rdcc->scaled_dims[u])))
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "unable to get the next power of 2")
- /* Inital 'power2up' values for scaled dimensions */
+ /* Initial 'power2up' values for scaled dimensions */
rdcc->scaled_power2up[u] = scaled_power2up;
/* Number of bits required to encode scaled dimension size */
@@ -2535,7 +2535,7 @@ H5D__chunk_read(H5D_io_info_t *io_info, const H5D_type_info_t *type_info, hsize_
HDassert((H5F_addr_defined(udata.chunk_block.offset) && udata.chunk_block.length > 0) ||
(!H5F_addr_defined(udata.chunk_block.offset) && udata.chunk_block.length == 0));
- /* Check for non-existant chunk & skip it if appropriate */
+ /* Check for non-existent chunk & skip it if appropriate */
if (H5F_addr_defined(udata.chunk_block.offset) || UINT_MAX != udata.idx_hint ||
!skip_missing_chunks) {
H5D_io_info_t *chk_io_info; /* Pointer to I/O info object for this chunk */
@@ -3603,7 +3603,7 @@ H5D__chunk_cache_prune(const H5D_t *dset, size_t size)
* traversing the list when pointer pN reaches wN percent of the original
* list. In other words, preemption method N gets to consider entries in
* approximate least recently used order w0 percent before method N+1
- * where 100% means tha method N will run to completion before method N+1
+ * where 100% means that method N will run to completion before method N+1
* begins. The pointers participating in the list traversal are each
* given a chance at preemption before any of the pointers are advanced.
*/
@@ -6043,7 +6043,7 @@ H5D__chunk_copy_cb(const H5D_chunk_rec_t *chunk_rec, void *_udata)
size_t buf_size = udata->buf_size; /* Size of chunk buffer */
const H5O_pline_t *pline = udata->pline; /* I/O pipeline for applying filters */
- /* needed for commpressed variable length data */
+ /* needed for compressed variable length data */
hbool_t must_filter = FALSE; /* Whether chunk must be filtered during copy */
size_t nbytes; /* Size of chunk in file (in bytes) */
H5Z_cb_t filter_cb; /* Filter failure callback struct */
diff --git a/src/H5Dint.c b/src/H5Dint.c
index e950096..6d3371d 100644
--- a/src/H5Dint.c
+++ b/src/H5Dint.c
@@ -410,7 +410,7 @@ done:
*
* Return:
* Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*-------------------------------------------------------------------------
*/
herr_t
@@ -735,7 +735,7 @@ H5D__calculate_minimum_header_size(H5F_t *file, H5D_t *dset, H5O_t *ohdr)
H5T_t * type = NULL;
H5O_fill_t *fill_prop = NULL;
hbool_t use_at_least_v18 = FALSE;
- const char continuation[1] = ""; /* requred for work-around */
+ const char continuation[1] = ""; /* required for work-around */
size_t get_value = 0;
size_t ret_value = 0;
@@ -782,7 +782,7 @@ H5D__calculate_minimum_header_size(H5F_t *file, H5D_t *dset, H5O_t *ohdr)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, 0, "can't get size of continuation message")
ret_value += get_value;
- /* Fill Value (backwards compatability) message size */
+ /* Fill Value (backwards compatibility) message size */
if (fill_prop->buf && !use_at_least_v18) {
H5O_fill_t old_fill_prop; /* Copy for writing "old" fill value */
@@ -871,7 +871,7 @@ H5D__prepare_minimized_oh(H5F_t *file, H5D_t *dset, H5O_loc_t *oloc)
if (ohdr_size == 0)
HGOTO_ERROR(H5E_OHDR, H5E_BADVALUE, FAIL, "computed header size is invalid")
- /* Special allocation of space for compact datsets is handled by the call here. */
+ /* Special allocation of space for compact datasets is handled by the call here. */
if (H5O_apply_ohdr(file, oh, dset->shared->dcpl_id, ohdr_size, (size_t)1, oloc) == FAIL)
HGOTO_ERROR(H5E_OHDR, H5E_BADVALUE, FAIL, "can't apply object header to file")
diff --git a/src/H5Dmpio.c b/src/H5Dmpio.c
index dfdf500..4f52127 100644
--- a/src/H5Dmpio.c
+++ b/src/H5Dmpio.c
@@ -664,7 +664,7 @@ H5D__mpio_opt_possible(const H5D_io_info_t *io_info, const H5S_t *file_space, co
/* Check to see if the process is reading the entire dataset */
if (H5S_GET_SELECT_TYPE(file_space) != H5S_SEL_ALL)
local_cause[1] |= H5D_MPIO_RANK0_NOT_H5S_ALL;
- /* Only perform this optimization for contigous datasets, currently */
+ /* Only perform this optimization for contiguous datasets, currently */
else if (H5D_CONTIGUOUS != io_info->dset->shared->layout.type)
/* Flag to do a MPI_Bcast of the data from one proc instead of
* having all the processes involved in the collective I/O.
@@ -1329,9 +1329,9 @@ H5D__link_chunk_collective_io(H5D_io_info_t *io_info, const H5D_type_info_t *typ
int sum_chunk, int mpi_rank, int mpi_size)
{
H5D_chunk_addr_info_t *chunk_addr_info_array = NULL;
- MPI_Datatype chunk_final_mtype; /* Final memory MPI datatype for all chunks with seletion */
+ MPI_Datatype chunk_final_mtype; /* Final memory MPI datatype for all chunks with selection */
hbool_t chunk_final_mtype_is_derived = FALSE;
- MPI_Datatype chunk_final_ftype; /* Final file MPI datatype for all chunks with seletion */
+ MPI_Datatype chunk_final_ftype; /* Final file MPI datatype for all chunks with selection */
hbool_t chunk_final_ftype_is_derived = FALSE;
H5D_storage_t ctg_store; /* Storage info for "fake" contiguous dataset */
size_t total_chunks;
@@ -2895,7 +2895,7 @@ done:
*
* 1) Each process provides two piece of information for all chunks having selection
* a) chunk index
- * b) wheather this chunk is regular(for MPI derived datatype not working case)
+ * b) whether this chunk is regular(for MPI derived datatype not working case)
*
* 2) Gather all the information to the root process
*
diff --git a/src/H5Dvirtual.c b/src/H5Dvirtual.c
index beee8c6..9ff3b06 100644
--- a/src/H5Dvirtual.c
+++ b/src/H5Dvirtual.c
@@ -195,7 +195,7 @@ H5D_virtual_check_mapping_pre(const H5S_t *vspace, const H5S_t *src_space,
"can't get number of elements in non-unlimited dimension")
if (nenu_vs != nenu_ss)
HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL,
- "numbers of elemnts in the non-unlimited dimensions is different for source and "
+ "numbers of elements in the non-unlimited dimensions is different for source and "
"virtual spaces")
} /* end if */
/* We will handle the printf case after parsing the source names */
@@ -1482,7 +1482,7 @@ H5D__virtual_set_extent_unlim(const H5D_t *dset)
storage->list[i].source_dset.dset->shared->space) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, FAIL, "can't copy source dataspace extent")
- /* Get source space dimenstions */
+ /* Get source space dimensions */
if (H5S_get_simple_extent_dims(storage->list[i].source_select, curr_dims, NULL) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get source space dimensions")
@@ -1960,7 +1960,7 @@ H5D__virtual_init_all(const H5D_t *dset)
storage->list[i].source_dset.dset->shared->space) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, FAIL, "can't copy source dataspace extent")
- /* Get source space dimenstions */
+ /* Get source space dimensions */
if (H5S_get_simple_extent_dims(storage->list[i].source_select, source_dims, NULL) < 0)
HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get source space dimensions")
@@ -2344,7 +2344,7 @@ H5D__virtual_is_data_cached(const H5D_shared_t *shared_dset)
if (storage->list[i].psfn_nsubs || storage->list[i].psdn_nsubs) {
/* Iterate over sub-source dsets */
for (j = storage->list[i].sub_dset_io_start; j < storage->list[i].sub_dset_io_end; j++)
- /* Check for cahced data in source dset */
+ /* Check for cached data in source dset */
if (storage->list[i].sub_dset[j].dset &&
storage->list[i].sub_dset[j].dset->shared->layout.ops->is_data_cached &&
storage->list[i].sub_dset[j].dset->shared->layout.ops->is_data_cached(
@@ -2678,7 +2678,7 @@ H5D__virtual_post_io(H5O_storage_virtual_t *storage)
/*-------------------------------------------------------------------------
* Function: H5D__virtual_read_one
*
- * Purpose: Read from a singe source dataset in a virtual dataset.
+ * Purpose: Read from a single source dataset in a virtual dataset.
*
* Return: Non-negative on success/Negative on failure
*
@@ -2868,7 +2868,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5D__virtual_write_one
*
- * Purpose: Write to a singe source dataset in a virtual dataset.
+ * Purpose: Write to a single source dataset in a virtual dataset.
*
* Return: Non-negative on success/Negative on failure
*
diff --git a/src/H5E.c b/src/H5E.c
index 10d31f0..5bfda52 100644
--- a/src/H5E.c
+++ b/src/H5E.c
@@ -720,7 +720,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5E__close_msg
*
- * Purpose: Private function to close an error messge.
+ * Purpose: Private function to close an error message.
*
* Return: SUCCEED/FAIL
*
diff --git a/src/H5F.c b/src/H5F.c
index 3d044c5..4e7ce26 100644
--- a/src/H5F.c
+++ b/src/H5F.c
@@ -1427,7 +1427,7 @@ done:
* 1) The file being opened has v3 superblock
* 2) The file is opened with H5F_ACC_RDWR
* 3) The file is not already marked for SWMR writing
- * 4) Current implementaion for opened objects:
+ * 4) Current implementation for opened objects:
* --only allow datasets and groups without attributes
* --disallow named datatype with/without attributes
* --disallow opened attributes attached to objects
diff --git a/src/H5FA.c b/src/H5FA.c
index ad69ee2..cb244ff 100644
--- a/src/H5FA.c
+++ b/src/H5FA.c
@@ -95,7 +95,7 @@ H5FL_BLK_DEFINE(fa_native_elmt);
/*-------------------------------------------------------------------------
* Function: H5FA__new
*
- * Purpose: Allocate and initialize a new fixe array wrapper in memory
+ * Purpose: Allocate and initialize a new fixed array wrapper in memory
*
* Return: Pointer to farray wrapper success
* NULL on failure
diff --git a/src/H5FAcache.c b/src/H5FAcache.c
index 367b4b0..e6e32da 100644
--- a/src/H5FAcache.c
+++ b/src/H5FAcache.c
@@ -943,7 +943,7 @@ done:
* file space block set.
*
* This function is needed when the data block is paged, as
- * the datablock header and all its pages are allocted as a
+ * the datablock header and all its pages are allocated as a
* single contiguous chunk of file space, and must be
* deallocated the same way.
*
diff --git a/src/H5FD.c b/src/H5FD.c
index 11cbdc4..427057e 100644
--- a/src/H5FD.c
+++ b/src/H5FD.c
@@ -1324,7 +1324,7 @@ H5FD_get_fs_type_map(const H5FD_t *file, H5FD_mem_t *type_map)
HDassert(file->cls);
HDassert(type_map);
- /* Check for VFD class providing a type map retrieval rouine */
+ /* Check for VFD class providing a type map retrieval routine */
if (file->cls->get_type_map) {
/* Retrieve type mapping for this file */
if ((file->cls->get_type_map)(file, type_map) < 0)
diff --git a/src/H5FDcore.c b/src/H5FDcore.c
index 0b96891..94b88b0 100644
--- a/src/H5FDcore.c
+++ b/src/H5FDcore.c
@@ -1510,7 +1510,7 @@ done:
* Addendum -- 12/2/11
* For file images opened with the core file driver, it is
* necessary that we avoid reallocating the core file driver's
- * buffer uneccessarily.
+ * buffer unnecessarily.
*
* To this end, I have made the following functional changes
* to this function.
diff --git a/src/H5FDhdfs.c b/src/H5FDhdfs.c
index 002aab6..9a5aae0 100644
--- a/src/H5FDhdfs.c
+++ b/src/H5FDhdfs.c
@@ -567,16 +567,16 @@ done:
* Function: H5FD__hdfs_validate_config()
*
* Purpose: Test to see if the supplied instance of H5FD_hdfs_fapl_t
- * contains internally consistant data. Return SUCCEED if so,
+ * contains internally consistent data. Return SUCCEED if so,
* and FAIL otherwise.
*
- * Note the difference between internally consistant and
+ * Note the difference between internally consistent and
* correct. As we will have to try to access the target
* object to determine whether the supplied data is correct,
- * we will settle for internal consistancy at this point
+ * we will settle for internal consistency at this point
*
* Return: SUCCEED if instance of H5FD_hdfs_fapl_t contains internally
- * consistant data, FAIL otherwise.
+ * consistent data, FAIL otherwise.
*
* Programmer: Jacob Smith
* 9/10/17
diff --git a/src/H5FDlog.c b/src/H5FDlog.c
index 7214b75..5873fc1 100644
--- a/src/H5FDlog.c
+++ b/src/H5FDlog.c
@@ -1559,7 +1559,7 @@ H5FD__log_write(H5FD_t *_file, H5FD_mem_t type, hid_t H5_ATTR_UNUSED dxpl_id, ha
HDfprintf(file->logfp, "%10" PRIuHADDR "-%10" PRIuHADDR " (%10zu bytes) (%s) Written", orig_addr,
(orig_addr + orig_size) - 1, orig_size, flavors[type]);
- /* Check if this is the first write into a "default" section, grabbed by the metadata agregation
+ /* Check if this is the first write into a "default" section, grabbed by the metadata aggregation
* algorithm */
if (file->fa.flags & H5FD_LOG_FLAVOR) {
if ((H5FD_mem_t)file->flavor[orig_addr] == H5FD_MEM_DEFAULT) {
diff --git a/src/H5FDmirror.c b/src/H5FDmirror.c
index f36b9c3..e0e240c 100644
--- a/src/H5FDmirror.c
+++ b/src/H5FDmirror.c
@@ -1168,7 +1168,7 @@ done:
/* -------------------------------------------------------------------------
* Function: H5FD__mirror_fapl_get
*
- * Purpose: Get the file access propety list which could be used to create
+ * Purpose: Get the file access property list which could be used to create
* an identical file.
*
* Return: Success: pointer to the new file access property list value.
diff --git a/src/H5FDros3.c b/src/H5FDros3.c
index decc341..f6765d5 100644
--- a/src/H5FDros3.c
+++ b/src/H5FDros3.c
@@ -380,7 +380,7 @@ H5FD__ros3_term(void)
* Function: H5Pset_fapl_ros3
*
* Purpose: Modify the file access property list to use the H5FD_ROS3
- * driver defined in this source file. All driver specfic
+ * driver defined in this source file. All driver specific
* properties are passed in as a pointer to a suitably
* initialized instance of H5FD_ros3_fapl_t
*
@@ -423,16 +423,16 @@ done:
* Function: H5FD__ros3_validate_config()
*
* Purpose: Test to see if the supplied instance of H5FD_ros3_fapl_t
- * contains internally consistant data. Return SUCCEED if so,
+ * contains internally consistent data. Return SUCCEED if so,
* and FAIL otherwise.
*
- * Note the difference between internally consistant and
+ * Note the difference between internally consistent and
* correct. As we will have to try to access the target
* object to determine whether the supplied data is correct,
- * we will settle for internal consistancy at this point
+ * we will settle for internal consistency at this point
*
* Return: SUCCEED if instance of H5FD_ros3_fapl_t contains internally
- * consistant data, FAIL otherwise.
+ * consistent data, FAIL otherwise.
*
* Programmer: Jacob Smith
* 9/10/17
diff --git a/src/H5FSprivate.h b/src/H5FSprivate.h
index fdc5c91..4cd18b9 100644
--- a/src/H5FSprivate.h
+++ b/src/H5FSprivate.h
@@ -116,7 +116,7 @@ typedef struct H5FS_section_class_t {
const struct H5FS_section_class_t *, const uint8_t *, haddr_t, hsize_t,
unsigned *); /* Routine to deserialize a buffer into a "live" section */
htri_t (*can_merge)(const H5FS_section_info_t *, const H5FS_section_info_t *,
- void *); /* Routine to determine if two nodes are mergable */
+ void *); /* Routine to determine if two nodes are mergeable */
herr_t (*merge)(H5FS_section_info_t **, H5FS_section_info_t *, void *); /* Routine to merge two nodes */
htri_t (*can_shrink)(const H5FS_section_info_t *,
void *); /* Routine to determine if node can shrink container */
diff --git a/src/H5FSsection.c b/src/H5FSsection.c
index 00e5103..5fa2ad7 100644
--- a/src/H5FSsection.c
+++ b/src/H5FSsection.c
@@ -2089,11 +2089,11 @@ H5FS_sect_change_class(H5F_t *f, H5FS_t *fspace, H5FS_section_info_t *sect, uint
} /* end else */
} /* end if */
- /* Check if the section's class change will affect the mergable list */
+ /* Check if the section's class change will affect the mergeable list */
if ((old_cls->flags & H5FS_CLS_SEPAR_OBJ) != (new_cls->flags & H5FS_CLS_SEPAR_OBJ)) {
- hbool_t to_mergable; /* Flag if the section is changing to a mergable section */
+ hbool_t to_mergable; /* Flag if the section is changing to a mergeable section */
- /* Determine if this section is becoming mergable or is becoming separate */
+ /* Determine if this section is becoming mergeable or is becoming separate */
if (old_cls->flags & H5FS_CLS_SEPAR_OBJ)
to_mergable = TRUE;
else
diff --git a/src/H5Fefc.c b/src/H5Fefc.c
index 3f9a22f..e883717 100644
--- a/src/H5Fefc.c
+++ b/src/H5Fefc.c
@@ -450,7 +450,7 @@ H5F__efc_release_real(H5F_efc_t *efc)
/* Sanity checks */
HDassert(efc);
- /* Lock the EFC to prevent manipulation of the EFC wile we are releasing it.
+ /* Lock the EFC to prevent manipulation of the EFC while we are releasing it.
* The EFC should never be locked when we enter this function because that
* would require a cycle, a cycle would necessarily invoke
* H5F__efc_try_close(), and that function checks the status of the lock
diff --git a/src/H5Fint.c b/src/H5Fint.c
index a8039b2..9b56139 100644
--- a/src/H5Fint.c
+++ b/src/H5Fint.c
@@ -1060,7 +1060,7 @@ H5F__is_hdf5(const char *name, hid_t fapl_id)
{
H5FD_t * file = NULL; /* Low-level file struct */
H5F_shared_t *shared = NULL; /* Shared part of file */
- haddr_t sig_addr = HADDR_UNDEF; /* Addess of hdf5 file signature */
+ haddr_t sig_addr = HADDR_UNDEF; /* Address of hdf5 file signature */
htri_t ret_value = FAIL; /* Return value */
FUNC_ENTER_PACKAGE
@@ -1157,7 +1157,7 @@ H5F__new(H5F_shared_t *shared, unsigned flags, hid_t fcpl_id, hid_t fapl_id, H5F
/* Initialization for handling file space (for paged aggregation) */
f->shared->pgend_meta_thres = H5F_FILE_SPACE_PGEND_META_THRES;
- /* intialize point of no return */
+ /* initialize point of no return */
f->shared->point_of_no_return = FALSE;
/* Copy the file creation and file access property lists into the
@@ -1294,7 +1294,7 @@ H5F__new(H5F_shared_t *shared, unsigned flags, hid_t fcpl_id, hid_t fapl_id, H5F
f->shared->read_attempts = H5F_METADATA_READ_ATTEMPTS;
}
- /* Determine the # of bins for metdata read retries */
+ /* Determine the # of bins for metadata read retries */
if (H5F_set_retries(f) < 0)
HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, NULL, "can't set retries and retries_nbins")
@@ -1819,7 +1819,7 @@ H5F_open(const char *name, unsigned flags, hid_t fcpl_id, hid_t fapl_id)
* or creating it) so we can compare it with files that are already
* open. If that fails then we try again with the full set of flags
* (only if they're different than the original failed attempt).
- * However, if the file driver can't distinquish between files then
+ * However, if the file driver can't distinguish between files then
* there's no reason to open the file tentatively because it's the
* application's responsibility to prevent this situation (there's no
* way for us to detect it here anyway).
@@ -3116,7 +3116,7 @@ H5F__get_file_image(H5F_t *file, void *buf_ptr, size_t buf_len)
* file driver. However, this test will not work if there is some
* other file driver sitting on top of the multi file driver.
*
- * I'm not sure if this is possible at present, but in all likelyhood,
+ * I'm not sure if this is possible at present, but in all likelihood,
* it will become possible in the future. On the other hand, we may
* remove the split/multi file drivers before then.
*
@@ -3570,7 +3570,7 @@ done:
* 1) The file being opened has v3 superblock
* 2) The file is opened with H5F_ACC_RDWR
* 3) The file is not already marked for SWMR writing
- * 4) Current implementaion for opened objects:
+ * 4) Current implementation for opened objects:
* --only allow datasets and groups without attributes
* --disallow named datatype with/without attributes
* --disallow opened attributes attached to objects
diff --git a/src/H5Fmount.c b/src/H5Fmount.c
index 06f2a63..5c0eb27 100644
--- a/src/H5Fmount.c
+++ b/src/H5Fmount.c
@@ -104,7 +104,7 @@ H5F__mount(H5G_loc_t *loc, const char *name, H5F_t *child, hid_t H5_ATTR_UNUSED
H5F_t * parent = NULL; /*file containing mount point */
unsigned lt, rt, md; /*binary search indices */
int cmp; /*binary search comparison value*/
- H5G_loc_t mp_loc; /* entry of moint point to be opened */
+ H5G_loc_t mp_loc; /* entry of mount point to be opened */
H5G_name_t mp_path; /* Mount point group hier. path */
H5O_loc_t mp_oloc; /* Mount point object location */
H5G_loc_t root_loc; /* Group location of root of file to mount */
@@ -489,7 +489,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Funmount
*
- * Purpose: Given a mount point, dissassociate the mount point's file
+ * Purpose: Given a mount point, disassociate the mount point's file
* from the file mounted there. Do not close either file.
*
* The mount point can either be the group in the parent or the
diff --git a/src/H5Fsuper.c b/src/H5Fsuper.c
index 4b4d5e0..60b045a 100644
--- a/src/H5Fsuper.c
+++ b/src/H5Fsuper.c
@@ -374,7 +374,7 @@ H5F__super_read(H5F_t *f, H5P_genplist_t *fa_plist, hbool_t initial_read)
if (0 == mpi_rank) {
herr_t status;
- /* Try detecting file's siganture */
+ /* Try detecting file's signature */
/* (Don't leave before Bcast, to avoid hang on error) */
H5E_BEGIN_TRY
{
@@ -466,7 +466,7 @@ H5F__super_read(H5F_t *f, H5P_genplist_t *fa_plist, hbool_t initial_read)
*
* After upgrading low_bound, the library will check to ensure that the
* superblock version does not exceed the version allowed by high_bound.
- * Otherise fail file open.
+ * Otherwise fail file open.
*
* For details, please see RFC:Setting Bounds for Object Creation in HDF5 1.10.0.
*/
diff --git a/src/H5Fsuper_cache.c b/src/H5Fsuper_cache.c
index 28ff82f..b9dc74a 100644
--- a/src/H5Fsuper_cache.c
+++ b/src/H5Fsuper_cache.c
@@ -698,7 +698,7 @@ H5F__cache_superblock_serialize(const H5F_t *f, void *_image, size_t H5_ATTR_UNU
HGOTO_ERROR(H5E_RESOURCE, H5E_CANTGET, FAIL, "driver get_eoa request failed")
H5F_addr_encode(f, &image, (rel_eof + sblock->base_addr));
- /* Encode the driver informaton block address */
+ /* Encode the driver information block address */
H5F_addr_encode(f, &image, sblock->driver_addr);
/* Encode the root group object entry, including the cached stab info */
@@ -797,7 +797,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5F__cache_drvrinfo_get_initial_load_size
*
- * Purpose: Compute the intiial size of the data structure on disk.
+ * Purpose: Compute the initial size of the data structure on disk.
*
* Return: Non-negative on success/Negative on failure
*
diff --git a/src/H5Gcache.c b/src/H5Gcache.c
index 4713193..65ed1e6 100644
--- a/src/H5Gcache.c
+++ b/src/H5Gcache.c
@@ -134,7 +134,7 @@ H5G__cache_node_get_initial_load_size(void *_udata, size_t *image_len)
* Function: H5G__cache_node_deserialize
*
* Purpose: Given a buffer containing the on disk image of a symbol table
- * node, allocate an instance of H5G_node_t, load the contence of the
+ * node, allocate an instance of H5G_node_t, load the contents of the
* image into it, and return a pointer to the instance.
*
* Note that deserializing the image requires access to the file
@@ -240,7 +240,7 @@ H5G__cache_node_image_len(const void *_thing, size_t *image_len)
/*-------------------------------------------------------------------------
* Function: H5G__cache_node_serialize
*
- * Purpose: Given a correctly sized buffer and an instace of H5G_node_t,
+ * Purpose: Given a correctly sized buffer and an instance of H5G_node_t,
* serialize the contents of the instance of H5G_node_t, and write
* this data into the supplied buffer. This buffer will be written
* to disk.
diff --git a/src/H5Gcompact.c b/src/H5Gcompact.c
index af188d8..e13f10f 100644
--- a/src/H5Gcompact.c
+++ b/src/H5Gcompact.c
@@ -419,7 +419,7 @@ done:
* Function: H5G__compact_lookup_cb
*
* Purpose: Callback routine for searching 'link' messages for a particular
- * name & gettting object location for it
+ * name & getting object location for it
*
* Return: SUCCEED/FAIL
*
diff --git a/src/H5Gnode.c b/src/H5Gnode.c
index 9a7da0f..c0c9903 100644
--- a/src/H5Gnode.c
+++ b/src/H5Gnode.c
@@ -1206,7 +1206,7 @@ H5G__node_copy(H5F_t *f, const void H5_ATTR_UNUSED *_lt_key, haddr_t addr, const
&(sn->entry[i]); /* Convenience variable to refer to current source group entry */
H5O_link_t lnk; /* Link to insert */
const char * name; /* Name of source object */
- H5G_entry_t tmp_src_ent; /* Temperary copy. Change will not affect the cache */
+ H5G_entry_t tmp_src_ent; /* Temporary copy. Change will not affect the cache */
H5O_type_t obj_type = H5O_TYPE_UNKNOWN; /* Target object type */
H5G_copy_file_ud_t *cpy_udata; /* Copy file udata */
H5G_obj_create_t gcrt_info; /* Group creation info */
diff --git a/src/H5HFcache.c b/src/H5HFcache.c
index 7b5078e..5d5c7a0 100644
--- a/src/H5HFcache.c
+++ b/src/H5HFcache.c
@@ -796,7 +796,7 @@ H5HF__cache_hdr_serialize(const H5F_t *f, void *_image, size_t H5_ATTR_NDEBUG_UN
/* Encode I/O filter information */
if (H5O_msg_encode(hdr->f, H5O_PLINE_ID, FALSE, image, &(hdr->pline)) < 0)
- HGOTO_ERROR(H5E_HEAP, H5E_CANTENCODE, FAIL, "can't encode I/O pipeline fiters")
+ HGOTO_ERROR(H5E_HEAP, H5E_CANTENCODE, FAIL, "can't encode I/O pipeline filters")
image += hdr->filter_len;
} /* end if */
diff --git a/src/H5HFdbg.c b/src/H5HFdbg.c
index a244035..cbdf340 100644
--- a/src/H5HFdbg.c
+++ b/src/H5HFdbg.c
@@ -50,7 +50,7 @@
/* User data for direct block debugging iterator callback */
typedef struct {
FILE * stream; /* Stream for output */
- int indent; /* Indention amount */
+ int indent; /* Indentation amount */
int fwidth; /* Field width mount */
haddr_t dblock_addr; /* Direct block's address */
hsize_t dblock_size; /* Direct block's size */
@@ -63,7 +63,7 @@ typedef struct {
typedef struct {
H5FS_t *fspace; /* Free space manager */
FILE * stream; /* Stream for output */
- int indent; /* Indention amount */
+ int indent; /* Indentation amount */
int fwidth; /* Field width mount */
} H5HF_debug_iter_ud2_t;
diff --git a/src/H5HFdblock.c b/src/H5HFdblock.c
index 628dcc4..cf0ba34 100644
--- a/src/H5HFdblock.c
+++ b/src/H5HFdblock.c
@@ -432,7 +432,7 @@ H5HF__man_dblock_protect(H5HF_hdr_t *hdr, haddr_t dblock_addr, size_t dblock_siz
H5HF_indirect_t *par_iblock, unsigned par_entry, unsigned flags)
{
H5HF_direct_t * dblock; /* Direct block from cache */
- H5HF_dblock_cache_ud_t udata; /* parent and other infor for deserializing direct block */
+ H5HF_dblock_cache_ud_t udata; /* parent and other info for deserializing direct block */
H5HF_direct_t * ret_value = NULL; /* Return value */
FUNC_ENTER_PACKAGE
diff --git a/src/H5HFspace.c b/src/H5HFspace.c
index d62267c..67f9b6f 100644
--- a/src/H5HFspace.c
+++ b/src/H5HFspace.c
@@ -383,7 +383,7 @@ H5HF__space_create_root(const H5HF_hdr_t *hdr, H5HF_indirect_t *root_iblock)
/* Only need to scan the sections if the free space has been initialized */
if (hdr->fspace)
- /* Iterate over all sections, seting the parent pointers in 'single' sections to the new indirect
+ /* Iterate over all sections, setting the parent pointers in 'single' sections to the new indirect
* block */
if (H5FS_sect_iterate(hdr->f, hdr->fspace, H5HF__space_create_root_cb, root_iblock) < 0)
HGOTO_ERROR(H5E_FSPACE, H5E_BADITER, FAIL, "can't iterate over sections to set parent pointers")
diff --git a/src/H5I.c b/src/H5I.c
index efb4fd5..67a3786 100644
--- a/src/H5I.c
+++ b/src/H5I.c
@@ -779,7 +779,7 @@ H5I__iterate_pub_cb(void H5_ATTR_UNUSED *obj, hid_t id, void *_udata)
/*-------------------------------------------------------------------------
* Function: H5Iiterate
*
- * Purpose: Call the callback funciton op for each member of the id
+ * Purpose: Call the callback function op for each member of the id
* type type. op takes as parameters the id and a
* passthrough of op_data, and returns an herr_t. A positive
* return from op will cause the iteration to stop and
diff --git a/src/H5MFaggr.c b/src/H5MFaggr.c
index 45ddee2..ccf59f9 100644
--- a/src/H5MFaggr.c
+++ b/src/H5MFaggr.c
@@ -196,7 +196,7 @@ H5MF__aggr_alloc(H5F_t *f, H5F_blk_aggr_t *aggr, H5F_blk_aggr_t *other_aggr, H5F
haddr_t aggr_frag_addr = HADDR_UNDEF; /* Address of aggregrator fragment */
hsize_t aggr_frag_size = 0; /* Size of aggregator fragment */
hsize_t alignment; /* Alignment of this section */
- hsize_t aggr_mis_align = 0; /* Mis-alignment of aggregator */
+ hsize_t aggr_mis_align = 0; /* Misalignment of aggregator */
H5FD_mem_t alloc_type, other_alloc_type; /* Current aggregator & 'other' aggregator types */
#ifdef H5MF_AGGR_DEBUG
diff --git a/src/H5MFdbg.c b/src/H5MFdbg.c
index 7b89fd8..8064f12 100644
--- a/src/H5MFdbg.c
+++ b/src/H5MFdbg.c
@@ -50,7 +50,7 @@
typedef struct {
H5FS_t *fspace; /* Free space manager */
FILE * stream; /* Stream for output */
- int indent; /* Indention amount */
+ int indent; /* Indentation amount */
int fwidth; /* Field width amount */
} H5MF_debug_iter_ud_t;
diff --git a/src/H5Oainfo.c b/src/H5Oainfo.c
index 10c9ede..03ea3df 100644
--- a/src/H5Oainfo.c
+++ b/src/H5Oainfo.c
@@ -52,7 +52,7 @@ const H5O_msg_class_t H5O_MSG_AINFO[1] = {{
H5O_AINFO_ID, /*message id number */
"ainfo", /*message name for debugging */
sizeof(H5O_ainfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__ainfo_decode, /*decode message */
H5O__ainfo_encode, /*encode message */
H5O__ainfo_copy, /*copy the native value */
diff --git a/src/H5Oattr.c b/src/H5Oattr.c
index db53eaf..83c0370 100644
--- a/src/H5Oattr.c
+++ b/src/H5Oattr.c
@@ -65,7 +65,7 @@ const H5O_msg_class_t H5O_MSG_ATTR[1] = {{
H5O_ATTR_ID, /* message id number */
"attribute", /* message name for debugging */
sizeof(H5A_t), /* native message size */
- H5O_SHARE_IS_SHARABLE, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE, /* messages are shareable? */
H5O__attr_shared_decode, /* decode message */
H5O__attr_shared_encode, /* encode message */
H5O__attr_copy, /* copy the native value */
diff --git a/src/H5Oattribute.c b/src/H5Oattribute.c
index b21a654..371bbbc 100644
--- a/src/H5Oattribute.c
+++ b/src/H5Oattribute.c
@@ -239,19 +239,19 @@ H5O__attr_create(const H5O_loc_t *loc, H5A_t *attr)
/* Check if switching to "dense" attribute storage is possible */
if (!H5F_addr_defined(ainfo.fheap_addr)) {
- htri_t sharable; /* Whether the attribute will be shared */
+ htri_t shareable; /* Whether the attribute will be shared */
size_t raw_size = 0; /* Raw size of message */
- /* Check for attribute being sharable */
- if ((sharable = H5SM_can_share(loc->file, NULL, NULL, H5O_ATTR_ID, attr)) < 0)
+ /* Check for attribute being shareable */
+ if ((shareable = H5SM_can_share(loc->file, NULL, NULL, H5O_ATTR_ID, attr)) < 0)
HGOTO_ERROR(H5E_ATTR, H5E_BADMESG, FAIL, "can't determine attribute sharing status")
- else if (sharable == FALSE) {
+ else if (shareable == FALSE) {
/* Compute the size needed to encode the attribute */
raw_size = (H5O_MSG_ATTR->raw_size)(loc->file, FALSE, attr);
} /* end if */
/* Check for condititions for switching to "dense" attribute storage are met */
- if (ainfo.nattrs == oh->max_compact || (!sharable && raw_size >= H5O_MESG_MAX_SIZE)) {
+ if (ainfo.nattrs == oh->max_compact || (!shareable && raw_size >= H5O_MESG_MAX_SIZE)) {
H5O_iter_cvt_t udata; /* User data for callback */
H5O_mesg_operator_t op; /* Wrapper for operator */
diff --git a/src/H5Obogus.c b/src/H5Obogus.c
index c7fcdb9..ea231d6 100644
--- a/src/H5Obogus.c
+++ b/src/H5Obogus.c
@@ -47,7 +47,7 @@ const H5O_msg_class_t H5O_MSG_BOGUS_VALID[1] = {{
H5O_BOGUS_VALID_ID, /*message id number */
"bogus valid", /*message name for debugging */
0, /*native message size */
- H5O_SHARE_IS_SHARABLE, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE, /* messages are shareable? */
H5O__bogus_decode, /*decode message */
H5O__bogus_encode, /*encode message */
NULL, /*copy the native value */
@@ -71,7 +71,7 @@ const H5O_msg_class_t H5O_MSG_BOGUS_INVALID[1] = {{
H5O_BOGUS_INVALID_ID, /*message id number */
"bogus invalid", /*message name for debugging */
0, /*native message size */
- H5O_SHARE_IS_SHARABLE, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE, /* messages are shareable? */
H5O__bogus_decode, /*decode message */
H5O__bogus_encode, /*encode message */
NULL, /*copy the native value */
diff --git a/src/H5Obtreek.c b/src/H5Obtreek.c
index a783366..1eaf3e5 100644
--- a/src/H5Obtreek.c
+++ b/src/H5Obtreek.c
@@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_BTREEK[1] = {{
H5O_BTREEK_ID, /*message id number */
"v1 B-tree 'K' values", /*message name for debugging */
sizeof(H5O_btreek_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__btreek_decode, /*decode message */
H5O__btreek_encode, /*encode message */
H5O__btreek_copy, /*copy the native value */
diff --git a/src/H5Ocache.c b/src/H5Ocache.c
index 3aae0cb..ba47da3 100644
--- a/src/H5Ocache.c
+++ b/src/H5Ocache.c
@@ -275,7 +275,7 @@ H5O__cache_verify_chksum(const void *_image, size_t len, void *_udata)
*
* Note that the object header is read with with a speculative read.
* If the initial read is too small, make note of this fact and return
- * without error. H5C__load_entry() will note the size discrepency
+ * without error. H5C__load_entry() will note the size discrepancy
* and retry the deserialize operation with the correct size read.
*
* Return: Success: Pointer to in core representation
diff --git a/src/H5Ocache_image.c b/src/H5Ocache_image.c
index 2603f71..8f60a0e 100644
--- a/src/H5Ocache_image.c
+++ b/src/H5Ocache_image.c
@@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_MDCI[1] = {{
H5O_MDCI_MSG_ID, /* message id number */
"mdci", /* message name for debugging */
sizeof(H5O_mdci_t), /* native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__mdci_decode, /* decode message */
H5O__mdci_encode, /* encode message */
H5O__mdci_copy, /* copy method */
diff --git a/src/H5Ocont.c b/src/H5Ocont.c
index 080225a..9dd04fe 100644
--- a/src/H5Ocont.c
+++ b/src/H5Ocont.c
@@ -47,7 +47,7 @@ const H5O_msg_class_t H5O_MSG_CONT[1] = {{
H5O_CONT_ID, /*message id number */
"hdr continuation", /*message name for debugging */
sizeof(H5O_cont_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__cont_decode, /*decode message */
H5O__cont_encode, /*encode message */
NULL, /*no copy method */
diff --git a/src/H5Ocopy.c b/src/H5Ocopy.c
index 6cf402b..ce80eaa 100644
--- a/src/H5Ocopy.c
+++ b/src/H5Ocopy.c
@@ -1094,7 +1094,7 @@ H5O__copy_header(const H5O_loc_t *oloc_src, H5O_loc_t *oloc_dst /*out */, hid_t
if (H5P_get(ocpy_plist, H5O_CPY_OPTION_NAME, &cpy_option) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get object copy flag")
- /* Retrieve the marge committed datatype list */
+ /* Retrieve the merge committed datatype list */
if (H5P_peek(ocpy_plist, H5O_CPY_MERGE_COMM_DT_LIST_NAME, &dt_list) < 0)
HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get merge committed datatype list")
diff --git a/src/H5Odrvinfo.c b/src/H5Odrvinfo.c
index 8bebc6d..69ab5b2 100644
--- a/src/H5Odrvinfo.c
+++ b/src/H5Odrvinfo.c
@@ -38,7 +38,7 @@ const H5O_msg_class_t H5O_MSG_DRVINFO[1] = {{
H5O_DRVINFO_ID, /*message id number */
"driver info", /*message name for debugging */
sizeof(H5O_drvinfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__drvinfo_decode, /*decode message */
H5O__drvinfo_encode, /*encode message */
H5O__drvinfo_copy, /*copy the native value */
diff --git a/src/H5Odtype.c b/src/H5Odtype.c
index d17e3d0..9af79f4 100644
--- a/src/H5Odtype.c
+++ b/src/H5Odtype.c
@@ -89,7 +89,7 @@ const H5O_msg_class_t H5O_MSG_DTYPE[1] = {{
H5O_DTYPE_ID, /* message id number */
"datatype", /* message name for debugging */
sizeof(H5T_t), /* native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__dtype_shared_decode, /* decode message */
H5O__dtype_shared_encode, /* encode message */
H5O__dtype_copy, /* copy the native value */
diff --git a/src/H5Oefl.c b/src/H5Oefl.c
index d950249..931fe0b 100644
--- a/src/H5Oefl.c
+++ b/src/H5Oefl.c
@@ -41,7 +41,7 @@ const H5O_msg_class_t H5O_MSG_EFL[1] = {{
H5O_EFL_ID, /*message id number */
"external file list", /*message name for debugging */
sizeof(H5O_efl_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__efl_decode, /*decode message */
H5O__efl_encode, /*encode message */
H5O__efl_copy, /*copy native value */
diff --git a/src/H5Ofill.c b/src/H5Ofill.c
index 2314a84..093eba7 100644
--- a/src/H5Ofill.c
+++ b/src/H5Ofill.c
@@ -107,7 +107,7 @@ const H5O_msg_class_t H5O_MSG_FILL[1] = {{
H5O_FILL_ID, /*message id number */
"fill", /*message name for debugging */
sizeof(H5O_fill_t), /*native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__fill_shared_decode, /*decode message */
H5O__fill_shared_encode, /*encode message */
H5O__fill_copy, /*copy the native value */
@@ -131,7 +131,7 @@ const H5O_msg_class_t H5O_MSG_FILL_NEW[1] = {{
H5O_FILL_NEW_ID, /*message id number */
"fill_new", /*message name for debugging */
sizeof(H5O_fill_t), /*native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__fill_new_shared_decode, /*decode message */
H5O__fill_new_shared_encode, /*encode message */
H5O__fill_copy, /*copy the native value */
diff --git a/src/H5Ofsinfo.c b/src/H5Ofsinfo.c
index 0ab5abb..9f6514a 100644
--- a/src/H5Ofsinfo.c
+++ b/src/H5Ofsinfo.c
@@ -44,7 +44,7 @@ const H5O_msg_class_t H5O_MSG_FSINFO[1] = {{
H5O_FSINFO_ID, /* message id number */
"fsinfo", /* message name for debugging */
sizeof(H5O_fsinfo_t), /* native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__fsinfo_decode, /* decode message */
H5O__fsinfo_encode, /* encode message */
H5O__fsinfo_copy, /* copy the native value */
@@ -175,7 +175,7 @@ H5O__fsinfo_decode(H5F_t *f, H5O_t H5_ATTR_UNUSED *open_oh, unsigned H5_ATTR_UNU
H5F_DECODE_LENGTH(f, p, fsinfo->threshold); /* Free-space section threshold */
H5F_DECODE_LENGTH(f, p, fsinfo->page_size); /* File space page size */
- UINT16DECODE(p, fsinfo->pgend_meta_thres); /* Page end metdata threshold */
+ UINT16DECODE(p, fsinfo->pgend_meta_thres); /* Page end metadata threshold */
H5F_addr_decode(f, &p,
&(fsinfo->eoa_pre_fsm_fsalloc)); /* EOA before free-space header and section info */
diff --git a/src/H5Oginfo.c b/src/H5Oginfo.c
index 8540cc4..304890c 100644
--- a/src/H5Oginfo.c
+++ b/src/H5Oginfo.c
@@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_GINFO[1] = {{
H5O_GINFO_ID, /*message id number */
"ginfo", /*message name for debugging */
sizeof(H5O_ginfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__ginfo_decode, /*decode message */
H5O__ginfo_encode, /*encode message */
H5O__ginfo_copy, /*copy the native value */
diff --git a/src/H5Olayout.c b/src/H5Olayout.c
index 651e317..c939e72 100644
--- a/src/H5Olayout.c
+++ b/src/H5Olayout.c
@@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_LAYOUT[1] = {{
H5O_LAYOUT_ID, /* message id number */
"layout", /* message name for debugging */
sizeof(H5O_layout_t), /* native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__layout_decode, /* decode message */
H5O__layout_encode, /* encode message */
H5O__layout_copy, /* copy the native value */
diff --git a/src/H5Olinfo.c b/src/H5Olinfo.c
index 711f199..eacc916 100644
--- a/src/H5Olinfo.c
+++ b/src/H5Olinfo.c
@@ -53,7 +53,7 @@ const H5O_msg_class_t H5O_MSG_LINFO[1] = {{
H5O_LINFO_ID, /*message id number */
"linfo", /*message name for debugging */
sizeof(H5O_linfo_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__linfo_decode, /*decode message */
H5O__linfo_encode, /*encode message */
H5O__linfo_copy, /*copy the native value */
diff --git a/src/H5Olink.c b/src/H5Olink.c
index 75456f4..51c44a3 100644
--- a/src/H5Olink.c
+++ b/src/H5Olink.c
@@ -56,7 +56,7 @@ const H5O_msg_class_t H5O_MSG_LINK[1] = {{
H5O_LINK_ID, /*message id number */
"link", /*message name for debugging */
sizeof(H5O_link_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__link_decode, /*decode message */
H5O__link_encode, /*encode message */
H5O__link_copy, /*copy the native value */
diff --git a/src/H5Omessage.c b/src/H5Omessage.c
index 15edf1e..fa20aa1 100644
--- a/src/H5Omessage.c
+++ b/src/H5Omessage.c
@@ -371,7 +371,7 @@ H5O__msg_write_real(H5F_t *f, H5O_t *oh, const H5O_msg_class_t *type, unsigned m
HDassert(((H5O_shared_t *)idx_msg->native)->type != H5O_SHARE_TYPE_COMMITTED);
/* Also, sanity check that a message doesn't switch status from being
- * shared (or sharable) to being unsharable. (Which could cause
+ * shared (or shareable) to being unshareable. (Which could cause
* a message to increase in size in the object header)
*/
HDassert(!(mesg_flags & H5O_MSG_FLAG_DONTSHARE));
@@ -1131,7 +1131,7 @@ done:
* Nov 19 2004
*
* Description:
- * This function interates over the object headers of an object
+ * This function iterates over the object headers of an object
* specified with 'loc' of type 'type_id'. For each object header of the
* object, the 'op_data' and some additional information (specified below) are
* passed to the 'op' function.
@@ -1194,7 +1194,7 @@ done:
* Sep 6 2005
*
* Description:
- * This function interates over the object headers of an object
+ * This function iterates over the object headers of an object
* specified with 'ent' of type 'type_id'. For each object header of the
* object, the 'op_data' and some additional information (specified below) are
* passed to the 'op' function.
@@ -1544,7 +1544,7 @@ H5O_msg_is_shared(unsigned type_id, const void *mesg)
HDassert(type);
HDassert(mesg);
- /* If messages in a class aren't sharable, then obviously this message isn't shared! :-) */
+ /* If messages in a class aren't shareable, then obviously this message isn't shared! :-) */
if (type->share_flags & H5O_SHARE_IS_SHARABLE)
ret_value = H5O_IS_STORED_SHARED(((const H5O_shared_t *)mesg)->type);
else
diff --git a/src/H5Omtime.c b/src/H5Omtime.c
index d48d226..dfc881e 100644
--- a/src/H5Omtime.c
+++ b/src/H5Omtime.c
@@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_MTIME[1] = {{
H5O_MTIME_ID, /*message id number */
"mtime", /*message name for debugging */
sizeof(time_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__mtime_decode, /*decode message */
H5O__mtime_encode, /*encode message */
H5O__mtime_copy, /*copy the native value */
@@ -68,7 +68,7 @@ const H5O_msg_class_t H5O_MSG_MTIME_NEW[1] = {{
H5O_MTIME_NEW_ID, /*message id number */
"mtime_new", /*message name for debugging */
sizeof(time_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__mtime_new_decode, /*decode message */
H5O__mtime_new_encode, /*encode message */
H5O__mtime_copy, /*copy the native value */
diff --git a/src/H5Oname.c b/src/H5Oname.c
index 979b240..ab263a2 100644
--- a/src/H5Oname.c
+++ b/src/H5Oname.c
@@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_NAME[1] = {{
H5O_NAME_ID, /*message id number */
"name", /*message name for debugging */
sizeof(H5O_name_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__name_decode, /*decode message */
H5O__name_encode, /*encode message */
H5O__name_copy, /*copy the native value */
diff --git a/src/H5Onull.c b/src/H5Onull.c
index 9b377fb..0f3143c 100644
--- a/src/H5Onull.c
+++ b/src/H5Onull.c
@@ -31,7 +31,7 @@ const H5O_msg_class_t H5O_MSG_NULL[1] = {{
H5O_NULL_ID, /*message id number */
"null", /*message name for debugging */
0, /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
NULL, /*no decode method */
NULL, /*no encode method */
NULL, /*no copy method */
diff --git a/src/H5Opline.c b/src/H5Opline.c
index 4b76da9..ffc4557 100644
--- a/src/H5Opline.c
+++ b/src/H5Opline.c
@@ -67,7 +67,7 @@ const H5O_msg_class_t H5O_MSG_PLINE[1] = {{
H5O_PLINE_ID, /* message id number */
"filter pipeline", /* message name for debugging */
sizeof(H5O_pline_t), /* native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__pline_shared_decode, /* decode message */
H5O__pline_shared_encode, /* encode message */
H5O__pline_copy, /* copy the native value */
@@ -86,7 +86,7 @@ const H5O_msg_class_t H5O_MSG_PLINE[1] = {{
H5O__pline_shared_debug /* debug the message */
}};
-/* Format version bounds for filter pipleline */
+/* Format version bounds for filter pipeline */
const unsigned H5O_pline_ver_bounds[] = {
H5O_PLINE_VERSION_1, /* H5F_LIBVER_EARLIEST */
H5O_PLINE_VERSION_2, /* H5F_LIBVER_V18 */
diff --git a/src/H5Orefcount.c b/src/H5Orefcount.c
index 52eee14..500ab0f 100644
--- a/src/H5Orefcount.c
+++ b/src/H5Orefcount.c
@@ -45,7 +45,7 @@ const H5O_msg_class_t H5O_MSG_REFCOUNT[1] = {{
H5O_REFCOUNT_ID, /*message id number */
"refcount", /*message name for debugging */
sizeof(H5O_refcount_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__refcount_decode, /*decode message */
H5O__refcount_encode, /*encode message */
H5O__refcount_copy, /*copy the native value */
diff --git a/src/H5Osdspace.c b/src/H5Osdspace.c
index dab989f..574103d 100644
--- a/src/H5Osdspace.c
+++ b/src/H5Osdspace.c
@@ -61,7 +61,7 @@ const H5O_msg_class_t H5O_MSG_SDSPACE[1] = {{
H5O_SDSPACE_ID, /* message id number */
"dataspace", /* message name for debugging */
sizeof(H5S_extent_t), /* native message size */
- H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */
+ H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */
H5O__sdspace_shared_decode, /* decode message */
H5O__sdspace_shared_encode, /* encode message */
H5O__sdspace_copy, /* copy the native value */
diff --git a/src/H5Oshmesg.c b/src/H5Oshmesg.c
index 1c894a5..a4e8b8f 100644
--- a/src/H5Oshmesg.c
+++ b/src/H5Oshmesg.c
@@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_SHMESG[1] = {{
H5O_SHMESG_ID, /*message id number */
"shared message table", /*message name for debugging */
sizeof(H5O_shmesg_table_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__shmesg_decode, /*decode message */
H5O__shmesg_encode, /*encode message */
H5O__shmesg_copy, /*copy the native value */
diff --git a/src/H5Ostab.c b/src/H5Ostab.c
index fa030bd..213b8a0 100644
--- a/src/H5Ostab.c
+++ b/src/H5Ostab.c
@@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_STAB[1] = {{
H5O_STAB_ID, /*message id number */
"stab", /*message name for debugging */
sizeof(H5O_stab_t), /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
H5O__stab_decode, /*decode message */
H5O__stab_encode, /*encode message */
H5O__stab_copy, /*copy the native value */
diff --git a/src/H5Ounknown.c b/src/H5Ounknown.c
index 2b8a7bb..dc43a6c 100644
--- a/src/H5Ounknown.c
+++ b/src/H5Ounknown.c
@@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_UNKNOWN[1] = {{
H5O_UNKNOWN_ID, /*message id number */
"unknown", /*message name for debugging */
0, /*native message size */
- 0, /* messages are sharable? */
+ 0, /* messages are shareable? */
NULL, /*decode message */
NULL, /*encode message */
NULL, /*copy the native value */
diff --git a/src/H5P.c b/src/H5P.c
index 45a9267..110565b 100644
--- a/src/H5P.c
+++ b/src/H5P.c
@@ -678,7 +678,7 @@ done:
NAME
H5Pexist
PURPOSE
- Routine to query the existance of a property in a property object.
+ Routine to query the existence of a property in a property object.
USAGE
htri_t H5P_exist(id, name)
hid_t id; IN: Property object ID to check
@@ -712,7 +712,7 @@ H5Pexist(hid_t id, const char *name)
if (!name || !*name)
HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property name");
- /* Check for the existance of the property in the list or class */
+ /* Check for the existence of the property in the list or class */
if (H5I_GENPROP_LST == H5I_get_type(id)) {
if (NULL == (plist = (H5P_genplist_t *)H5I_object(id)))
HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a property list");
@@ -858,7 +858,7 @@ done:
Failure: H5I_INVALID_HID (negative)
DESCRIPTION
Decodes a property list from a binary buffer. The contents of the buffer
- contain the values for the correponding properties of the plist. The decode
+ contain the values for the corresponding properties of the plist. The decode
callback of a certain property decodes its value from the buffer and sets it
in the property list.
GLOBAL VARIABLES
@@ -1299,7 +1299,7 @@ done:
Returns non-negative on success, negative on failure.
DESCRIPTION
Removes a property from a property list. Both properties which were
- in existance when the property list was created (i.e. properties registered
+ in existence when the property list was created (i.e. properties registered
with H5Pregister2) and properties added to the list after it was created
(i.e. added with H5Pinsert2) may be removed from a property list.
Properties do not need to be removed a property list before the list itself
diff --git a/src/H5PL.c b/src/H5PL.c
index caa9dc4..bbdef22 100644
--- a/src/H5PL.c
+++ b/src/H5PL.c
@@ -67,7 +67,7 @@
* the special "::" string.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -99,7 +99,7 @@ done:
* Positive if one or more of the plugin types are enabled
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -128,7 +128,7 @@ done:
* Purpose: Insert a plugin search path at the end of the list.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -160,7 +160,7 @@ done:
* Purpose: Insert a plugin search path at the beginning of the list.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -234,7 +234,7 @@ done:
* other paths after the index.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
*-------------------------------------------------------------------------
*/
@@ -270,11 +270,11 @@ done:
/*-------------------------------------------------------------------------
* Function: H5PLremove
*
- * Purpose: Remove the plugin path at the specifed index and compact
+ * Purpose: Remove the plugin path at the specified index and compact
* the list.
*
* Return: Success: Non-negative
- * Failture: Negative
+ * Failure: Negative
*
* Return: Non-negative or success.
*
diff --git a/src/H5PLpath.c b/src/H5PLpath.c
index d16d8b2..a2bc178 100644
--- a/src/H5PLpath.c
+++ b/src/H5PLpath.c
@@ -322,7 +322,7 @@ H5PL__close_path_table(void)
* Purpose: Gets the number of plugin paths that have been stored.
*
* Return: Success: The number of paths
- * Failture: Can't fail
+ * Failure: Can't fail
*-------------------------------------------------------------------------
*/
unsigned
diff --git a/src/H5Pdxpl.c b/src/H5Pdxpl.c
index 0638ffc..9349b04 100644
--- a/src/H5Pdxpl.c
+++ b/src/H5Pdxpl.c
@@ -1968,7 +1968,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Pget_mpio_actual_io_mode
*
- * Purpose: Retrieves the type of I/O actually preformed when collective I/O
+ * Purpose: Retrieves the type of I/O actually performed when collective I/O
* is requested.
*
* Return: Non-negative on success/Negative on failure
diff --git a/src/H5Pencdec.c b/src/H5Pencdec.c
index e2a97f8..01573d1 100644
--- a/src/H5Pencdec.c
+++ b/src/H5Pencdec.c
@@ -663,7 +663,7 @@ done:
on failure.
DESCRIPTION
Decodes a property list from a binary buffer. The contents of the buffer
- contain the values for the correponding properties of the plist. The decode
+ contain the values for the corresponding properties of the plist. The decode
callback of a certain property decodes its value from the buffer and sets it
in the property list.
GLOBAL VARIABLES
diff --git a/src/H5Pfapl.c b/src/H5Pfapl.c
index 2e345f1..9799734 100644
--- a/src/H5Pfapl.c
+++ b/src/H5Pfapl.c
@@ -500,7 +500,7 @@ static const size_t H5F_def_page_buf_size_g = H5F_ACS_PAGE_BUFFER_SIZE_DEF; /*
static const unsigned H5F_def_page_buf_min_meta_perc_g =
H5F_ACS_PAGE_BUFFER_MIN_META_PERC_DEF; /* Default page buffer minimum metadata size */
static const unsigned H5F_def_page_buf_min_raw_perc_g =
- H5F_ACS_PAGE_BUFFER_MIN_RAW_PERC_DEF; /* Default page buffer mininum raw data size */
+ H5F_ACS_PAGE_BUFFER_MIN_RAW_PERC_DEF; /* Default page buffer minimum raw data size */
static const hbool_t H5F_def_use_file_locking_g =
H5F_ACS_USE_FILE_LOCKING_DEF; /* Default use file locking flag */
static const hbool_t H5F_def_ignore_disabled_file_locks_g =
@@ -2347,7 +2347,7 @@ done:
*
* Currently, the only two valid combinations for this routine are:
* LOW = H5F_FORMAT_EARLIEST and HIGH = H5F_FORMAT_LATEST (the default
- * setting, which creates objects with the ealiest version possible for
+ * setting, which creates objects with the earliest version possible for
* each object, but no upper limit on the version allowed to be created if
* a newer version of an object's format is required to support a feature
* requested with an HDF5 library API routine), and LOW = H5F_FORMAT_LATEST
@@ -2363,7 +2363,7 @@ done:
* Note: Eventually we want to add more values to the H5F_libver_t
* enumerated type that indicate library release values where the file
* format was changed (like "H5F_FORMAT_1_2_0" for the file format changes
- * in the 1.2.x release branch and possily even "H5F_FORMAT_1_4_2" for
+ * in the 1.2.x release branch and possibly even "H5F_FORMAT_1_4_2" for
* a change mid-way through the 1.4.x release branch, etc).
*
* Adding more values will allow applications to make settings like the
@@ -2809,7 +2809,7 @@ H5Pset_file_image_callbacks(hid_t fapl_id, H5FD_file_image_callbacks_t *callback
HDassert(callbacks_ptr->udata_copy);
HDassert(callbacks_ptr->udata_free);
if ((info.callbacks.udata = callbacks_ptr->udata_copy(callbacks_ptr->udata)) == NULL)
- HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't copy the suppplied udata")
+ HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't copy the supplied udata")
} /* end if */
/* Set values */
diff --git a/src/H5Pint.c b/src/H5Pint.c
index f66e2b9..6907faf 100644
--- a/src/H5Pint.c
+++ b/src/H5Pint.c
@@ -3176,7 +3176,7 @@ done:
NAME
H5P_exist_plist
PURPOSE
- Internal routine to query the existance of a property in a property list.
+ Internal routine to query the existence of a property in a property list.
USAGE
htri_t H5P_exist_plist(plist, name)
const H5P_genplist_t *plist; IN: Property list to check
@@ -3235,7 +3235,7 @@ done:
NAME
H5P__exist_pclass
PURPOSE
- Internal routine to query the existance of a property in a property class.
+ Internal routine to query the existence of a property in a property class.
USAGE
herr_t H5P__exist_pclass(pclass, name)
H5P_genclass_t *pclass; IN: Property class to check
@@ -3751,7 +3751,7 @@ H5P__cmp_plist_cb(H5P_genprop_t *prop, void *_udata)
/* Check if the property exists in the second property list */
if ((prop2_exist = H5P_exist_plist(udata->plist2, prop->name)) < 0)
- HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, H5_ITER_ERROR, "can't lookup existance of property?")
+ HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, H5_ITER_ERROR, "can't lookup existence of property?")
if (prop2_exist) {
const H5P_genprop_t *prop2; /* Pointer to property in second plist */
@@ -4702,7 +4702,7 @@ done:
Returns non-negative on success, negative on failure.
DESCRIPTION
Removes a property from a property list. Both properties which were
- in existance when the property list was created (i.e. properties registered
+ in existence when the property list was created (i.e. properties registered
with H5Pregister2) and properties added to the list after it was created
(i.e. added with H5Pinsert2) may be removed from a property list.
Properties do not need to be removed a property list before the list itself
@@ -4783,7 +4783,7 @@ H5P__copy_prop_plist(hid_t dst_id, hid_t src_id, const char *name)
NULL == (dst_plist = (H5P_genplist_t *)H5I_object(dst_id)))
HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, FAIL, "property object doesn't exist")
- /* If the property exists in the destination alread */
+ /* If the property exists in the destination already */
if (NULL != H5P__find_prop_plist(dst_plist, name)) {
/* Delete the property from the destination list, calling the 'close' callback if necessary */
if (H5P_remove(dst_plist, name) < 0)
diff --git a/src/H5Pocpl.c b/src/H5Pocpl.c
index cc32582..5d39cf6 100644
--- a/src/H5Pocpl.c
+++ b/src/H5Pocpl.c
@@ -482,7 +482,7 @@ done:
* transfer property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -543,7 +543,7 @@ done:
* transfer property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -609,7 +609,7 @@ done:
* creation property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -671,7 +671,7 @@ done:
* creation property list. The FLAGS argument specifies certain
* general properties of the filter and is documented below.
* The CD_VALUES is an array of CD_NELMTS integers which are
- * auxiliary data for the filter. The integer vlues will be
+ * auxiliary data for the filter. The integer values will be
* stored in the dataset object header as part of the filter
* information.
*
@@ -782,7 +782,7 @@ done:
* CD_NELMTS indicates the number of entries in the CD_VALUES
* array allocated by the caller while on exit it contains the
* number of values defined by the filter. FILTER_CONFIG is a bit
- * field contaning encode/decode flags from H5Zpublic.h. The IDX
+ * field containing encode/decode flags from H5Zpublic.h. The IDX
* should be a value between zero and N-1 as described for
* H5Pget_nfilters() and the function will return failure if the
* filter number is out of range.
@@ -867,7 +867,7 @@ done:
* CD_NELMTS indicates the number of entries in the CD_VALUES
* array allocated by the caller while on exit it contains the
* number of values defined by the filter. FILTER_CONFIG is a bit
- * field contaning encode/decode flags from H5Zpublic.h. The ID
+ * field containing encode/decode flags from H5Zpublic.h. The ID
* should be the filter ID to retrieve the parameters for. If the
* filter is not set for the property list, an error will be returned.
*
@@ -916,7 +916,7 @@ done:
* CD_NELMTS indicates the number of entries in the CD_VALUES
* array allocated by the caller while on exit it contains the
* number of values defined by the filter. FILTER_CONFIG is a bit
- * field contaning encode/decode flags from H5Zpublic.h. The ID
+ * field containing encode/decode flags from H5Zpublic.h. The ID
* should be the filter ID to retrieve the parameters for. If the
* filter is not set for the property list, an error will be returned.
*
diff --git a/src/H5S.c b/src/H5S.c
index d45a6f1..d74aaac 100644
--- a/src/H5S.c
+++ b/src/H5S.c
@@ -118,7 +118,7 @@ H5S__init_package(void)
if (H5I_register_type(H5I_DATASPACE_CLS) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTINIT, FAIL, "unable to initialize dataspace ID class")
- /* Initialize the atom group for the dataspace selction iterator IDs */
+ /* Initialize the atom group for the dataspace selection iterator IDs */
if (H5I_register_type(H5I_SPACE_SEL_ITER_CLS) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTINIT, FAIL,
"unable to initialize dataspace selection iterator ID class")
diff --git a/src/H5SL.c b/src/H5SL.c
index 9d3f5b3..1a2ef22 100644
--- a/src/H5SL.c
+++ b/src/H5SL.c
@@ -289,7 +289,7 @@
}
/* Macro used to reduce the level of a node by 1. Does not update the head node
- * "current level". PREV is the previous node of the currrent height of X. */
+ * "current level". PREV is the previous node of the current height of X. */
#define H5SL_DEMOTE(X, PREV) \
{ \
size_t _lvl = X->level; \
@@ -305,7 +305,7 @@
#define H5SL_INSERT(CMP, SLIST, X, TYPE, KEY, HASHVAL) \
{ \
H5SL_node_t *_last = X; /* Lowest node in the current gap */ \
- H5SL_node_t *_next = NULL; /* Highest node in the currect gap */ \
+ H5SL_node_t *_next = NULL; /* Highest node in the current gap */ \
H5SL_node_t *_drop; /* Low node of the gap to drop into */ \
int _count; /* Number of nodes in the current gap */ \
int _i; \
@@ -365,7 +365,7 @@
else { \
H5SL_node_t *_last = X; /* Lowest node in the current gap */ \
H5SL_node_t *_llast = X; /* Lowest node in the previous gap */ \
- H5SL_node_t *_next = NULL; /* Highest node in the currect gap */ \
+ H5SL_node_t *_next = NULL; /* Highest node in the current gap */ \
H5SL_node_t *_drop = NULL; /* Low node of the gap to drop into */ \
H5SL_node_t *_ldrop = NULL; /* Low node of gap before the one to drop into */ \
H5SL_node_t *_head = SLIST->header; /* Head of the skip list */ \
@@ -2020,7 +2020,7 @@ H5SL_next(H5SL_node_t *slist_node)
NAME
H5SL_prev
PURPOSE
- Gets a pointer to the previos node in a skip list
+ Gets a pointer to the previous node in a skip list
USAGE
H5SL_node_t *H5SL_prev(slist_node)
H5SL_node_t *slist_node; IN: Pointer to skip list node
diff --git a/src/H5SM.c b/src/H5SM.c
index fa446cf..fb0df07 100644
--- a/src/H5SM.c
+++ b/src/H5SM.c
@@ -117,7 +117,7 @@ H5SM_init(H5F_t *f, H5P_genplist_t *fc_plist, const H5O_loc_t *ext_loc)
haddr_t table_addr = HADDR_UNDEF; /* Address of SOHM master table in file */
unsigned list_max, btree_min; /* Phase change limits for SOHM indices */
unsigned index_type_flags[H5O_SHMESG_MAX_NINDEXES]; /* Messages types stored in each index */
- unsigned minsizes[H5O_SHMESG_MAX_NINDEXES]; /* Message size sharing threshhold for each index */
+ unsigned minsizes[H5O_SHMESG_MAX_NINDEXES]; /* Message size sharing threshold for each index */
unsigned type_flags_used; /* Message type flags used, for sanity checking */
unsigned x; /* Local index variable */
herr_t ret_value = SUCCEED; /* Return value */
@@ -1602,7 +1602,7 @@ H5SM_delete(H5F_t *f, H5O_t *open_oh, H5O_shared_t *sh_mesg)
*/
if (H5SM__delete_from_index(f, open_oh, &(table->indexes[index_num]), sh_mesg, &cache_flags, &mesg_size,
&mesg_buf) < 0)
- HGOTO_ERROR(H5E_SOHM, H5E_CANTDELETE, FAIL, "unable to delete mesage from SOHM index")
+ HGOTO_ERROR(H5E_SOHM, H5E_CANTDELETE, FAIL, "unable to delete message from SOHM index")
/* Release the master SOHM table */
if (H5AC_unprotect(f, H5AC_SOHM_TABLE, H5F_SOHM_ADDR(f), table, cache_flags) < 0)
diff --git a/src/H5Shyper.c b/src/H5Shyper.c
index 8f593ea..70e9d54 100644
--- a/src/H5Shyper.c
+++ b/src/H5Shyper.c
@@ -2060,7 +2060,7 @@ H5S__hyper_iter_get_seq_list_opt(H5S_sel_iter_t *iter, size_t maxseq, size_t max
/* Increment the offset and count for the other dimensions */
temp_dim = (int)fast_dim - 1;
while (temp_dim >= 0) {
- /* Move to the next row in the curent dimension */
+ /* Move to the next row in the current dimension */
offset[temp_dim]++;
tmp_block[temp_dim]++;
@@ -2178,7 +2178,7 @@ H5S__hyper_iter_get_seq_list_opt(H5S_sel_iter_t *iter, size_t maxseq, size_t max
/* Increment the offset and count for the other dimensions */
temp_dim = (int)fast_dim - 1;
while (temp_dim >= 0) {
- /* Move to the next row in the curent dimension */
+ /* Move to the next row in the current dimension */
offset[temp_dim]++;
tmp_block[temp_dim]++;
@@ -2492,7 +2492,7 @@ H5S__hyper_iter_get_seq_list_single(H5S_sel_iter_t *iter, size_t maxseq, size_t
/* Increment the offset and count for the other dimensions */
while (temp_dim >= 0) {
- /* Move to the next row in the curent dimension */
+ /* Move to the next row in the current dimension */
offset[temp_dim]++;
tmp_block[temp_dim]++;
@@ -3584,7 +3584,7 @@ H5S__hyper_get_version_enc_size(const H5S_t *space, hsize_t block_count, uint32_
H5F_libver_t low_bound; /* The 'low' bound of library format versions */
H5F_libver_t high_bound; /* The 'high' bound of library format versions */
htri_t is_regular; /* A regular hyperslab or not */
- uint32_t tmp_version; /* Local temporay version */
+ uint32_t tmp_version; /* Local temporary version */
unsigned u; /* Local index variable */
herr_t ret_value = SUCCEED; /* Return value */
@@ -3972,7 +3972,7 @@ H5S__hyper_serialize(const H5S_t *space, uint8_t **p)
pp = (*p);
HDassert(pp);
- /* Set some convienence values */
+ /* Set some convenience values */
ndims = space->extent.rank;
diminfo = space->select.sel_info.hslab->diminfo.opt;
@@ -4075,7 +4075,7 @@ H5S__hyper_serialize(const H5S_t *space, uint8_t **p)
else {
HDassert(version == H5S_HYPER_VERSION_1);
- /* Set some convienence values */
+ /* Set some convenience values */
fast_dim = ndims - 1;
/* Encode number of hyperslabs */
@@ -4225,7 +4225,7 @@ H5S__hyper_deserialize(H5S_t **space, const uint8_t **p)
{
H5S_t *tmp_space = NULL; /* Pointer to actual dataspace to use,
either *space or a newly allocated one */
- hsize_t dims[H5S_MAX_RANK]; /* Dimenion sizes */
+ hsize_t dims[H5S_MAX_RANK]; /* Dimension sizes */
hsize_t start[H5S_MAX_RANK]; /* hyperslab start information */
hsize_t block[H5S_MAX_RANK]; /* hyperslab block information */
uint32_t version; /* Version number */
@@ -4618,7 +4618,7 @@ H5S__get_select_hyper_blocklist(H5S_t *space, hsize_t startblock, hsize_t numblo
hbool_t done; /* Whether we are done with the iteration */
unsigned u; /* Counter */
- /* Set some convienence values */
+ /* Set some convenience values */
ndims = space->extent.rank;
fast_dim = ndims - 1;
@@ -5112,7 +5112,7 @@ H5S__hyper_is_contiguous(const H5S_t *space)
large_contiguous = TRUE; /* assume true and reset if the dimensions don't match */
small_contiguous = FALSE; /* assume false initially */
- /* Check for a "large contigous" block */
+ /* Check for a "large contiguous" block */
for (u = 0; u < space->extent.rank; u++) {
if (diminfo[u].count > 1) {
large_contiguous = FALSE;
@@ -11516,7 +11516,7 @@ H5S__hyper_proj_int_iterate(const H5S_hyper_span_info_t *ss_span_info,
udata->skip += (count - 1) * (udata->skip - old_skip);
} /* end if */
else {
- /* Third case: agorithm added skip and nelem (in that
+ /* Third case: algorithm added skip and nelem (in that
* order). Add the same skip and nelem once for each item
* remaining in count. */
hsize_t skip_add;
diff --git a/src/H5Smpio.c b/src/H5Smpio.c
index 7b85209..df4c3c7 100644
--- a/src/H5Smpio.c
+++ b/src/H5Smpio.c
@@ -1145,7 +1145,7 @@ H5S__obtain_datatype(H5S_hyper_span_info_t *spans, const hsize_t *down, size_t e
/* If this is the fastest changing dimension, it is the base case for derived datatype. */
span = spans->head;
if (NULL == span->down) {
- hbool_t large_block = FALSE; /* Wether the block length is larger than 32 bit integer */
+ hbool_t large_block = FALSE; /* Whether the block length is larger than 32 bit integer */
outercount = 0;
while (span) {
diff --git a/src/H5Spoint.c b/src/H5Spoint.c
index 929499f..65aefe2 100644
--- a/src/H5Spoint.c
+++ b/src/H5Spoint.c
@@ -1504,7 +1504,7 @@ done:
dataspace they are located within. The list of points is formatted as
follows: <coordinate> followed by the next coordinate, etc. until all the
point information in the selection have been put into the user's buffer.
- The points are returned in the order they will be interated through
+ The points are returned in the order they will be iterated through
when a selection is read/written from/to disk.
GLOBAL VARIABLES
COMMENTS, BUGS, ASSUMPTIONS
@@ -1578,7 +1578,7 @@ H5S__get_select_elem_pointlist(const H5S_t *space, hsize_t startpoint, hsize_t n
dataspace they are located within. The list of points is formatted as
follows: <coordinate> followed by the next coordinate, etc. until all the
point information in the selection have been put into the user's buffer.
- The points are returned in the order they will be interated through
+ The points are returned in the order they will be iterated through
when a selection is read/written from/to disk.
GLOBAL VARIABLES
COMMENTS, BUGS, ASSUMPTIONS
diff --git a/src/H5Sselect.c b/src/H5Sselect.c
index 86e9043..241708a 100644
--- a/src/H5Sselect.c
+++ b/src/H5Sselect.c
@@ -2247,7 +2247,7 @@ H5S_select_construct_projection(const H5S_t *base_space, H5S_t **new_space_ptr,
if (new_space_rank == 0) {
hssize_t npoints; /* Number of points selected */
- /* Retreve the number of elements selected */
+ /* Retrieve the number of elements selected */
if ((npoints = (hssize_t)H5S_GET_SELECT_NPOINTS(base_space)) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_CANTGET, FAIL, "unable to get number of points selected")
HDassert(npoints <= 1);
@@ -2316,12 +2316,12 @@ H5S_select_construct_projection(const H5S_t *base_space, H5S_t **new_space_ptr,
* true on selections of different rank iff:
*
* 1) the selection in the lower rank dataspace matches that
- * in the dimensions with the fastest changing indicies in
+ * in the dimensions with the fastest changing indices in
* the larger rank dataspace, and
*
* 2) the selection has thickness 1 in all ranks that appear
* only in the higher rank dataspace (i.e. those with
- * more slowly changing indicies).
+ * more slowly changing indices).
*/
if (new_space_rank > base_space_rank) {
hsize_t tmp_dim_size = 1; /* Temporary dimension value, for filling arrays */
diff --git a/src/H5Stest.c b/src/H5Stest.c
index a851549..2782086 100644
--- a/src/H5Stest.c
+++ b/src/H5Stest.c
@@ -185,12 +185,12 @@ H5S__check_spans_tail_ptr(const H5S_hyper_span_info_t *span_lst)
if (NULL != cur_elem->down)
if ((ret_value = H5S__check_spans_tail_ptr(cur_elem->down)) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
- "the seletion has inconsistent tail pointers")
+ "the selection has inconsistent tail pointers")
cur_elem = cur_elem->next;
} /* end while */
if (actual_tail != span_lst->tail)
- HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the seletion has inconsistent tail pointers")
+ HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the selection has inconsistent tail pointers")
done:
FUNC_LEAVE_NOAPI(ret_value)
@@ -231,7 +231,7 @@ H5S__check_points_tail_ptr(const H5S_pnt_list_t *pnt_lst)
cur_elem = cur_elem->next;
} /* end while */
if (actual_tail != pnt_lst->tail)
- HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the seletion has inconsistent tail pointers")
+ HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the selection has inconsistent tail pointers")
done:
FUNC_LEAVE_NOAPI(ret_value)
@@ -314,7 +314,7 @@ H5S__check_internal_consistency(const H5S_t *space)
if ((NULL != hslab) && (NULL != hslab->span_lst))
if (H5S__check_spans_tail_ptr(hslab->span_lst) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
- "the seletion has inconsistent tail pointers")
+ "the selection has inconsistent tail pointers")
} /* end if */
else if (space->select.type->type == H5S_SEL_POINTS) {
H5S_pnt_list_t *pnt_lst = space->select.sel_info.pnt_lst;
@@ -322,7 +322,7 @@ H5S__check_internal_consistency(const H5S_t *space)
if (NULL != pnt_lst)
if (H5S__check_points_tail_ptr(pnt_lst) < 0)
HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL,
- "the seletion has inconsistent tail pointers")
+ "the selection has inconsistent tail pointers")
} /* end else-if */
done:
diff --git a/src/H5T.c b/src/H5T.c
index 8426040..3bb220a 100644
--- a/src/H5T.c
+++ b/src/H5T.c
@@ -4207,7 +4207,7 @@ done:
*
* Return: Success: non-negative
*
- * Failure: nagative
+ * Failure: negative
*
* Programmer: Robb Matzke
* Tuesday, December 22, 1998
diff --git a/src/H5TS.c b/src/H5TS.c
index 321a83d..72d158f 100644
--- a/src/H5TS.c
+++ b/src/H5TS.c
@@ -41,7 +41,7 @@
/* Local Typedefs */
/******************/
-/* Cancelability structure */
+/* Cancellability structure */
typedef struct H5TS_cancel_struct {
int previous_state;
unsigned int cancel_count;
@@ -548,9 +548,9 @@ H5TS_mutex_unlock(H5TS_mutex_t *mutex)
* Creates a cancellation counter for a thread if it is the first time
* the thread is entering the library.
*
- * if counter value is zero, then set cancelability type of the thread
+ * if counter value is zero, then set cancellability type of the thread
* to PTHREAD_CANCEL_DISABLE as thread is entering the library and store
- * the previous cancelability type into cancellation counter.
+ * the previous cancellability type into cancellation counter.
*
* Increase the counter value by 1.
*
@@ -613,8 +613,8 @@ H5TS_cancel_count_inc(void)
* Returns: SUCCEED/FAIL
*
* Description:
- * If counter value is one, then set cancelability type of the thread
- * to the previous cancelability type stored in the cancellation counter.
+ * If counter value is one, then set cancellability type of the thread
+ * to the previous cancellability type stored in the cancellation counter.
* (the thread is leaving the library).
*
* Decrement the counter value by 1.
@@ -658,7 +658,7 @@ H5TS_cancel_count_dec(void)
* H5TS_rec_entry_count, or NULL on failure.
*
* Description:
- * Allocate and initalize an instance of H5TS_rec_entry_count.
+ * Allocate and initialize an instance of H5TS_rec_entry_count.
*
* Programmer: John Mainzer
* August 28, 2020
@@ -824,7 +824,7 @@ H5TS_rw_lock_destroy(H5TS_rw_lock_t *rw_lock)
}
else {
- /* We are commited to the destroy at this point. Set magic
+ /* We are committed to the destroy at this point. Set magic
* to an invalid value, and call the appropriate pthread
* destroy routines. Call them all, even if one fails along
* the way.
diff --git a/src/H5TSprivate.h b/src/H5TSprivate.h
index e324b38..7bcf34d 100644
--- a/src/H5TSprivate.h
+++ b/src/H5TSprivate.h
@@ -285,7 +285,7 @@ typedef struct H5TS_rw_lock_stats_t {
* If this structure is allocated dynamically, remember to set
* it to some invalid value before discarding the structure.
*
- * policy Integer containing a code indicating the precidence policy
+ * policy Integer containing a code indicating the precedence policy
* used by the R/W lock. The supported policies are listed
* below:
*
@@ -362,13 +362,13 @@ typedef struct H5TS_rw_lock_t {
*
* Structure H5TS_rec_entry_count
*
- * Strucure associated with the reader_rec_entry_count_key defined in
+ * Structure associated with the reader_rec_entry_count_key defined in
* H5TS_rw_lock_t.
*
* The primary purpose of this structure is to maintain a count of recursive
* locks so that the lock can be dropped when the count drops to zero.
*
- * Aditional fields are included for purposes of sanity checking.
+ * Additional fields are included for purposes of sanity checking.
*
* Individual fields are discussed below.
*
diff --git a/src/H5Tbit.c b/src/H5Tbit.c
index 04e2a97..20f3855 100644
--- a/src/H5Tbit.c
+++ b/src/H5Tbit.c
@@ -228,7 +228,7 @@ done:
* Purpose: Return a small bit sequence as a number. Bit vector starts
* at OFFSET and is SIZE bits long.
*
- * Return: The bit sequence interpretted as an unsigned integer
+ * Return: The bit sequence interpreted as an unsigned integer
*
*-------------------------------------------------------------------------
*/
@@ -538,7 +538,7 @@ H5T__bit_inc(uint8_t *buf, size_t start, size_t size)
/*-------------------------------------------------------------------------
* Function: H5T__bit_dec
*
- * Purpose: Decrement part of a bit field by substracting 1. The bit
+ * Purpose: Decrement part of a bit field by subtracting 1. The bit
* field starts with bit position START and is SIZE bits long.
*
* Return: The "borrow-in" value. It's TRUE if underflows, FALSE
@@ -564,9 +564,9 @@ H5T__bit_dec(uint8_t *buf, size_t start, size_t size)
if ((size + start - 1) / 8 > idx) {
/* The bit sequence doesn't end in the same byte as starts */
- /* Example: a sequence like 11000100 and start = 3. We substract 00001000 from
+ /* Example: a sequence like 11000100 and start = 3. We subtract 00001000 from
* it and get 10111100. If a sequence is 00000111, we do right shift for START
- * bits and get 00000000. So we need to borrow from higher byte when we substract
+ * bits and get 00000000. So we need to borrow from higher byte when we subtract
* 00001000.
*/
if (!(buf[idx] >> pos))
@@ -597,7 +597,7 @@ H5T__bit_dec(uint8_t *buf, size_t start, size_t size)
else {
/* The bit sequence ends in the same byte as starts */
- /* Example: a sequence like 11000100 and pos=3, size=3. We substract 00001000
+ /* Example: a sequence like 11000100 and pos=3, size=3. We subtract 00001000
* and get 10111100. A bit is borrowed from 6th bit(buf[idx]>>6=00000010, tmp>>6=00000011,
* not equal). We need to put this bit back by increment 1000000.
*/
diff --git a/src/H5Tconv.c b/src/H5Tconv.c
index ff3a53f..c349757 100644
--- a/src/H5Tconv.c
+++ b/src/H5Tconv.c
@@ -4867,7 +4867,7 @@ H5T__conv_s_s(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelmts, siz
size_t olap; /*num overlapping elements */
size_t nchars = 0; /*number of characters copied */
uint8_t *s, *sp, *d, *dp; /*src and dst traversal pointers*/
- uint8_t *dbuf = NULL; /*temp buf for overlap convers. */
+ uint8_t *dbuf = NULL; /*temp buf for overlap converts. */
herr_t ret_value = SUCCEED; /* Return value */
FUNC_ENTER_PACKAGE
@@ -7145,7 +7145,7 @@ H5T__conv_schar_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7205,7 +7205,7 @@ H5T__conv_uchar_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7265,7 +7265,7 @@ H5T__conv_short_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7325,7 +7325,7 @@ H5T__conv_ushort_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7385,7 +7385,7 @@ H5T__conv_int_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelm
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7445,7 +7445,7 @@ H5T__conv_uint_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7505,7 +7505,7 @@ H5T__conv_long_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7565,7 +7565,7 @@ H5T__conv_ulong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7625,7 +7625,7 @@ H5T__conv_llong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7687,7 +7687,7 @@ H5T__conv_ullong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7797,7 +7797,7 @@ H5T__conv_double_uchar(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7819,7 +7819,7 @@ H5T__conv_ldouble_schar(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7929,7 +7929,7 @@ H5T__conv_double_ushort(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -7951,7 +7951,7 @@ H5T__conv_ldouble_short(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8061,7 +8061,7 @@ H5T__conv_double_uint(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8083,7 +8083,7 @@ H5T__conv_ldouble_int(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8193,7 +8193,7 @@ H5T__conv_double_ulong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8215,7 +8215,7 @@ H5T__conv_ldouble_long(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8325,7 +8325,7 @@ H5T__conv_double_ullong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
@@ -8349,7 +8349,7 @@ H5T__conv_ldouble_llong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n
* Return: Non-negative on success/Negative on failure
*
* Programmer: Raymond Lu
- * Tuesday, Febuary 1, 2005
+ * Tuesday, February 1, 2005
*
*-------------------------------------------------------------------------
*/
diff --git a/src/H5Tfields.c b/src/H5Tfields.c
index baeae6d..68bc0a3 100644
--- a/src/H5Tfields.c
+++ b/src/H5Tfields.c
@@ -12,7 +12,7 @@
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
/*
- * Module Info: This module contains commond functionality for fields in
+ * Module Info: This module contains command functionality for fields in
* enumerated & compound datatypes in the H5T interface.
*/
diff --git a/src/H5Tnative.c b/src/H5Tnative.c
index c9a2907..4529e57 100644
--- a/src/H5Tnative.c
+++ b/src/H5Tnative.c
@@ -56,7 +56,7 @@ static herr_t H5T__cmp_offset(size_t *comp_size, size_t *offset, size_t elem_siz
* H5T_NATIVE_LDOUBLE
*
* Compound, array, enum, and VL types all choose among these
- * types for theire members. Time, Bifield, Opaque, Reference
+ * types for their members. Time, Bifield, Opaque, Reference
* types are only copy out.
*
* Return: Success: Returns the native data type if successful.
diff --git a/src/H5Topaque.c b/src/H5Topaque.c
index f097fca..c5c2788 100644
--- a/src/H5Topaque.c
+++ b/src/H5Topaque.c
@@ -70,7 +70,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Tget_tag
*
- * Purpose: Get tha tag associated with an opaque datatype.
+ * Purpose: Get the tag associated with an opaque datatype.
*
* Return: A pointer to an allocated string. The caller should free
* the string. NULL is returned for errors.
diff --git a/src/H5VLcallback.c b/src/H5VLcallback.c
index cde3f84..8e420a0 100644
--- a/src/H5VLcallback.c
+++ b/src/H5VLcallback.c
@@ -4789,7 +4789,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL__link_copy
*
- * Purpose: Copys a link from src to dst.
+ * Purpose: Copies a link from src to dst.
*
* Return: Success: Non-negative
* Failure: Negative
@@ -4820,7 +4820,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL_link_copy
*
- * Purpose: Copys a link from src to dst.
+ * Purpose: Copies a link from src to dst.
*
* Return: Success: Non-negative
* Failure: Negative
@@ -6086,7 +6086,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL__request_wait
*
- * Purpose: Waits on an asychronous request through the VOL
+ * Purpose: Waits on an asynchronous request through the VOL
*
* Note: Releases the request if the operation has completed and the
* connector callback succeeds
@@ -6123,7 +6123,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5VL_request_wait
*
- * Purpose: Waits on an asychronous request through the VOL
+ * Purpose: Waits on an asynchronous request through the VOL
*
* Note: Releases the request if the operation has completed and the
* connector callback succeeds
diff --git a/src/H5VLint.c b/src/H5VLint.c
index 4a329a9..02d6f16 100644
--- a/src/H5VLint.c
+++ b/src/H5VLint.c
@@ -1300,7 +1300,7 @@ H5VL__register_connector_by_name(const char *name, hbool_t app_ref, hid_t vipl_i
if (H5I_iterate(H5I_VOL, H5VL__get_connector_cb, &op_data, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_BADITER, H5I_INVALID_HID, "can't iterate over VOL ids")
- /* If connector alread registered, increment ref count on ID and return ID */
+ /* If connector already registered, increment ref count on ID and return ID */
if (op_data.found_id != H5I_INVALID_HID) {
if (H5I_inc_ref(op_data.found_id, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_CANTINC, H5I_INVALID_HID,
@@ -1360,7 +1360,7 @@ H5VL__register_connector_by_value(H5VL_class_value_t value, hbool_t app_ref, hid
if (H5I_iterate(H5I_VOL, H5VL__get_connector_cb, &op_data, TRUE) < 0)
HGOTO_ERROR(H5E_VOL, H5E_BADITER, H5I_INVALID_HID, "can't iterate over VOL ids")
- /* If connector alread registered, increment ref count on ID and return ID */
+ /* If connector already registered, increment ref count on ID and return ID */
if (op_data.found_id != H5I_INVALID_HID) {
if (H5I_inc_ref(op_data.found_id, app_ref) < 0)
HGOTO_ERROR(H5E_VOL, H5E_CANTINC, H5I_INVALID_HID,
diff --git a/src/H5VLnative.c b/src/H5VLnative.c
index 48c5a80..7a12f56 100644
--- a/src/H5VLnative.c
+++ b/src/H5VLnative.c
@@ -222,7 +222,7 @@ H5VL__native_term(void)
* Purpose: Query the connector class.
*
* Note: This routine is in this file so that it can return the address
- * of the staticly declared class struct.
+ * of the statically declared class struct.
*
* Returns: SUCCEED (Can't fail)
*
@@ -296,7 +296,7 @@ done:
herr_t
H5VL__native_get_file_addr_len(void *obj, H5I_type_t obj_type, size_t *addr_len)
{
- H5F_t *file = NULL; /* File stuct pointer */
+ H5F_t *file = NULL; /* File struct pointer */
herr_t ret_value = SUCCEED;
FUNC_ENTER_NOAPI(FAIL)
diff --git a/src/H5VLnative_introspect.c b/src/H5VLnative_introspect.c
index 9d1244b..15652ee 100644
--- a/src/H5VLnative_introspect.c
+++ b/src/H5VLnative_introspect.c
@@ -22,7 +22,7 @@
#include "H5VLnative_private.h" /* Native VOL connector */
/* Note: H5VL__native_introspect_get_conn_cls is in src/H5VLnative.c so that
- * it can return the address of the staticly declared class struct.
+ * it can return the address of the statically declared class struct.
*/
/*---------------------------------------------------------------------------
diff --git a/src/H5VLpassthru.c b/src/H5VLpassthru.c
index 85f36d4..ca3c265 100644
--- a/src/H5VLpassthru.c
+++ b/src/H5VLpassthru.c
@@ -45,7 +45,7 @@
/* Macros */
/**********/
-/* Whether to display log messge when callback is invoked */
+/* Whether to display log message when callback is invoked */
/* (Uncomment to enable) */
/* #define ENABLE_PASSTHRU_LOGGING */
diff --git a/src/H5VM.c b/src/H5VM.c
index 4d65d75..209ac25 100644
--- a/src/H5VM.c
+++ b/src/H5VM.c
@@ -1244,7 +1244,7 @@ H5VM_chunk_index_scaled(unsigned ndims, const hsize_t *coord, const uint32_t *ch
* Function: H5VM_opvv
*
* Purpose: Perform an operation on a source & destination sequences
- * of offset/length pairs. Each set of sequnces has an array
+ * of offset/length pairs. Each set of sequences has an array
* of lengths, an array of offsets, the maximum number of
* sequences and the current sequence to start at in the sequence.
*
diff --git a/src/H5WB.c b/src/H5WB.c
index df7627f..84592bd 100644
--- a/src/H5WB.c
+++ b/src/H5WB.c
@@ -18,7 +18,7 @@
* Quincey Koziol
*
* Purpose: Implements the "wrapped buffer" code for wrapping
- * an existing [staticly sized] buffer, in order to
+ * an existing [statically sized] buffer, in order to
* avoid lots of memory allocation calls.
*
*-------------------------------------------------------------------------
diff --git a/src/H5Z.c b/src/H5Z.c
index 05c4da7..2cd1349 100644
--- a/src/H5Z.c
+++ b/src/H5Z.c
@@ -1496,7 +1496,7 @@ done:
/*-------------------------------------------------------------------------
* Function: H5Z_filter_in_pline
*
- * Purpose: Check wheter a filter is in the filter pipeline using the
+ * Purpose: Check whether a filter is in the filter pipeline using the
* filter ID. This function is very similar to H5Z_filter_info
*
* Return: TRUE - found filter
diff --git a/src/H5Znbit.c b/src/H5Znbit.c
index 3ed0ec6..36d1a52 100644
--- a/src/H5Znbit.c
+++ b/src/H5Znbit.c
@@ -1015,7 +1015,7 @@ done:
* assume one byte has 8 bit
* assume padding bit is 0
* assume size of unsigned char is one byte
- * assume one data item of certain datatype is stored continously in bytes
+ * assume one data item of certain datatype is stored continuously in bytes
* atomic datatype is treated on byte basis
*/
diff --git a/src/H5Zscaleoffset.c b/src/H5Zscaleoffset.c
index c7be014..21bbccb 100644
--- a/src/H5Zscaleoffset.c
+++ b/src/H5Zscaleoffset.c
@@ -659,7 +659,7 @@ H5Z_class2_t H5Z_SCALEOFFSET[1] = {{
buf[i] = (type)(buf[i] + (type)(minval)); \
} while (0)
-/* Retrive minimum value of floating-point type */
+/* Retrieve minimum value of floating-point type */
#define H5Z_scaleoffset_get_min(type, minval, min) \
{ \
if (sizeof(type) <= sizeof(long long)) \
@@ -1388,7 +1388,7 @@ done:
* assume one byte has 8 bit
* assume padding bit is 0
* assume size of unsigned char is one byte
- * assume one data item of certain datatype is stored continously in bytes
+ * assume one data item of certain datatype is stored continuously in bytes
* atomic datatype is treated on byte basis
*/
diff --git a/src/H5Ztrans.c b/src/H5Ztrans.c
index ea532cf..092b289 100644
--- a/src/H5Ztrans.c
+++ b/src/H5Ztrans.c
@@ -1352,7 +1352,7 @@ H5Z__op_is_numbs2(H5Z_node *_tree)
* Function: H5Z__xform_reduce_tree
*
* Purpose: Simplifies parse tree passed in by performing any obvious
- * and trivial arithemtic calculations.
+ * and trivial arithmetic calculations.
*
* Return: None.
*
@@ -1692,7 +1692,7 @@ H5Z_xform_noop(const H5Z_data_xform_t *data_xform_prop)
if (data_xform_prop) {
ret_value = FALSE;
- /* Check for trivial data tranformation: expression = "x" */
+ /* Check for trivial data transformation: expression = "x" */
if ((HDstrlen(data_xform_prop->xform_exp) == 1) && data_xform_prop->dat_val_pointers &&
(data_xform_prop->dat_val_pointers->num_ptrs == 1)) {
ret_value = TRUE;
@@ -1723,7 +1723,7 @@ H5Z_xform_extract_xform_str(const H5Z_data_xform_t *data_xform_prop)
/* There should be no way that this can be NULL since the function
* that calls this one checks to make sure it isn't before
- * pasing them */
+ * passing them */
HDassert(data_xform_prop);
FUNC_LEAVE_NOAPI(data_xform_prop->xform_exp)
diff --git a/src/H5detect.c b/src/H5detect.c
index ed2ae5f..1db0f06 100644
--- a/src/H5detect.c
+++ b/src/H5detect.c
@@ -808,7 +808,7 @@ done:\n\
fprintf(rawoutstream, "/* sigprocmask() support: no */\n");
#endif
- /* Print the statics of signal handlers called for debugging */
+ /* Print the statistics of signal handlers called for debugging */
fprintf(rawoutstream, "\n"
"/******************************/\n"
"/* signal handlers statistics */\n"
@@ -1602,7 +1602,7 @@ verify_signal_handlers(int signum, void (*handler)(int))
}
else {
if (val == signum) {
- /* return from signum handler. Record a sucess. */
+ /* return from signum handler. Record a success. */
nsuccesses++;
}
else {
diff --git a/src/H5mpi.c b/src/H5mpi.c
index 34ebce8..15fb785 100644
--- a/src/H5mpi.c
+++ b/src/H5mpi.c
@@ -38,7 +38,7 @@ static hsize_t bigio_count_g = H5_MAX_MPI_COUNT;
/*-------------------------------------------------------------------------
* Function: H5_mpi_set_bigio_count
*
- * Purpose: Allow us to programatically change the switch point
+ * Purpose: Allow us to programmatically change the switch point
* when we utilize derived datatypes. This is of
* particular interest for allowing nightly testing
*
diff --git a/src/H5private.h b/src/H5private.h
index b560d62..1bbdc52 100644
--- a/src/H5private.h
+++ b/src/H5private.h
@@ -111,7 +111,7 @@
#endif
/*
- * Longjumps are used to detect alignment constrants
+ * Longjumps are used to detect alignment constraints
*/
#ifdef H5_HAVE_SETJMP_H
#include <setjmp.h>
diff --git a/test/SWMR_POSIX_Order_UG.txt b/test/SWMR_POSIX_Order_UG.txt
index 2771af1..a4190bc 100644
--- a/test/SWMR_POSIX_Order_UG.txt
+++ b/test/SWMR_POSIX_Order_UG.txt
@@ -3,7 +3,7 @@ POSIX Write Order Test Instructions
Purpose
-------
-This documents shows the requirments, implementaion design and instructions
+This documents shows the requirements, implementation design and instructions
of building and running the POSIX Write Order test. The name of the
test is twriteorder and it resides in the test/ directory.
@@ -44,7 +44,7 @@ and they are ready for the reader processes to access.
If the system, in which the writer and reader processes run, the readers
will always get all chain-linked blocks correctly. If the order of write
-is not maintained, some reader processes may found unexpect block data.
+is not maintained, some reader processes may found unexpected block data.
Building the Tests
------------------
diff --git a/test/SWMR_UseCase_UG.txt b/test/SWMR_UseCase_UG.txt
index 1e3d1e6..677fdc4 100644
--- a/test/SWMR_UseCase_UG.txt
+++ b/test/SWMR_UseCase_UG.txt
@@ -2,7 +2,7 @@
User Guide for SWMR Use Case Programs
2. Purpose:
- This is a User Guide of the SWMR Use Case programs. It descibes the use
+ This is a User Guide of the SWMR Use Case programs. It describes the use
case program and explain how to run them.
2.1. Author and Dates:
@@ -87,7 +87,7 @@
It opens the HDF5 data file without the SWMR access mode (0 means
off). This likely will result in error. This option is provided for
- users to see the effect of the neede SWMR access mode for concurrent
+ users to see the effect of the needed SWMR access mode for concurrent
access.
3.4. Test Shell Script:
@@ -176,7 +176,7 @@
It opens the HDF5 data file without the SWMR access mode (0 means
off). This likely will result in error. This option is provided for
- users to see the effect of the neede SWMR access mode for concurrent
+ users to see the effect of the needed SWMR access mode for concurrent
access.
4.4. Test Shell Script:
diff --git a/test/accum.c b/test/accum.c
index 47aae39..73652a3 100644
--- a/test/accum.c
+++ b/test/accum.c
@@ -1348,7 +1348,7 @@ test_accum_adjust(H5F_t *f)
/* ==> Accumulator will try to resize, but see that it's getting too big */
/* ==> Size of new block is less than than half maximum size of accumulator */
/* ==> New block being appended to accumulator */
- /* ==> We can slide the dirty region down, to accomodate the request */
+ /* ==> We can slide the dirty region down, to accommodate the request */
/* ==> Max Buffer Size - (dirty offset + adjust size) >= 2 * size) */
/* ==> Need to adjust location of accumulator while appending */
/* ==> Accumulator will need to be reallocated */
@@ -1394,7 +1394,7 @@ test_accum_adjust(H5F_t *f)
/* ==> Accumulator will try to resize, but see that it's getting too big */
/* ==> Size of new block is less than than half maximum size of accumulator */
/* ==> New block being appended to accumulator */
- /* ==> We can slide the dirty region down, to accomodate the request */
+ /* ==> We can slide the dirty region down, to accommodate the request */
/* ==> Max Buffer Size - (dirty offset + adjust size) < 2 * size) */
/* ==> Need to adjust location of accumulator while appending */
if (accum_write((1024 * 1024) - 5, 10, wbuf) < 0)
diff --git a/test/atomic_writer.c b/test/atomic_writer.c
index 5a6310a..33252f1 100644
--- a/test/atomic_writer.c
+++ b/test/atomic_writer.c
@@ -112,7 +112,7 @@ int
main(int argc, char *argv[])
{
int fd = -1; /* file descriptor */
- ssize_t bytes_wrote; /* the nubmer of bytes written */
+ ssize_t bytes_wrote; /* the number of bytes written */
unsigned int *buf = NULL; /* buffer to hold written data */
unsigned int n, u, i; /* local index variable */
int temp; /* temporary variable */
diff --git a/test/big.c b/test/big.c
index 2bbae67..9f62b3a 100644
--- a/test/big.c
+++ b/test/big.c
@@ -43,7 +43,7 @@
* that disk space is allocated only when the contents are actually written.
* E.g., If one creates a new file, seeks forward 10 million bytes, writes
* 1 bytes and closes the file, then a sparse file, will show file size of
- * 10 million bytes but actaully uses only couple disk blocks, much smaller
+ * 10 million bytes but actually uses only couple disk blocks, much smaller
* than the formal file size.)
*
* One more consideration is that we want to distinguish an HDF5 library
@@ -70,7 +70,7 @@
#endif
/* Define Small, Large, Extra Large, Huge File which
- * corrspond to less than 2GB, 2GB, 4GB, and tens of GB file size.
+ * correspond to less than 2GB, 2GB, 4GB, and tens of GB file size.
* NO_FILE stands for "no file" to be tested.
*/
typedef enum fsizes_t { SFILE, LFILE, XLFILE, HUGEFILE, NO_FILE } fsizes_t;
@@ -609,7 +609,7 @@ usage(void)
"Examples:\n"
"\tbig -fsize 2.1e9 \t# test with file size just under 2GB\n"
"\tbig -fsize 2.2e9 \t# test with file size just above 2GB\n"
- "\tBe sure the file system can support the file size requested\n",
+ "\the sure the file system can support the file size requested\n",
(hsize_t)FAMILY_SIZE);
}
diff --git a/test/btree2.c b/test/btree2.c
index 3e03366..ec99229 100644
--- a/test/btree2.c
+++ b/test/btree2.c
@@ -696,7 +696,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_insert(bt2, &record) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
idx = 41;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
@@ -707,7 +707,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_find(bt2, &idx, NULL, NULL) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in B-tree with 1 record */
+ /* Attempt to find existent record in B-tree with 1 record */
idx = 42;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
TEST_ERROR
@@ -716,7 +716,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_find(bt2, &idx, NULL, NULL) != TRUE)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
idx = 0;
H5E_BEGIN_TRY
{
@@ -764,18 +764,18 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_insert(bt2, &record) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
idx = 41;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several record */
+ /* Attempt to find existent record in level-0 B-tree with several record */
idx = 56;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
idx = 0;
H5E_BEGIN_TRY
{
@@ -935,23 +935,23 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (idx != (INSERT_SPLIT_ROOT_NREC + 2))
TEST_ERROR
- /* Attempt to find non-existant record in level-1 B-tree */
+ /* Attempt to find non-existent record in level-1 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_SPLIT_ROOT_NREC + 10;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in root of level-1 B-tree */
+ /* Attempt to find existent record in root of level-1 B-tree */
idx = 33;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
- /* Attempt to find existant record in leaf of level-1 B-tree */
+ /* Attempt to find existent record in leaf of level-1 B-tree */
idx = 56;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
- /* Attempt to index non-existant record in level-1 B-tree */
+ /* Attempt to index non-existent record in level-1 B-tree */
idx = 0;
H5E_BEGIN_TRY
{
@@ -1671,13 +1671,13 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx != ((INSERT_SPLIT_ROOT_NREC * 29) + 5))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_SPLIT_ROOT_NREC * 30;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
idx = 948;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
@@ -1687,7 +1687,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
idx = 505;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
@@ -1697,7 +1697,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
idx = 555;
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
@@ -1707,7 +1707,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
idx = 0;
H5E_BEGIN_TRY
{
@@ -3060,7 +3060,7 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
if (idx != INSERT_MANY)
TEST_ERROR
- /* Attempt to find non-existant record in level-4 B-tree */
+ /* Attempt to find non-existent record in level-4 B-tree */
/* (Should not be found, but not fail) */
idx = INSERT_MANY * 2;
if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE)
@@ -3071,12 +3071,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
/* Pick random record */
idx = (hsize_t)(HDrandom() % INSERT_MANY);
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE)
FAIL_STACK_ERROR
} /* end for */
- /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */
+ /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY * 3), find_cb, NULL);
@@ -3099,12 +3099,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time);
/* Pick random record */
idx = (hsize_t)(HDrandom() % INSERT_MANY);
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
/* (in increasing order) */
if (H5B2_index(bt2, H5_ITER_INC, idx, find_cb, &idx) < 0)
FAIL_STACK_ERROR
- /* Attempt to find existant record in root of level-4 B-tree */
+ /* Attempt to find existent record in root of level-4 B-tree */
/* (in decreasing order) */
if (H5B2_index(bt2, H5_ITER_DEC, idx, find_dec_cb, &idx) < 0)
FAIL_STACK_ERROR
@@ -3210,7 +3210,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3226,7 +3226,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in B-tree with 1 record */
+ /* Attempt to find existent record in B-tree with 1 record */
find.key = 42;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
FAIL_STACK_ERROR
@@ -3240,7 +3240,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL);
@@ -3275,7 +3275,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &modify, modify_rec_cb, &modify) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in B-tree with 1 record */
+ /* Attempt to find non-existent record in B-tree with 1 record */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3305,7 +3305,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with 1 record */
+ /* Attempt to index non-existent record in B-tree with 1 record */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL);
@@ -3360,7 +3360,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
find.key = 10;
find.val = (hsize_t)-1;
@@ -3369,14 +3369,14 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several records */
+ /* Attempt to find existent record in level-0 B-tree with several records */
find.key = 56;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
TEST_ERROR
if (find.val != 12)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL);
@@ -3448,7 +3448,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (H5B2_update(bt2, &record, modify_rec_cb, &modify) < 0)
FAIL_STACK_ERROR
- /* Attempt to find non-existant record in level-0 B-tree with several records */
+ /* Attempt to find non-existent record in level-0 B-tree with several records */
/* (Should not be found, but not fail) */
find.key = 41;
find.val = (hsize_t)-1;
@@ -3457,14 +3457,14 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in level-0 B-tree with several record */
+ /* Attempt to find existent record in level-0 B-tree with several record */
find.key = 56;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
TEST_ERROR
if (find.val != 22)
TEST_ERROR
- /* Attempt to index non-existant record in B-tree with several records */
+ /* Attempt to index non-existent record in B-tree with several records */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL);
@@ -3692,7 +3692,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (idx.key != (INSERT_SPLIT_ROOT_NREC_REC + 2))
TEST_ERROR
- /* Attempt to find non-existant record in level-1 B-tree */
+ /* Attempt to find non-existent record in level-1 B-tree */
/* (Should not be found, but not fail) */
find.key = 800;
find.val = (hsize_t)-1;
@@ -3701,7 +3701,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-1 B-tree */
+ /* Attempt to find existent record in root of level-1 B-tree */
find.key = 33;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -3711,7 +3711,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (find.val != 67)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-1 B-tree */
+ /* Attempt to find existent record in leaf of level-1 B-tree */
find.key = 56;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -3721,7 +3721,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p
if (find.val != 113)
TEST_ERROR
- /* Attempt to index non-existant record in level-1 B-tree */
+ /* Attempt to index non-existent record in level-1 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC + 2), index_rec_cb, NULL);
@@ -4478,7 +4478,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4487,7 +4487,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1347;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4500,7 +4500,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 513;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4513,7 +4513,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4526,7 +4526,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -4643,7 +4643,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4652,7 +4652,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1344;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4665,7 +4665,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 512;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4678,7 +4678,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4691,7 +4691,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -4818,7 +4818,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 6))
TEST_ERROR
- /* Attempt to find non-existant record in level-2 B-tree */
+ /* Attempt to find non-existent record in level-2 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_SPLIT_ROOT_NREC_REC * 42;
find.val = (hsize_t)-1;
@@ -4827,7 +4827,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (find.val != (hsize_t)-1)
TEST_ERROR
- /* Attempt to find existant record in root of level-2 B-tree */
+ /* Attempt to find existent record in root of level-2 B-tree */
find.key = 1345;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4840,7 +4840,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)2) < 0)
TEST_ERROR
- /* Attempt to find existant record in internal node of level-2 B-tree */
+ /* Attempt to find existent record in internal node of level-2 B-tree */
find.key = 513;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4853,7 +4853,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)1) < 0)
TEST_ERROR
- /* Attempt to find existant record in leaf of level-2 B-tree */
+ /* Attempt to find existent record in leaf of level-2 B-tree */
find.key = 555;
find.val = (hsize_t)-1;
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
@@ -4866,7 +4866,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_
if (check_node_depth(bt2, &record, (unsigned)0) < 0)
TEST_ERROR
- /* Attempt to index non-existant record in level-2 B-tree */
+ /* Attempt to index non-existent record in level-2 B-tree */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL);
@@ -5051,7 +5051,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
if (iter.key != INSERT_MANY_REC)
TEST_ERROR
- /* Attempt to find non-existant record in level-4 B-tree */
+ /* Attempt to find non-existent record in level-4 B-tree */
/* (Should not be found, but not fail) */
find.key = INSERT_MANY_REC * 2;
find.val = (hsize_t)-1;
@@ -5066,14 +5066,14 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)(HDrandom() % INSERT_MANY_REC);
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE)
FAIL_STACK_ERROR
if (find.val != (find.key * 2))
TEST_ERROR
} /* end for */
- /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */
+ /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */
H5E_BEGIN_TRY
{
ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY_REC * 3), find_rec_cb, NULL);
@@ -5102,7 +5102,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)-1;
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
/* (in increasing order) */
if (H5B2_index(bt2, H5_ITER_INC, idx, index_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5115,7 +5115,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time);
find.key = (hsize_t)-1;
find.val = (hsize_t)-1;
- /* Attempt to find existant record in level-4 B-tree */
+ /* Attempt to find existent record in level-4 B-tree */
/* (in decreasing order) */
if (H5B2_index(bt2, H5_ITER_DEC, idx, index_rec_cb, &find) < 0)
FAIL_STACK_ERROR
@@ -5234,7 +5234,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
- TESTING("B-tree remove: non-existant record from 1 record B-tree");
+ TESTING("B-tree remove: non-existent record from 1 record B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5261,7 +5261,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
if (!H5F_addr_defined(root_addr))
TEST_ERROR
- /* Attempt to remove a non-existant record from a B-tree with 1 record */
+ /* Attempt to remove a non-existent record from a B-tree with 1 record */
record = 0;
H5E_BEGIN_TRY
{
@@ -5275,7 +5275,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
/* Attempt to remove a record from a B-tree with 1 record */
- TESTING("B-tree remove: existant record from 1 record B-tree");
+ TESTING("B-tree remove: existent record from 1 record B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5339,8 +5339,8 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
- /* Attempt to remove a non-existant record from a level-0 B-tree with mult. record */
- TESTING("B-tree remove: non-existant record from level-0 B-tree");
+ /* Attempt to remove a non-existent record from a level-0 B-tree with mult. record */
+ TESTING("B-tree remove: non-existent record from level-0 B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5359,7 +5359,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_
PASSED();
/* Attempt to remove a record from a level-0 B-tree with mult. record */
- TESTING("B-tree remove: mult. existant records from level-0 B-tree");
+ TESTING("B-tree remove: mult. existent records from level-0 B-tree");
/* Check for closing & re-opening the B-tree */
if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0)
@@ -5517,7 +5517,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt
herr_t ret; /* Generic error return value */
/* B-tree record removal tests */
- TESTING("B-tree remove: non-existant record from level-1 B-tree");
+ TESTING("B-tree remove: non-existent record from level-1 B-tree");
/* Create the file for the test */
if (create_file(&file, &f, fapl) < 0)
@@ -5550,7 +5550,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt
if (!H5F_addr_defined(root_addr))
TEST_ERROR
- /* Attempt to remove a non-existant record from a B-tree with 1 record */
+ /* Attempt to remove a non-existent record from a B-tree with 1 record */
record = (INSERT_SPLIT_ROOT_NREC * 2) + 1;
H5E_BEGIN_TRY
{
@@ -9605,7 +9605,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa
/*
* Test modifying records
*/
- TESTING("B-tree modify: attempt to modify non-existant record");
+ TESTING("B-tree modify: attempt to modify non-existent record");
/* Create the file for the test */
if (create_file(&file, &f, fapl) < 0)
@@ -9628,7 +9628,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa
if (bt2_stat.depth != 2)
TEST_ERROR
- /* Attempt to modify a non-existant record */
+ /* Attempt to modify a non-existent record */
record = 3;
modify = 4;
H5E_BEGIN_TRY
diff --git a/test/cache.c b/test/cache.c
index aa58777..95a985b 100644
--- a/test/cache.c
+++ b/test/cache.c
@@ -2774,8 +2774,8 @@ write_permitted_check(int
* Test the behaviour with different flags.
*
* This test was added primarily to test basic insert
- * pinned entry functionallity, but I through in explicit
- * tests for other functionallity that is tested implicitly
+ * pinned entry functionality, but I through in explicit
+ * tests for other functionality that is tested implicitly
* elsewhere.
*
* Return: void
@@ -2893,7 +2893,7 @@ check_insert_entry(unsigned paged)
}
}
- /* Thats all we can get from H5C_get_entry_status().
+ /* That's all we can get from H5C_get_entry_status().
* Now start looking at the cache data structures directly.
*/
@@ -5131,7 +5131,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr)
*
* 10/8/07 -- JRM
* Added a resize operation to this test to satisfy the new
- * requiremnt that any resize of an entry on flush will always
+ * requirement that any resize of an entry on flush will always
* be accompanied by a resize. Note that as a result, this
* test becomes redundant with later tests.
*/
@@ -5223,7 +5223,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr)
*
* 10/8/07 -- JRM
* Added a resize operation to this test to satisfy the new
- * requiremnt that any resize of an entry on flush will always
+ * requirement that any resize of an entry on flush will always
* be accompanied by a resize. Note that as a result, this
* test becomes redundant with later tests.
*/
@@ -9601,7 +9601,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
*
* (VET, 9) Y 10 KB N N - dirty (VET, 8)
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -10154,7 +10154,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
* returns to the serialize function for (VET, 9).
*
* When the serialize function for (VET, 9) is done with (VET, 8), it
- * calls H5C_unprotect(VET, 8), which markes (VET, 8) as dirty and
+ * calls H5C_unprotect(VET, 8), which marks (VET, 8) as dirty and
* unprotected, and places it at the head of the LRU.
*
* (VET, 0) is the next item on the LRU -- it is clean and is therefore
@@ -10553,7 +10553,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr)
* Added tests for pinned entries.
*
* JRM -- 5/17/06
- * Complete reqrite of pinned entry tests to accomodate
+ * Complete reqrite of pinned entry tests to accommodate
* the new H5C_mark_entry_dirty() call.
*
*-------------------------------------------------------------------------
@@ -12978,7 +12978,7 @@ check_expunge_entry(unsigned paged)
* Function: check_multiple_read_protect()
*
* Purpose: Verify that multiple, simultaneous read protects of a
- * single entry perform as expectd.
+ * single entry perform as expected.
*
* Return: void
*
@@ -13324,7 +13324,7 @@ check_multiple_read_protect(unsigned paged)
if (pass && (entry_ptr->header.is_pinned)) {
pass = FALSE;
- failure_mssg = "enty (0,4) still pinned.\n";
+ failure_mssg = "entry (0,4) still pinned.\n";
}
unprotect_entry(file_ptr, 0, 4, /* (0,4) unprotect */
@@ -15527,7 +15527,7 @@ check_destroy_protected_err(unsigned paged)
* Function: check_duplicate_insert_err()
*
* Purpose: Verify that an attempt to insert and entry that is
- * alread in the cache will generate an error.
+ * already in the cache will generate an error.
*
* Return: void
*
@@ -18083,7 +18083,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- force the hit rate to 100% again -- should be steady
+ /* eighth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -18179,7 +18179,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- if (pass) { /* twelth epoch -- at minimum size so no more ageouts */
+ if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */
rpt_fcn_called = FALSE;
i = 0;
while (pass && (i < 1000)) {
@@ -18460,7 +18460,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- still 100% hit rate
+ /* eighth epoch -- still 100% hit rate
*/
if (pass) {
@@ -18563,7 +18563,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* twelth epoch -- force the hit rate to 100% again -- should be steady
+ /* twelfth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -18939,7 +18939,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* eigth epoch -- force the hit rate to 100% again -- should be steady
+ /* eighth epoch -- force the hit rate to 100% again -- should be steady
* state.
*/
if (pass) {
@@ -19039,7 +19039,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- if (pass) { /* twelth epoch -- at minimum size so no more ageouts */
+ if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */
rpt_fcn_called = FALSE;
i = 0;
@@ -19690,7 +19690,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* twelth epoch -- hit rate 1.0 -- decrease as before.
+ /* twelfth epoch -- hit rate 1.0 -- decrease as before.
*/
if (pass) {
@@ -20703,7 +20703,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged)
if (show_progress)
HDfprintf(stderr, "check point %d\n", checkpoint++);
- /* protect and unprotect two more large entries -- shouldnt trigger a
+ /* protect and unprotect two more large entries -- shouldn't trigger a
* flash cache size increase.
*/
if (pass) {
@@ -23263,7 +23263,7 @@ check_auto_cache_resize_disable(unsigned paged)
* mode (add space), which is triggered whenever the size of a newly
* loaded or inserted entry, or the delta between old and new entry
* sizes exceeds some fraction of the current maximum cache size, and
- * the cache doesn't have enough free space to accomodate the new/
+ * the cache doesn't have enough free space to accommodate the new/
* resize entry without performing evictions. The range of permissible
* values for the flash_threshold (0.1 to 1.0 as of this writing), and
* for the flash_multiple (0.1 to 10.0) do not permit the facility to
@@ -23272,7 +23272,7 @@ check_auto_cache_resize_disable(unsigned paged)
* current max_cache_size equal to max_size.
*
* We have already tested the latter in check_auto_cache_resize(), so
- * we need only thest the former here. Do this by disabling flash
+ * we need only test the former here. Do this by disabling flash
* cache size increments via the flash_incr_mode, and then creating
* situations that would trigger flash cache size increases were that
* code enabled.
@@ -26345,7 +26345,7 @@ check_auto_cache_resize_input_errs(unsigned paged)
/*-------------------------------------------------------------------------
* Function: check_auto_cache_resize_aux_fcns()
*
- * Purpose: Verify that the auxilary functions associated with
+ * Purpose: Verify that the auxiliary functions associated with
* the automatic cache resize capability are operating
* correctly. These functions are:
*
@@ -26422,13 +26422,13 @@ check_auto_cache_resize_aux_fcns(unsigned paged)
};
if (paged)
- TESTING("automatic cache resize auxilary functions (paged aggregation)")
+ TESTING("automatic cache resize auxiliary functions (paged aggregation)")
else
- TESTING("automatic cache resize auxilary functions")
+ TESTING("automatic cache resize auxiliary functions")
pass = TRUE;
- /* allocate a cache, and then test the various auxilary functions.
+ /* allocate a cache, and then test the various auxiliary functions.
*/
if (pass) {
@@ -27075,7 +27075,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged)
if (file_ptr == NULL) {
pass = FALSE;
- failure_mssg = "bad return from cache intialization.\n";
+ failure_mssg = "bad return from cache initialization.\n";
}
else
cache_ptr = file_ptr->shared->cache;
@@ -27294,7 +27294,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged)
H5C__DIRTIED_FLAG); /* unsigned int flags */
}
- /* Expected staus is that after each insertion, the entry
+ /* Expected status is that after each insertion, the entry
* inserted 26 insertions ago has been flushed, and the
* entry currently getting inserted is now in the cache and
* dirty.
@@ -29699,7 +29699,7 @@ check_flush_deps_err(unsigned paged)
result =
H5C_destroy_flush_dependency(&((entries[entry_type])[0]), &((entries[entry_type])[1]));
if (result != FAIL)
- CACHE_ERROR("Destroying [non-existant] dependency when parent isn't protected")
+ CACHE_ERROR("Destroying [non-existent] dependency when parent isn't protected")
break;
/* Verify that parent entry has flush dependency */
@@ -33055,7 +33055,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged)
if (file_ptr == NULL) {
pass = FALSE;
- failure_mssg = "bad return from cache intialization.\n";
+ failure_mssg = "bad return from cache initialization.\n";
}
else
cache_ptr = file_ptr->shared->cache;
@@ -33565,7 +33565,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged)
* an entry that is no longer in the cache, and which may have
* been deleted.
*
- * This function contans tests for correct handling on this
+ * This function contains tests for correct handling on this
* situation.
*
* Do nothing if pass is FALSE on entry.
@@ -33796,7 +33796,7 @@ cedds__expunge_dirty_entry_in_flush_test(H5F_t *file_ptr)
*
* (HET, 3) Y 16 KB Y N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -34046,7 +34046,7 @@ cedds__H5C_make_space_in_cache(H5F_t *file_ptr)
*
* (HET, 3) Y 16 KB N N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
@@ -34461,7 +34461,7 @@ cedds__H5C__autoadjust__ageout__evict_aged_out_entries(H5F_t *file_ptr)
*
* (MET, 1-31) Y 64 KB N N - -
*
- * Recall that in this test bed, flush operations are excuted the
+ * Recall that in this test bed, flush operations are executed the
* first time the associated entry is flushed, and are then
* deleted.
*/
diff --git a/test/cache_api.c b/test/cache_api.c
index 55aaa0a..f79a531 100644
--- a/test/cache_api.c
+++ b/test/cache_api.c
@@ -662,7 +662,7 @@ check_file_mdc_api_calls(unsigned paged, hid_t fcpl_id)
* configured as per the default both by looking at its internal
* configuration, and via the H5Fget_mdc_config() call.
*
- * Then set serveral different configurations, and verify that
+ * Then set several different configurations, and verify that
* they took as per above.
*/
@@ -1054,7 +1054,7 @@ mdc_api_call_smoke_check(int express_test, unsigned paged, hid_t fcpl_id)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -1885,7 +1885,7 @@ check_file_mdc_api_errs(unsigned paged, hid_t fcpl_id)
pass = TRUE;
- /* Create a file for test purposes, and veify that its metadata cache
+ /* Create a file for test purposes, and verify that its metadata cache
* set to the default MDC configuration.
*/
@@ -2254,7 +2254,7 @@ main(void)
} /* end if */
/* Test with paged aggregation enabled or not */
- /* The "my_fcpl" passed to each test has the paged or non-paged strategy set up accordinly */
+ /* The "my_fcpl" passed to each test has the paged or non-paged strategy set up accordingly */
for (paged = FALSE; paged <= TRUE; paged++) {
hid_t my_fcpl = fcpl_id;
diff --git a/test/cache_common.c b/test/cache_common.c
index 59fda94..61c5565 100644
--- a/test/cache_common.c
+++ b/test/cache_common.c
@@ -1523,7 +1523,7 @@ add_flush_op(int target_type, int target_idx, int op_code, int type, int idx, hb
* Otherwise, set up a pinned entry dependency so we can
* test the pinned entry modifications to the flush routine.
*
- * Given the types and indicies of the pinned and pinning
+ * Given the types and indices of the pinned and pinning
* entries, add the pinned entry to the list of pinned
* entries in the pinning entry, increment the
* pinning reference count of the pinned entry, and
@@ -1768,7 +1768,7 @@ execute_flush_op(H5F_t *file_ptr, struct test_entry_t *entry_ptr, struct flush_o
case FLUSH_OP__EXPUNGE:
/* the expunge flush op exists to allow us to simulate the
- * case in which an entry is removed from the cashe as the
+ * case in which an entry is removed from the cache as the
* the result of the flush of a second entry. At present,
* this can only happen via the take ownership flag, but
* we will make this test feature more general to as to make
@@ -2313,7 +2313,7 @@ verify_clean(void)
* Function: verify_entry_status
*
* Purpose: Verify that a list of entries have the expected status.
- * If any discrepencies are found, set the failure message
+ * If any discrepancies are found, set the failure message
* and set pass to FALSE.
*
* Do nothing if pass is FALSE on entry.
@@ -3833,7 +3833,7 @@ row_major_scan_forward(H5F_t *file_ptr, int32_t max_index, int32_t lag, hbool_t
HDassert((!cache_ptr->slist_enabled) ||
(cache_ptr->slist_size == cache_ptr->dirty_index_size));
} /* end if */
- } /* end elsef */
+ } /* end elseif */
if (verbose)
HDfprintf(stdout, "\n");
@@ -5241,7 +5241,7 @@ resize_configs_are_equal(const H5C_auto_size_ctl_t *a, const H5C_auto_size_ctl_t
* has both internal and external configuration matching
* *config_ptr.
*
- * Do nothin on success. On failure, set pass to FALSE, and
+ * Do nothing on success. On failure, set pass to FALSE, and
* load an error message into failue_mssg. Note that
* failure_msg is assumed to be at least 128 bytes in length.
*
diff --git a/test/cache_image.c b/test/cache_image.c
index 90fe7af..d2313ba 100644
--- a/test/cache_image.c
+++ b/test/cache_image.c
@@ -136,7 +136,7 @@ create_datasets(hid_t file_id, int min_dset, int max_dset)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -470,7 +470,7 @@ delete_datasets(hid_t file_id, int min_dset, int max_dset)
* Set pass to FALSE and issue a suitable failure
* message if either the file contains a metadata cache image
* superblock extension and mdci_sbem_expected is TRUE, or
- * vise versa.
+ * vice versa.
*
* If mdci_sbem_expected is TRUE, also verify that the metadata
* cache has been advised of this.
@@ -544,7 +544,7 @@ open_hdf5_file(hbool_t create_file, hbool_t mdci_sbem_expected, hbool_t read_onl
if (show_progress)
HDfprintf(stdout, "%s: cp = %d.\n", fcn_name, cp++);
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = h5_fileaccess();
@@ -879,7 +879,7 @@ attempt_swmr_open_hdf5_file(const hbool_t create_file, const hbool_t set_mdci_fa
H5AC_cache_image_config_t cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE,
H5AC__CACHE_IMAGE__ENTRY_AGEOUT__NONE};
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = h5_fileaccess();
@@ -1505,7 +1505,7 @@ check_cache_image_ctl_flow_1(hbool_t single_file_vfd)
* processed as part of the first protect operation after the
* superblock is loaded.)
*
- * In this particular test, we preform the following operations:
+ * In this particular test, we perform the following operations:
*
* 1) Create a HDF5 file with the cache image FAPL entry.
*
@@ -7001,7 +7001,7 @@ cache_image_api_error_check_4(hbool_t single_file_vfd)
* image.
*
* The objective of this test is to create a test file
- * with both non-empty self referential presistant
+ * with both non-empty self referential persistent
* free space managers, and a cache image, and then
* verify that this situation is handled correctly if
* H5Fget_free_sections() is called before the metadata
diff --git a/test/cache_tagging.c b/test/cache_tagging.c
index d03caec..b01fc34 100644
--- a/test/cache_tagging.c
+++ b/test/cache_tagging.c
@@ -443,7 +443,7 @@ check_file_creation_tags(hid_t fcpl_id, int type)
/* Variable Declarations */
hid_t fid = -1; /* File Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose test outout */
+ int verbose = FALSE; /* verbose test output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -547,7 +547,7 @@ check_file_open_tags(hid_t fcpl, int type)
/* Variable Declarations */
hid_t fid = -1; /* File Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag; /* Root Group Tag */
@@ -677,7 +677,7 @@ check_group_creation_tags(void)
hid_t fid = -1; /* File Identifier */
hid_t gid = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = HADDR_UNDEF; /* Root Group Tag */
@@ -800,7 +800,7 @@ check_multi_group_creation_tags(void)
hid_t fid = -1; /* File Identifier */
hid_t gid = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
char gname[16]; /* group name buffer */
int i = 0; /* iterator */
@@ -952,7 +952,7 @@ check_link_iteration_tags(void)
hid_t sid = -1; /* Group Identifier */
hid_t did = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int i = 0; /* iterator */
haddr_t root_tag = 0; /* Root Group Tag Value */
@@ -1094,7 +1094,7 @@ check_dense_attribute_tags(void)
hid_t did = -1; /* Group Identifier */
hid_t dcpl = -1; /* Group Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int i = 0; /* iterator */
hid_t fapl = -1; /* File access property list */
@@ -1455,7 +1455,7 @@ check_attribute_creation_tags(hid_t fcpl, int type)
hid_t gid = -1; /* Group Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0; /* Root group tag */
@@ -1620,7 +1620,7 @@ check_attribute_open_tags(hid_t fcpl, int type)
hid_t gid = -1; /* Group Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -1787,7 +1787,7 @@ check_attribute_rename_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -2000,7 +2000,7 @@ check_attribute_delete_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -2191,7 +2191,7 @@ check_dataset_creation_tags(hid_t fcpl, int type)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2351,7 +2351,7 @@ check_dataset_creation_earlyalloc_tags(hid_t fcpl, int type)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2517,7 +2517,7 @@ check_dataset_open_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2669,7 +2669,7 @@ check_dataset_write_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -2836,7 +2836,7 @@ check_attribute_write_tags(hid_t fcpl, int type)
hid_t aid = -1; /* Attribute Identifier */
hid_t sid = -1; /* Dataset Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
int * data = NULL; /* data buffer */
int i, j, k = 0; /* iterators */
@@ -3022,7 +3022,7 @@ check_dataset_read_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3184,7 +3184,7 @@ check_dataset_size_retrieval(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3348,7 +3348,7 @@ check_dataset_extend_tags(void)
hid_t did = -1; /* Dataset Identifier */
hid_t sid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3795,7 +3795,7 @@ check_link_removal_tags(hid_t fcpl, int type)
hid_t sid = -1; /* Dataspace Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -3984,7 +3984,7 @@ check_link_getname_tags(void)
hid_t sid = -1; /* Dataspace Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t dcpl = -1; /* dataset creation pl */
hsize_t cdims[2] = {1, 1}; /* chunk dimensions */
@@ -4161,7 +4161,7 @@ check_external_link_creation_tags(void)
hid_t fid2 = -1; /* File Identifier */
hid_t gid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
hid_t fapl = -1; /* File access prop list */
haddr_t root_tag = 0;
@@ -4288,7 +4288,7 @@ check_external_link_open_tags(void)
hid_t gid = -1; /* Dataspace Identifier */
hid_t xid = -1; /* Dataspace Identifier */
#ifndef NDEBUG
- int verbose = FALSE; /* verbose file outout */
+ int verbose = FALSE; /* verbose file output */
#endif /* NDEBUG */
H5O_native_info_t ninfo; /* Native object info struct */
hid_t fapl = -1; /* File access prop list */
@@ -4579,7 +4579,7 @@ main(void)
/* Create a standard file creation property list */
fcpl_default = H5Pcreate(H5P_FILE_CREATE);
- /* Create an fcpl with shared messages and file space managment enabled */
+ /* Create an fcpl with shared messages and file space management enabled */
fcpl_shmesg_all = H5Pcreate(H5P_FILE_CREATE);
H5Pset_shared_mesg_nindexes(fcpl_shmesg_all, 1);
H5Pset_shared_mesg_index(fcpl_shmesg_all, 0, H5O_SHMESG_ALL_FLAG, 20);
diff --git a/test/chunk_info.c b/test/chunk_info.c
index 5c16a7d..f60cddd 100644
--- a/test/chunk_info.c
+++ b/test/chunk_info.c
@@ -291,7 +291,7 @@ index_type_str(H5D_chunk_index_t idx_type)
/*-------------------------------------------------------------------------
* Function: verify_selected_chunks (helper function)
*
- * Purpose: Reads the chunks within the boundery {start,end} and verify
+ * Purpose: Reads the chunks within the boundary {start,end} and verify
* the values against the populated data.
*
* Return: Success: SUCCEED
@@ -321,7 +321,7 @@ verify_selected_chunks(hid_t dset, hid_t plist, const hsize_t *start, const hsiz
for (jj = 0; jj < CHUNK_NY; jj++)
expected_buf[n][ii][jj] = (int)(ii * jj) + 1;
- /* Read each chunk within the boundery of {start,end} and verify the
+ /* Read each chunk within the boundary of {start,end} and verify the
values against the expected data */
chk_index = 0;
for (ii = start[0]; ii < end[0]; ii++)
diff --git a/test/cmpd_dset.c b/test/cmpd_dset.c
index 3235056..024b6c0 100644
--- a/test/cmpd_dset.c
+++ b/test/cmpd_dset.c
@@ -713,7 +713,7 @@ test_compound(char *filename, hid_t fapl)
/*
*######################################################################
* Step 12: Write an array into the middle third of the dataset
- * initializeing only members `b' and `d' to -1.
+ * initializing only members `b' and `d' to -1.
*/
TESTING("hyperslab part initialized write");
diff --git a/test/cork.c b/test/cork.c
index 1b99384..f4b8d1c 100644
--- a/test/cork.c
+++ b/test/cork.c
@@ -97,7 +97,7 @@ verify_old_dset_cork(void)
hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */
hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */
hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */
- int ** buf = NULL; /* Data bufer (pointers to fake 2D array) */
+ int ** buf = NULL; /* Data buffer (pointers to fake 2D array) */
int * buf_data = NULL; /* Data buffer (actual data) */
int i = 0, j = 0; /* Local index variables */
H5O_info2_t oinfo, oinfo2, oinfo3; /* Object metadata information */
@@ -509,7 +509,7 @@ verify_dset_cork(hbool_t swmr, hbool_t new_format)
hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */
hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */
hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */
- int ** buf = NULL; /* Data bufer (pointers to fake 2D array) */
+ int ** buf = NULL; /* Data buffer (pointers to fake 2D array) */
int * buf_data = NULL; /* Data buffer (actual data) */
int i = 0, j = 0; /* Local index variables */
H5O_info2_t oinfo, oinfo2, oinfo3; /* Object metadata information */
@@ -1883,7 +1883,7 @@ test_dset_cork(hbool_t swmr, hbool_t new_format)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
hid_t fapl = H5I_INVALID_HID; /* File access property list */
- hid_t gid = H5I_INVALID_HID; /* Groupd ID */
+ hid_t gid = H5I_INVALID_HID; /* Grouped ID */
hid_t did1 = H5I_INVALID_HID, did2 = H5I_INVALID_HID; /* Dataset IDs */
hid_t tid1 = H5I_INVALID_HID, tid2 = H5I_INVALID_HID; /* Datatype IDs */
hid_t sid = H5I_INVALID_HID; /* Dataspace ID */
diff --git a/test/direct_chunk.c b/test/direct_chunk.c
index b2b7cba..d4a1e7a 100644
--- a/test/direct_chunk.c
+++ b/test/direct_chunk.c
@@ -686,7 +686,7 @@ filter_bogus1(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= (int)ADD_ON;
buf_left -= sizeof(int);
@@ -722,7 +722,7 @@ filter_bogus2(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts,
size_t buf_left = *buf_size; /* Amount of data buffer left to process */
if (flags & H5Z_FLAG_REVERSE) { /* read */
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ /= (int)FACTOR;
buf_left -= sizeof(int);
@@ -1427,7 +1427,7 @@ error:
* Function: test_direct_chunk_read_no_cache
*
* Purpose: Test the basic functionality of H5Dread_chunk with the
- * chunk cache diabled.
+ * chunk cache disabled.
*
* Return: Success: 0
* Failure: 1
@@ -1963,7 +1963,7 @@ error:
* Function: test_read_unallocated_chunk
*
* Purpose: Tests the H5Dread_chunk and H5Dget_chunk_storage_size with valid
- * offets to chunks that have not been written to the dataset and are
+ * offsets to chunks that have not been written to the dataset and are
* not allocated in the chunk storage on disk.
*
* Return: Success: 0
@@ -2014,7 +2014,7 @@ test_read_unallocated_chunk(hid_t file)
if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0)
FAIL_STACK_ERROR;
- /* Write a single chunk to intialize the chunk storage */
+ /* Write a single chunk to initialize the chunk storage */
HDmemset(direct_buf, 0, CHUNK_NX * CHUNK_NY * sizeof(int));
offset[0] = 0;
offset[1] = 0;
@@ -2031,7 +2031,7 @@ test_read_unallocated_chunk(hid_t file)
offset[0] = i * CHUNK_NX;
offset[1] = j * CHUNK_NY;
- /* Read a non-existant chunk using the direct read function. */
+ /* Read a non-existent chunk using the direct read function. */
H5E_BEGIN_TRY
{
status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, &direct_buf);
@@ -2042,7 +2042,7 @@ test_read_unallocated_chunk(hid_t file)
if (status != -1)
TEST_ERROR
- /* Query the size of the non-existant chunk */
+ /* Query the size of the non-existent chunk */
direct_chunk_nbytes = ULONG_MAX;
H5E_BEGIN_TRY
{
diff --git a/test/dsets.c b/test/dsets.c
index 0423898..b1ebf58 100644
--- a/test/dsets.c
+++ b/test/dsets.c
@@ -2009,7 +2009,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2076,7 +2076,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2140,7 +2140,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -2207,7 +2207,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32,
if (status >= 0)
TEST_ERROR;
- /* Callback decides to continue inspite data is corrupted. */
+ /* Callback decides to continue in spite data is corrupted. */
if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0)
TEST_ERROR;
if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0)
@@ -3242,7 +3242,7 @@ test_nbit_int(hid_t file)
power = HDpow(2.0, (double)(precision - 1));
orig_data[i][j] = (int)(((long long)HDrandom() % (long long)power) << offset);
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -3820,7 +3820,7 @@ test_nbit_compound(hid_t file)
orig_data[i][j].s = (short)(((long long)HDrandom() % (long long)power) << offset[2]);
orig_data[i][j].f = float_val[i][j];
- /* some even-numbered integer values are negtive */
+ /* some even-numbered integer values are negative */
if ((i * size[1] + j + 1) % 2 == 0) {
orig_data[i][j].i = -orig_data[i][j].i;
orig_data[i][j].s = (short)-orig_data[i][j].s;
@@ -4534,7 +4534,7 @@ test_nbit_int_size(hid_t file)
goto error;
}
- /* Initiliaze data buffer with random data within correct range
+ /* Initialize data buffer with random data within correct range
* corresponding to the memory datatype's precision and offset.
*/
for (i = 0; i < DSET_DIM1; i++)
@@ -4747,7 +4747,7 @@ test_nbit_flt_size(hid_t file)
} /* end if */
/*
- * Initiliaze data buffer with random data
+ * Initialize data buffer with random data
*/
for (i = 0; i < DSET_DIM1; i++)
for (j = 0; j < DSET_DIM2; j++)
@@ -4906,7 +4906,7 @@ test_scaleoffset_int(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = HDrandom() % 10000;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5050,7 +5050,7 @@ test_scaleoffset_int_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (int)HDrandom() % 10000;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -5172,7 +5172,7 @@ test_scaleoffset_float(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = (float)(HDrandom() % 100000) / 1000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5318,7 +5318,7 @@ test_scaleoffset_float_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (float)(HDrandom() % 100000) / 1000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -5439,7 +5439,7 @@ test_scaleoffset_double(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[i][j] = (float)(HDrandom() % 10000000) / 10000000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((i * size[1] + j + 1) % 2 == 0)
orig_data[i][j] = -orig_data[i][j];
}
@@ -5585,7 +5585,7 @@ test_scaleoffset_double_2(hid_t file)
for (j = 0; j < (size_t)size[1]; j++) {
orig_data[0][j] = (float)(HDrandom() % 10000000) / 10000000.0F;
- /* even-numbered values are negtive */
+ /* even-numbered values are negative */
if ((j + 1) % 2 == 0)
orig_data[0][j] = -orig_data[0][j];
}
@@ -8897,7 +8897,7 @@ test_chunk_cache(hid_t fapl)
if (H5Pclose(dapl2) < 0)
FAIL_STACK_ERROR
- /* Similary, test use of H5Dcreate2 with H5P_DEFAULT */
+ /* Similarly, test use of H5Dcreate2 with H5P_DEFAULT */
if (H5Dclose(dsid) < 0)
FAIL_STACK_ERROR
if ((dsid = H5Dcreate2(fid, "dset2", H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT)) < 0)
@@ -11982,7 +11982,7 @@ error:
* Purpose: Checks that a file created with either:
* (a) SWMR-write + non-latest-format
* (b) write + latest format
- * will generate datset with latest chunk indexing type.
+ * will generate dataset with latest chunk indexing type.
*
* Return: Success: 0
* Failure: -1
@@ -14716,7 +14716,7 @@ test_versionbounds(void)
hid_t dcpl = -1; /* Dataset creation property list */
hid_t srcspace = -1; /* Source dataspaces */
hid_t vspace = -1; /* Virtual dset dataspaces */
- hid_t srcdset = -1; /* Source datset */
+ hid_t srcdset = -1; /* Source dataset */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[1] = {3}; /* Data space current size */
char srcfilename[FILENAME_BUF_SIZE];
@@ -14852,7 +14852,7 @@ error:
/*-----------------------------------------------------------------------------
* Function: test_object_header_minimization_dcpl
*
- * Purpose: Test the "datset object header minimization" property as part of
+ * Purpose: Test the "dataset object header minimization" property as part of
* the DCPL.
*
* Return: Success/pass: 0
@@ -15195,7 +15195,7 @@ main(void)
double rdcc_w0;
int nerrors = 0;
const char *envval;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
int i;
/* Don't run this test using certain file drivers */
@@ -15203,7 +15203,7 @@ main(void)
if (envval == NULL)
envval = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0);
/* Set the random # seed */
diff --git a/test/dt_arith.c b/test/dt_arith.c
index 3f47153..e6927f5 100644
--- a/test/dt_arith.c
+++ b/test/dt_arith.c
@@ -27,7 +27,7 @@
#define FP_EPSILON 0.000001F
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
@@ -223,7 +223,7 @@ static int without_hardware_g = 0;
/*The number of values if multiplied by 10 for each step.*/ \
num_norm = (SRC_MAX_10_EXP - SRC_MIN_10_EXP); \
/*Reduce the number of values by 2^factor. MULTIPLY=10^(2^factor). Using this algorithm \
- *instead of arithmatic operation to avoid any conversion*/ \
+ *instead of arithmetic operation to avoid any conversion*/ \
num_norm >>= factor; \
\
/*Total number of values*/ \
@@ -640,7 +640,7 @@ test_particular_fp_integer(void)
float src_f = (float)INT_MAX;
int dst_i;
int fill_value = 13;
- int endian; /*endianess */
+ int endian; /*endianness */
unsigned int fails_this_test = 0;
size_t j;
@@ -818,7 +818,7 @@ test_derived_flt(void)
size_t src_size, dst_size;
unsigned char *buf = NULL, *saved_buf = NULL;
int * aligned = NULL;
- int endian; /*endianess */
+ int endian; /*endianness */
size_t nelmts = NTESTELEM;
unsigned int fails_this_test = 0;
const size_t max_fails = 40; /*max number of failures*/
@@ -1267,7 +1267,7 @@ test_derived_integer(void)
char filename[1024];
size_t src_size, dst_size;
unsigned char *buf = NULL, *saved_buf = NULL;
- int endian; /*endianess */
+ int endian; /*endianness */
size_t nelmts = NTESTELEM;
unsigned int fails_this_test = 0;
const size_t max_fails = 40; /*max number of failures*/
@@ -1591,7 +1591,7 @@ test_conv_int_1(const char *name, hid_t src, hid_t dst)
dtype_t src_type, dst_type; /*data types */
const char * src_type_name = NULL; /*source type name */
const char * dst_type_name = NULL; /*destination type name */
- int endian; /*machine endianess */
+ int endian; /*machine endianness */
size_t src_size, dst_size; /*type sizes */
unsigned char * buf = NULL; /*buffer for conversion */
unsigned char * saved = NULL; /*original values */
@@ -2831,8 +2831,8 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst)
int overflow = 0; /*overflow occurred */
int uflow = 0; /*underflow debug counters*/
size_t j, k; /*counters */
- int sendian; /* source type endianess */
- int dendian; /* Destination type endianess */
+ int sendian; /* source type endianness */
+ int dendian; /* Destination type endianness */
size_t dst_ebias; /* Destination type's exponent bias */
size_t src_epos; /* Source type's exponent position */
size_t src_esize; /* Source type's exponent size */
@@ -3147,7 +3147,7 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst)
/*
* Assume same if both results are NaN. There are many NaN bit
- * patterns and the software doesn't attemt to emulate the
+ * patterns and the software doesn't attempt to emulate the
* hardware in this regard. Instead, software uses a single bit
* pattern for NaN by setting the significand to all ones.
*/
@@ -3426,8 +3426,8 @@ test_conv_int_fp(const char *name, int run_test, hid_t src, hid_t dst)
dtype_t dst_type; /*data types */
const char * src_type_name = NULL; /*source type name */
const char * dst_type_name = NULL; /*destination type name */
- int sendian; /*source endianess */
- int dendian; /*destination endianess */
+ int sendian; /*source endianness */
+ int dendian; /*destination endianness */
size_t src_size, dst_size; /*type sizes */
unsigned char * buf = NULL; /*buffer for conversion */
unsigned char * saved = NULL; /*original values */
diff --git a/test/dtypes.c b/test/dtypes.c
index bf2ad5a..59ad0e3 100644
--- a/test/dtypes.c
+++ b/test/dtypes.c
@@ -29,7 +29,7 @@
#define ARRAY_DIM 4
/*
- * Offset from alinged memory returned by malloc(). This can be used to test
+ * Offset from aligned memory returned by malloc(). This can be used to test
* that type conversions handle non-aligned buffers correctly.
*/
#define ALIGNMENT 1
@@ -2637,7 +2637,7 @@ test_compound_13(void)
data_out.x[u] = (unsigned char)u;
data_out.y = 99.99F;
- /* Set latest_format in access propertly list to enable the latest
+ /* Set latest_format in access property list to enable the latest
* compound datatype format.
*/
if ((fapl_id = H5Pcreate(H5P_FILE_ACCESS)) < 0)
@@ -3792,7 +3792,7 @@ error:
*
* Modifications:
* Raymond Lu
- * Wednesday, Febuary 9, 2005
+ * Wednesday, February 9, 2005
* Added test for H5Tenum_valueof, H5Tenum_nameof, and
* H5Tget_member_value.
*-------------------------------------------------------------------------
@@ -6845,7 +6845,7 @@ test_set_order(void)
{
hid_t dtype; /* Datatype ID */
H5T_order_t order; /* Byte order */
- hsize_t dims[2] = {3, 4}; /* Array dimenstions */
+ hsize_t dims[2] = {3, 4}; /* Array dimensions */
herr_t ret; /* Generic return value */
TESTING("H5Tset/get_order");
@@ -7093,7 +7093,7 @@ test_set_order_compound(hid_t fapl)
hid_t file = -1;
hid_t cmpd = -1, memb_cmpd = -1, memb_array1 = -1, memb_array2 = -1, cmpd_array = -1;
hid_t vl_id = -1;
- hsize_t dims[2] = {3, 4}; /* Array dimenstions */
+ hsize_t dims[2] = {3, 4}; /* Array dimensions */
char filename[1024];
herr_t ret; /* Generic return value */
@@ -7948,7 +7948,7 @@ test_deprec(hid_t fapl)
unsigned u; /* Local index variable */
herr_t status; /* Generic routine value */
- TESTING("deprected API routines for datatypes");
+ TESTING("deprecated API routines for datatypes");
/* Create an array datatype with an atomic base type */
/* (dimension permutations allowed, but not stored) */
diff --git a/test/earray.c b/test/earray.c
index 65664fb..4ed5b7f 100644
--- a/test/earray.c
+++ b/test/earray.c
@@ -1284,7 +1284,7 @@ typedef struct eiter_fw_t {
/*-------------------------------------------------------------------------
* Function: eiter_fw_init
*
- * Purpose: Initialize element interator (forward iteration)
+ * Purpose: Initialize element iterator (forward iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1428,7 +1428,7 @@ eiter_fw_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa
/*-------------------------------------------------------------------------
* Function: eiter_fw_term
*
- * Purpose: Shut down element interator (forward iteration)
+ * Purpose: Shut down element iterator (forward iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1472,7 +1472,7 @@ typedef struct eiter_rv_t {
/*-------------------------------------------------------------------------
* Function: eiter_rv_init
*
- * Purpose: Initialize element interator (reverse iteration)
+ * Purpose: Initialize element iterator (reverse iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1653,7 +1653,7 @@ eiter_rv_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa
/*-------------------------------------------------------------------------
* Function: eiter_rv_term
*
- * Purpose: Shut down element interator (reverse iteration)
+ * Purpose: Shut down element iterator (reverse iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1694,7 +1694,7 @@ typedef struct eiter_rnd_t {
/*-------------------------------------------------------------------------
* Function: eiter_rnd_init
*
- * Purpose: Initialize element interator (random iteration)
+ * Purpose: Initialize element iterator (random iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1803,7 +1803,7 @@ eiter_rnd_max(const void *in_eiter)
/*-------------------------------------------------------------------------
* Function: eiter_rnd_term
*
- * Purpose: Shut down element interator (random iteration)
+ * Purpose: Shut down element iterator (random iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1843,7 +1843,7 @@ static const earray_iter_t ea_iter_rnd = {
/*-------------------------------------------------------------------------
* Function: eiter_rnd2_init
*
- * Purpose: Initialize element interator (random #2 iteration)
+ * Purpose: Initialize element iterator (random #2 iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1926,7 +1926,7 @@ typedef struct eiter_cyc_t {
/*-------------------------------------------------------------------------
* Function: eiter_cyc_init
*
- * Purpose: Initialize element interator (cyclic iteration)
+ * Purpose: Initialize element iterator (cyclic iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -2019,7 +2019,7 @@ eiter_cyc_max(const void *in_eiter)
/*-------------------------------------------------------------------------
* Function: eiter_cyc_term
*
- * Purpose: Shut down element interator (cyclic iteration)
+ * Purpose: Shut down element iterator (cyclic iteration)
*
* Return: Success: 0
* Failure: -1
diff --git a/test/enc_dec_plist_cross_platform.c b/test/enc_dec_plist_cross_platform.c
index 9ea5cf9..3030d7c 100644
--- a/test/enc_dec_plist_cross_platform.c
+++ b/test/enc_dec_plist_cross_platform.c
@@ -28,7 +28,7 @@ int
main(void)
{
if (VERBOSE_MED)
- HDprintf("Encode/Decode property list endianess\n");
+ HDprintf("Encode/Decode property list endianness\n");
/******* ENCODE/DECODE DCPLS *****/
TESTING("Default DCPL Encoding/Decoding");
diff --git a/test/enum.c b/test/enum.c
index 3923892..512fd9d 100644
--- a/test/enum.c
+++ b/test/enum.c
@@ -552,7 +552,7 @@ test_value_dsnt_exist(void)
if ((datatype_id = H5Tenum_create(H5T_NATIVE_INT)) < 0)
goto error;
- /* These calls should fail, since no memebrs exist yet */
+ /* These calls should fail, since no members exist yet */
if (H5Tenum_valueof(datatype_id, "SAX", &val) >= 0)
goto error;
val = 3;
diff --git a/test/external.c b/test/external.c
index 1e4e32a..11185b3 100644
--- a/test/external.c
+++ b/test/external.c
@@ -1217,7 +1217,7 @@ test_path_relative_cwd(hid_t fapl)
}
H5E_END_TRY;
if (dset3 >= 0)
- FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded");
+ FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded");
/* Read the entire dataset and compare with the original */
HDmemset(whole, 0, sizeof(whole));
@@ -1233,7 +1233,7 @@ test_path_relative_cwd(hid_t fapl)
if (H5Dclose(dset) < 0)
FAIL_STACK_ERROR
- /* Open dataset (use a differend prefix than for create.
+ /* Open dataset (use a different prefix than for create.
* This works because the dataset was closed.
*/
if (H5Pset_efile_prefix(dapl2, "${ORIGIN}/.") < 0)
@@ -1254,7 +1254,7 @@ test_path_relative_cwd(hid_t fapl)
}
H5E_END_TRY;
if (dset3 >= 0)
- FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded");
+ FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded");
/* Read the entire dataset and compare with the original */
HDmemset(whole, 0, sizeof(whole));
diff --git a/test/farray.c b/test/farray.c
index fdf384b..44dda99 100644
--- a/test/farray.c
+++ b/test/farray.c
@@ -969,7 +969,7 @@ typedef struct fiter_fw_t {
/*-------------------------------------------------------------------------
* Function: fiter_fw_init
*
- * Purpose: Initialize element interator (forward iteration)
+ * Purpose: Initialize element iterator (forward iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1021,7 +1021,7 @@ fiter_fw_next(void *_fiter)
/*-------------------------------------------------------------------------
* Function: fiter_term
*
- * Purpose: Shut down element interator (simple iterators)
+ * Purpose: Shut down element iterator (simple iterators)
*
* Return: Success: 0
* Failure: -1
@@ -1055,7 +1055,7 @@ typedef struct fiter_rv_t {
/*-------------------------------------------------------------------------
* Function: fiter_rv_init
*
- * Purpose: Initialize element interator (reverse iteration)
+ * Purpose: Initialize element iterator (reverse iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1120,7 +1120,7 @@ typedef struct fiter_rnd_t {
/*-------------------------------------------------------------------------
* Function: fiter_rnd_init
*
- * Purpose: Initialize element interator (random iteration)
+ * Purpose: Initialize element iterator (random iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
@@ -1193,7 +1193,7 @@ fiter_rnd_next(void *_fiter)
/*-------------------------------------------------------------------------
* Function: fiter_rnd_term
*
- * Purpose: Shut down element interator (random iteration)
+ * Purpose: Shut down element iterator (random iteration)
*
* Return: Success: 0
* Failure: -1
@@ -1235,7 +1235,7 @@ typedef struct fiter_cyc_t {
/*-------------------------------------------------------------------------
* Function: fiter_cyc_init
*
- * Purpose: Initialize element interator (cyclic iteration)
+ * Purpose: Initialize element iterator (cyclic iteration)
*
* Return: Success: Pointer to iteration status object
* Failure: NULL
diff --git a/test/fheap.c b/test/fheap.c
index c98034f..5ca45fc 100644
--- a/test/fheap.c
+++ b/test/fheap.c
@@ -2470,7 +2470,7 @@ test_id_limits(hid_t fapl, H5HF_create_t *cparam, hid_t fcpl)
size_t id_len; /* Size of fractal heap IDs */
size_t tiny_max_len; /* Max. length of tiny objects */
hbool_t tiny_len_extended; /* Do tiny objects use two bytes for the length? */
- hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */
+ hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */
/* Set the filename to use for this test (dependent on fapl) */
h5_fixname(FILENAME[0], fapl, filename, sizeof(filename));
@@ -13786,7 +13786,7 @@ test_filtered_huge(hid_t fapl, H5HF_create_t *cparam, fheap_test_param_t *tparam
fheap_heap_state_t state; /* State of fractal heap */
unsigned deflate_level; /* Deflation level */
size_t old_actual_id_len = 0; /* Old actual ID length */
- hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */
+ hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */
hbool_t pline_init = FALSE; /* Whether the I/O pipeline has been initialized */
/* Test description */
const char *base_desc = "insert 'huge' object into heap with I/O filters, then remove %s";
@@ -16362,7 +16362,7 @@ main(void)
unsigned num_pb_fs = 1; /* The number of settings to test for page buffering and file space handling */
int ExpressMode; /* Express testing level */
const char *envval; /* Environment variable */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
hbool_t api_ctx_pushed = FALSE; /* Whether API context pushed */
/* Don't run this test using certain file drivers */
@@ -16370,7 +16370,7 @@ main(void)
if (envval == NULL)
envval = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0);
/* Reset library */
diff --git a/test/file_image.c b/test/file_image.c
index 0373468..6950bfb 100644
--- a/test/file_image.c
+++ b/test/file_image.c
@@ -145,7 +145,7 @@ test_properties(void)
/* Check that sizes are the same, and that the buffers are identical but separate */
VERIFY(temp_size == size, "Sizes of buffers don't match");
- VERIFY(temp2 != NULL, "Recieved buffer not set");
+ VERIFY(temp2 != NULL, "Received buffer not set");
VERIFY(temp2 != buffer, "Retrieved buffer is the same as original");
VERIFY(temp2 != temp, "Retrieved buffer is the same as previously retrieved buffer");
VERIFY(0 == HDmemcmp(temp2, buffer, size), "Buffers contain different data");
@@ -260,7 +260,7 @@ free_cb(void *ptr, H5FD_file_image_op_t op, void *udata)
* Function: udata_copy_cb
*
* Purpose: This function allows calls to the udata_copy callback to be tracked.
- * No copying actualy takes place; it is easier to deal with only one
+ * No copying actually takes place; it is easier to deal with only one
* instance of the udata.
*
* Returns: A pointer to the same udata that was passed in.
@@ -954,7 +954,7 @@ test_get_file_image(const char *test_banner, const int file_name_num, hid_t fapl
err = H5Fclose(core_file_id);
VERIFY(err == SUCCEED, "H5Fclose(core_file_id) failed.");
- /* dicard core fapl */
+ /* discard core fapl */
err = H5Pclose(core_fapl_id);
VERIFY(err == SUCCEED, "H5Pclose(core_fapl_id) failed.");
@@ -1078,7 +1078,7 @@ test_get_file_image_error_rejection(void)
VERIFY(bytes_read < 0, "H5Fget_file_image(2 -- test 1) succeeded.");
/* Call H5Fget_file_image() with good buffer and buffer size,
- * but non-existant file_id. Should fail.
+ * but non-existent file_id. Should fail.
*/
H5E_BEGIN_TRY
{
diff --git a/test/fillval.c b/test/fillval.c
index 8de6ef1..b2e88bc 100644
--- a/test/fillval.c
+++ b/test/fillval.c
@@ -1918,7 +1918,7 @@ test_extend_cases(hid_t file, hid_t _dcpl, const char *dset_name, const hsize_t
if ((fspace = H5Dget_space(dset)) < 0)
TEST_ERROR
- /* Create dataspace for single element sized bufer */
+ /* Create dataspace for single element sized buffer */
if ((mspace = H5Screate_simple(5, one, NULL)) < 0)
TEST_ERROR
@@ -2142,7 +2142,7 @@ test_extend(hid_t fapl, const char *base_name, H5D_layout_t layout)
#if 1
/*
- * Remove this when contiguous datasets can be exended to some
+ * Remove this when contiguous datasets can be extended to some
* predetermined fininte size, even if it's just in the slowest varying
* dimension. If it's removed prematurely then you'll get one of the
* errors described above or `unable to select fill value region'.
diff --git a/test/filter_fail.c b/test/filter_fail.c
index a1bcf28..64cc41a 100644
--- a/test/filter_fail.c
+++ b/test/filter_fail.c
@@ -160,7 +160,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled)
points[i] = i;
/* Write data. If the chunk cache is enabled, H5Dwrite should succeed. If it is
- * diabled, H5Dwrite should fail. */
+ * disabled, H5Dwrite should fail. */
if (cache_enabled) {
if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, sid, H5P_DEFAULT, points) < 0)
TEST_ERROR
@@ -186,7 +186,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled)
TEST_ERROR
/* Close dataset. If the chunk cache is enabled, the flushing of chunks should fail
- * during H5Dclose. If it is diabled, H5Dwrite should fail but H5Dclose should succeed. */
+ * during H5Dclose. If it is disabled, H5Dwrite should fail but H5Dclose should succeed. */
if (cache_enabled) {
H5E_BEGIN_TRY
{
diff --git a/test/filter_plugin1_dsets.c b/test/filter_plugin1_dsets.c
index 5d6c1ef..29b6e73 100644
--- a/test/filter_plugin1_dsets.c
+++ b/test/filter_plugin1_dsets.c
@@ -79,7 +79,7 @@ add_sub_value(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_value
value = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) {
- /* READ - Substract the given value from all the data values */
+ /* READ - Subtract the given value from all the data values */
while (buf_left > 0) {
*int_ptr++ -= value;
buf_left -= sizeof(int);
diff --git a/test/filter_plugin3_dsets.c b/test/filter_plugin3_dsets.c
index 618ce06..1d399e5 100644
--- a/test/filter_plugin3_dsets.c
+++ b/test/filter_plugin3_dsets.c
@@ -108,7 +108,7 @@ add_sub_value_hdf5(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
value = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) {
- /* READ - Substract the given value from all the data values */
+ /* READ - Subtract the given value from all the data values */
while (buf_left > 0) {
*int_ptr++ -= value;
buf_left -= sizeof(int);
diff --git a/test/flushrefresh.c b/test/flushrefresh.c
index 6bf6260..0010cfd 100644
--- a/test/flushrefresh.c
+++ b/test/flushrefresh.c
@@ -63,10 +63,10 @@
#define NOT_FLUSHED "NOT_FLUSHED"
/* Error Handling */
-/* For errors occuring in the main process, use the standard TEST_ERROR macro.
+/* For errors occurring in the main process, use the standard TEST_ERROR macro.
For errors occurring in the spawned process (from the test script), use
the PROCESS_ERROR macro, which will send a signal to the main process so the
- main process can propogate errors correctly. */
+ main process can propagate errors correctly. */
FILE *errorfile;
#define ERRFILE "flushrefresh_ERROR"
#define PROCESS_ERROR \
@@ -222,7 +222,7 @@ test_flush(void)
* object, a series of verifications will occur on each object in the file.
*
* Each verification consists of spawning off a new process and determining
- * if the object can be opened and its information retreived in said
+ * if the object can be opened and its information retrieved in said
* alternate process. It reports the results, which are compared to an
* expected value (either that the object can be found on disk, or that it
* cannot).
@@ -695,7 +695,7 @@ test_refresh(void)
* that an attribute has been added, and is only visible after a
* successful call to a H5*refresh function).
*
- * As with the flush case, the implemention is a bit tricky as it's
+ * As with the flush case, the implementation is a bit tricky as it's
* dealing with signals going back and forth between the two processes
* to ensure the timing is correct, but basically, an example:
*
@@ -1166,7 +1166,7 @@ error:
* Purpose: This function is used to communicate with the verification
* process spawned by the start_refresh_verification_process
* function. It gives it the go-ahead to call H5*refresh
- * on an object and conlcude the refresh verification.
+ * on an object and conclude the refresh verification.
*
* Return: 0 on Success, 1 on Failure
*
@@ -1371,7 +1371,7 @@ error:
*
* Purpose: This function checks the status of external verification
* processes to see if they've succeeded. It checks for the
- * existance of flushrefresh_ERROR file. If present, that indicates
+ * existence of flushrefresh_ERROR file. If present, that indicates
* an external verification process has failed, and this function
* thus fails as well. If not present, then nothing else has
* failed, and this function succeeds.
diff --git a/test/gen_bounds.c b/test/gen_bounds.c
index 8148f16..87a9c18 100644
--- a/test/gen_bounds.c
+++ b/test/gen_bounds.c
@@ -858,12 +858,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_REG_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
@@ -886,12 +886,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_IRR_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT,
H5P_DEFAULT, H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
@@ -915,12 +915,12 @@ gen_sel_files(const char *filename, H5F_libver_t low_bound, H5F_libver_t high_bo
if (H5Rcreate(&ref_wbuf[0], fid, DATASET, H5R_DATASET_REGION, sid) < 0)
TEST_ERROR;
- /* Create the reference datset */
+ /* Create the reference dataset */
if ((ref_did = H5Dcreate2(fid, SEL_EX_PT_DSET, H5T_STD_REF_DSETREG, ref_sid, H5P_DEFAULT, H5P_DEFAULT,
H5P_DEFAULT)) < 0)
TEST_ERROR;
- /* Write to the reference datset */
+ /* Write to the reference dataset */
if (H5Dwrite(ref_did, H5T_STD_REF_DSETREG, H5S_ALL, H5S_ALL, H5P_DEFAULT, ref_wbuf) < 0)
TEST_ERROR;
diff --git a/test/gen_old_group.c b/test/gen_old_group.c
index 1b4e0b2..15e67a4 100644
--- a/test/gen_old_group.c
+++ b/test/gen_old_group.c
@@ -19,7 +19,7 @@
* empty "symbol table" group.
*
* This file is used in the v1.7 branch (after the "compact group"
- * checkin) to test compatability. Compile and run this
+ * checkin) to test compatibility. Compile and run this
* program (with the 1.6.x branch), it will generate a file*
* called "group_old.h5". You need to move it to the test
* directory in the HDF5 v1.7 source tree. The test/stab.c
diff --git a/test/gen_plist.c b/test/gen_plist.c
index 10c6423..25d48cf 100644
--- a/test/gen_plist.c
+++ b/test/gen_plist.c
@@ -81,7 +81,7 @@ main(void)
H5AC_cache_image_config_t my_cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE,
-1};
- /* check endianess */
+ /* check endianness */
{
short int word = 0x0001;
char * byte = (char *)&word;
diff --git a/test/genall5.c b/test/genall5.c
index 4fa6c01..53d1cdd 100644
--- a/test/genall5.c
+++ b/test/genall5.c
@@ -14,7 +14,7 @@
/* Programmer: John Mainzer
* 9/23/15
*
- * This file contains a heavily edited and functionaly reduce
+ * This file contains a heavily edited and functionally reduce
* version of the test code first written by Quincey in a file
* of the same name.
*/
diff --git a/test/getname.c b/test/getname.c
index 200c426..698079d 100644
--- a/test/getname.c
+++ b/test/getname.c
@@ -1526,7 +1526,7 @@ test_main(hid_t file_id, hid_t fapl)
{
char name[NAME_BUF_SIZE]; /* Buffer to hold name and its size */
- /* Get name for non commited datatype, it should fail */
+ /* Get name for non committed datatype, it should fail */
H5E_BEGIN_TRY
{
if (H5Iget_name(type_id, name, NAME_BUF_SIZE) > 0)
diff --git a/test/gheap.c b/test/gheap.c
index fa761b4..c83ff68 100644
--- a/test/gheap.c
+++ b/test/gheap.c
@@ -353,7 +353,7 @@ error:
* Function: test_4
*
* Purpose: Tests the H5HG_remove() feature by writing lots of objects
- * and occassionally removing some. When we're done they're all
+ * and occasionally removing some. When we're done they're all
* removed.
*
* Return: Success: 0
diff --git a/test/h5test.c b/test/h5test.c
index bb2f5cc..0a7f2fd 100644
--- a/test/h5test.c
+++ b/test/h5test.c
@@ -1828,7 +1828,7 @@ h5_send_message(const char *send, const char *arg1, const char *arg2)
* signal file from disk, and only continues once it has
* successfully done so (i.e., only after another process has
* called the "h5_send_message" function to write the signal file).
- * This functon will then immediately remove the file (i.e.,
+ * This function will then immediately remove the file (i.e.,
* to indicate that it has been received and can be reused),
* and then exits, allowing the calling function to continue.
*
diff --git a/test/hdfs.c b/test/hdfs.c
index 5465e77..be67ee4 100644
--- a/test/hdfs.c
+++ b/test/hdfs.c
@@ -1454,7 +1454,7 @@ test_noops_and_autofails(void)
JSVERIFY(SUCCEED, H5FDlock(file, FALSE), "lock issue")
JSVERIFY(SUCCEED, H5FDunlock(file), "unlock issue")
/* Lock/unlock with null file or similar error crashes tests.
- * HDassert in calling heirarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
+ * HDassert in calling hierarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
*/
/************
diff --git a/test/links.c b/test/links.c
index 2de1fca..2aeef14 100644
--- a/test/links.c
+++ b/test/links.c
@@ -8199,7 +8199,7 @@ external_link_dangling(hid_t fapl, hbool_t new_format)
if (status >= 0) {
H5_FAILED();
HDputs(
- " Retreiving name of object by index through dangling file external link should have failed.");
+ " Retrieving name of object by index through dangling file external link should have failed.");
} /* end if */
/* Close root group */
@@ -8672,7 +8672,7 @@ external_link_abstar(hid_t fapl, hbool_t new_format)
}
H5E_END_TRY;
- /* should be able to find the target file with abolute path */
+ /* should be able to find the target file with absolute path */
if (gid < 0) {
H5_FAILED();
HDputs(" Should have found the file in tmp_links directory.");
@@ -16591,7 +16591,7 @@ link_filters(hid_t fapl, hbool_t new_format)
TEST_ERROR
filesize_unfiltered = h5_get_file_size(filename, fapl);
- /* Set deflate fitler */
+ /* Set deflate filter */
if (H5Pset_deflate(fcpl, 6) < 0)
TEST_ERROR
@@ -22556,7 +22556,7 @@ main(void)
for (minimize_dset_oh = 0; minimize_dset_oh <= 1; minimize_dset_oh++) {
if (minimize_dset_oh) {
- HDprintf("\n-Testing with minimzed dataset object headers-\n");
+ HDprintf("\n-Testing with minimized dataset object headers-\n");
dcpl_g = H5Pcreate(H5P_DATASET_CREATE);
if (0 > dcpl_g)
TEST_ERROR
diff --git a/test/mf.c b/test/mf.c
index 9eb7807..facfeb7 100644
--- a/test/mf.c
+++ b/test/mf.c
@@ -223,7 +223,7 @@ test_mf_eoa(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MM_alloc() of file allocation");
@@ -371,7 +371,7 @@ test_mf_eoa_shrink(const char *env_h5_drvr, hid_t fapl)
haddr_t addr = 0;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_shrink() of file allocation: test 1");
@@ -658,7 +658,7 @@ test_mf_eoa_extend(const char *env_h5_drvr, hid_t fapl)
htri_t was_extended;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_extend() of file allocation: test 1");
@@ -1934,7 +1934,7 @@ test_mf_fs_extend(hid_t fapl)
if (check_stats(f, f->shared->fs_man[H5FD_MEM_SUPER], &state))
TEST_ERROR
- /* Remove section A from free-space manger */
+ /* Remove section A from free-space manager */
if (H5MF__find_sect(f, H5FD_MEM_SUPER, (hsize_t)(TBLOCK_SIZE30 - 10), f->shared->fs_man[H5FD_MEM_SUPER],
&tmp) != TRUE)
TEST_ERROR
@@ -2012,7 +2012,7 @@ test_mf_fs_absorb(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
H5MF_free_section_t *sect_node = NULL;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("A free-space section absorbs an aggregator: test 1");
@@ -2188,7 +2188,7 @@ test_mf_aggr_alloc1(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2;
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 1");
@@ -2339,7 +2339,7 @@ test_mf_aggr_alloc2(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, addr3;
haddr_t ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 2");
@@ -2497,7 +2497,7 @@ test_mf_aggr_alloc3(const char *env_h5_drvr, hid_t fapl)
hsize_t ma_size = 0, new_ma_size = 0;
haddr_t sdata_addr = HADDR_UNDEF;
hsize_t sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator: test 3");
@@ -2665,7 +2665,7 @@ test_mf_aggr_alloc4(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, saddr1, saddr2, saddr3;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 4");
@@ -2814,7 +2814,7 @@ test_mf_aggr_alloc5(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1, addr2, addr3;
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 5");
@@ -2952,7 +2952,7 @@ test_mf_aggr_alloc6(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
H5FS_stat_t state;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 6");
@@ -3126,7 +3126,7 @@ test_mf_aggr_alloc7(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, sdata_size = 0;
H5FS_stat_t state;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_alloc() of meta/sdata aggregator:test 7");
@@ -3297,7 +3297,7 @@ test_mf_aggr_extend(const char *env_h5_drvr, hid_t fapl)
haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0;
htri_t was_extended;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_extend() of meta/sdata aggregator: test 1");
@@ -3595,7 +3595,7 @@ test_mf_aggr_absorb(const char *env_h5_drvr, hid_t fapl)
haddr_t new_sdata_addr = HADDR_UNDEF;
hsize_t ma_size = 0, new_ma_size = 0;
hsize_t sdata_size = 0, new_sdata_size = 0;
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
TESTING("H5MF_try_shrink() of meta/sdata aggregator: test 1");
@@ -5838,7 +5838,7 @@ error:
*
* Alignment = 4096 aggr->alloc_size = 2048
*
- * Allocate first block (30) from meta_aggr: (meta_aggr is emtpy)
+ * Allocate first block (30) from meta_aggr: (meta_aggr is empty)
* request-size is > meta_aggr->size and < meta_aggr->alloc_size
* Result:
* A block of meta_aggr->alloc_size is allocated from file allocation
@@ -7105,7 +7105,7 @@ test_mf_fs_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format)
else
TESTING("File's free-space is going away with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
if (contig_addr_vfd) {
@@ -7296,7 +7296,7 @@ test_mf_strat_thres_persist(const char *env_h5_drvr, hid_t fapl, hbool_t new_for
else
TESTING("File space strategy/persisting/threshold with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Set the filename to use for this test (dependent on fapl) */
@@ -7466,7 +7466,7 @@ test_mf_strat_thres_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format
else
TESTING("File space merge/shrink for section size < threshold with old library format")
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Set the filename to use for this test (dependent on fapl) */
@@ -8024,7 +8024,7 @@ test_page_try_shrink(const char *env_h5_drvr, hid_t fapl)
haddr_t addr1; /* Address for small metadata block */
haddr_t saddr1; /* Address for small raw data block */
haddr_t gaddr1; /* Address for large data block */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t status; /* status from shrinking */
h5_stat_size_t file_size; /* File size */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8154,7 +8154,7 @@ test_page_small_try_extend(const char *env_h5_drvr, hid_t fapl)
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t addr1, addr2, addr3; /* Addresses for small metadata blocks */
haddr_t saddr1; /* Address for small raw data block */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t was_extended; /* Whether the block can be extended or not */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8333,7 +8333,7 @@ test_page_large_try_extend(const char *env_h5_drvr, hid_t fapl)
hid_t fcpl = -1; /* File creation property list */
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for large data blocks */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
htri_t was_extended; /* Whether the block can be extended or not */
char filename[FILENAME_LEN]; /* Filename to use */
@@ -8496,7 +8496,7 @@ test_page_large(const char *env_h5_drvr, hid_t fapl)
hid_t fcpl = -1; /* File creation property list ID */
H5F_t * f = NULL; /* Internal file object pointer */
haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for blocks */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
H5FS_stat_t fs_stat; /* Information for free-space manager */
h5_stat_size_t file_size; /* File size */
char filename[FILENAME_LEN]; /* Filename to use */
diff --git a/test/mirror_vfd.c b/test/mirror_vfd.c
index 3556cd2..7047146 100644
--- a/test/mirror_vfd.c
+++ b/test/mirror_vfd.c
@@ -128,7 +128,7 @@ mybzero(void *dest, size_t size)
* Function: _populate_filepath
*
* Purpose: Given a directory name and a base name, concatenate the two and
- * run h5fixname() to get the "actual" path to the intented target.
+ * run h5fixname() to get the "actual" path to the intended target.
* `h5suffix' should be FALSE to keep the base name unaltered;
* TRUE will append the '.h5' h5suffix to the basename...
* FALSE -> h5fixname_no_suffix(), TRUE -> h5fixname()
diff --git a/test/mount.c b/test/mount.c
index c5ac9b3..b23de30 100644
--- a/test/mount.c
+++ b/test/mount.c
@@ -753,7 +753,7 @@ test_move(hid_t fapl)
H5E_END_TRY;
if (status >= 0) {
H5_FAILED();
- HDputs(" Moving an object across files should't have been possible");
+ HDputs(" Moving an object across files shouldn't have been possible");
TEST_ERROR
} /* end if */
@@ -1835,7 +1835,7 @@ test_mount_after_unmount(hid_t fapl)
if (HDstrcmp(objname, "/X/M/Y") != 0)
TEST_ERROR
- /* Rename object in file #3 that is "disconnected" from name hiearchy */
+ /* Rename object in file #3 that is "disconnected" from name hierarchy */
/* (It is "disconnected" because it's parent file has been unmounted) */
if (H5Lmove(gidAMX, "M/Y", gidAMX, "M/Z", H5P_DEFAULT, H5P_DEFAULT) < 0)
TEST_ERROR
diff --git a/test/mtime.c b/test/mtime.c
index 43dc502..fdc499a 100644
--- a/test/mtime.c
+++ b/test/mtime.c
@@ -116,7 +116,7 @@ main(void)
if (0 == oi1.ctime) {
SKIPPED();
HDputs(" The modification time could not be decoded on this OS.");
- HDputs(" Modification times will be mantained in the file but");
+ HDputs(" Modification times will be maintained in the file but");
HDputs(" cannot be queried on this system. See H5O_mtime_decode().");
return 0;
}
diff --git a/test/objcopy.c b/test/objcopy.c
index 9aae621..c70bbea 100644
--- a/test/objcopy.c
+++ b/test/objcopy.c
@@ -2789,7 +2789,7 @@ test_copy_dataset_simple_empty(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): empty and openend contiguous dataset");
+ TESTING("H5Ocopy(): empty and opened contiguous dataset");
}
else {
TESTING("H5Ocopy(): empty contiguous dataset");
@@ -8679,7 +8679,7 @@ test_copy_soft_link(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t dst_fa
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): openend object through soft link");
+ TESTING("H5Ocopy(): opened object through soft link");
}
else {
TESTING("H5Ocopy(): object through soft link");
@@ -9674,7 +9674,7 @@ test_copy_dataset_contig_named_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl
char src_filename[NAME_BUF_SIZE];
char dst_filename[NAME_BUF_SIZE];
- TESTING("H5Ocopy(): contigous dataset with named VLEN datatype");
+ TESTING("H5Ocopy(): contiguous dataset with named VLEN datatype");
/* set initial data values */
for (i = 0; i < DIM_SIZE_1; i++) {
@@ -10434,10 +10434,10 @@ test_copy_dataset_contig_vl_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h
char dst_filename[NAME_BUF_SIZE];
if (test_open) {
- TESTING("H5Ocopy(): contigous and opened dataset with nested VLEN datatype");
+ TESTING("H5Ocopy(): contiguous and opened dataset with nested VLEN datatype");
}
else {
- TESTING("H5Ocopy(): contigous dataset with nested VLEN datatype");
+ TESTING("H5Ocopy(): contiguous dataset with nested VLEN datatype");
}
/* set initial data values */
@@ -13763,7 +13763,7 @@ test_copy_cdt_hier_merge(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t d
TEST_ERROR
/*
- * Test 3: copy datsets in /g0 at SRC to DST group /uncopied
+ * Test 3: copy datasets in /g0 at SRC to DST group /uncopied
*/
if (H5Ocopy(fid_src, NAME_GROUP_TOP "/" SRC_DSET1, fid_dst, NAME_GROUP_UNCOPIED "/" SRC_DSET1, ocpypl_id,
H5P_DEFAULT) < 0)
diff --git a/test/ohdr.c b/test/ohdr.c
index e0b5e88..ecea18e 100644
--- a/test/ohdr.c
+++ b/test/ohdr.c
@@ -584,7 +584,7 @@ test_unknown(unsigned bogus_id, char *filename, hid_t fapl)
TESTING("object with unknown header message & 'shareable' flag set");
- /* Open the dataset with the unknown header message, adn "shareable" flag */
+ /* Open the dataset with the unknown header message, and "shareable" flag */
if ((did = H5Dopen2(loc_bogus, "Dataset5", H5P_DEFAULT)) < 0)
FAIL_STACK_ERROR
if (H5Dclose(did) < 0)
@@ -1089,13 +1089,13 @@ test_minimized_dset_ohdr_size_comparisons(hid_t fapl_id)
/* IDs for non-minimized file open */
hid_t file_f_id = -1; /* lower 'f' for standard file setting */
hid_t dset_f_x_id = -1; /* 'x' for default */
- hid_t dset_f_N_id = -1; /* 'N' for explcit non-minimized dset */
+ hid_t dset_f_N_id = -1; /* 'N' for explicit non-minimized dset */
hid_t dset_f_Y_id = -1; /* 'Y' for minimized dset */
/* IDs for minimized file open */
hid_t file_F_id = -1; /* upper 'F' for minimized file setting */
hid_t dset_F_x_id = -1; /* 'x' for default */
- hid_t dset_F_N_id = -1; /* 'N' for explcit non-minimized dset */
+ hid_t dset_F_N_id = -1; /* 'N' for explicit non-minimized dset */
hid_t dset_F_Y_id = -1; /* 'Y' for minimized dset */
char filename_a[512] = "";
diff --git a/test/page_buffer.c b/test/page_buffer.c
index 37f2acc..8918639 100644
--- a/test/page_buffer.c
+++ b/test/page_buffer.c
@@ -75,7 +75,7 @@ const char *FILENAME[] = {"filepaged", NULL};
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -215,7 +215,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -375,7 +375,7 @@ error:
* Any data mis-matches or unexpected failures or successes
* reported by the HDF5 library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -549,7 +549,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -827,7 +827,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -1081,7 +1081,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -1502,7 +1502,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 2)
TEST_ERROR;
- /* adding more meta entires should replace meta entries since raw data
+ /* adding more meta entries should replace meta entries since raw data
* is at its minimum
*/
if (H5F_block_write(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 600), sizeof(int) * 100, data) < 0)
@@ -1517,7 +1517,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 2)
TEST_ERROR;
- /* bring existing raw entires up the LRU */
+ /* bring existing raw entries up the LRU */
if (H5F_block_read(f, H5FD_MEM_DRAW, raw_addr + (sizeof(int) * 750), sizeof(int) * 100, data) < 0)
FAIL_STACK_ERROR;
@@ -1533,7 +1533,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr)
if (f->shared->page_buf->raw_count != 3)
TEST_ERROR;
- /* adding 2 meta entries should replace 2 entires at the bottom of the LRU */
+ /* adding 2 meta entries should replace 2 entries at the bottom of the LRU */
if (H5F_block_read(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 98), sizeof(int) * 100, data) < 0)
FAIL_STACK_ERROR;
@@ -1716,7 +1716,7 @@ error:
* Any data mis-matches or failures reported by the HDF5
* library result in test failure.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
@@ -2019,7 +2019,7 @@ error:
* At present, page buffering should be disabled in parallel
* builds. Verify this.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: John Mainzer
@@ -2134,7 +2134,7 @@ error:
*
* Purpose: Main function for the page buffer tests.
*
- * Return: 0 if test is sucessful
+ * Return: 0 if test is successful
* 1 if test fails
*
* Programmer: unknown
diff --git a/test/ros3.c b/test/ros3.c
index 2d4bcf0..1f56d70 100644
--- a/test/ros3.c
+++ b/test/ros3.c
@@ -1456,7 +1456,7 @@ test_noops_and_autofails(void)
JSVERIFY(SUCCEED, H5FDlock(file, FALSE), NULL)
JSVERIFY(SUCCEED, H5FDunlock(file), NULL)
/* Lock/unlock with null file or similar error crashes tests.
- * HDassert in calling heirarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
+ * HDassert in calling hierarchy, `H5FD[un]lock()` and `H5FD_[un]lock()`
*/
/************
@@ -1772,7 +1772,7 @@ main(void)
bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL");
if (bucket_url_env == NULL || bucket_url_env[0] == '\0') {
- HDprintf("WARNING: S3 bucket url is not defined in enviornment "
+ HDprintf("WARNING: S3 bucket url is not defined in environment "
"variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n");
}
else {
diff --git a/test/s3comms.c b/test/s3comms.c
index 22d647f..03b8ce6 100644
--- a/test/s3comms.c
+++ b/test/s3comms.c
@@ -244,7 +244,7 @@ jserr_str(const char *expected, const char *actual, const char *reason)
}
#ifdef JSVERIFY_EXP_ACT
-/* VERIFY rountines with paramter order (<expected>, <actual> [, <msg> ])
+/* VERIFY routines with parameter order (<expected>, <actual> [, <msg> ])
*/
/*----------------------------------------------------------------------------
@@ -435,7 +435,7 @@ error:
*
* Purpose:
*
- * Demonstrate the construction of a Canoncial Request (and Signed Headers)
+ * Demonstrate the construction of a Canonical Request (and Signed Headers)
*
* Elided / not yet implemented:
* Query strings
@@ -730,7 +730,7 @@ test_hrb_init_request(void)
FALSE,
},
{
- "null verb substitues to GET",
+ "null verb substitutes to GET",
NULL,
"/MYPATH/MYFILE.tiff",
"/MYPATH/MYFILE.tiff",
@@ -1144,9 +1144,9 @@ test_hrb_node_set(void)
for (mock_i = 0; test->given[mock_i] != NULL; mock_i += 2) {
const char *name = test->given[mock_i];
- const char *valu = test->given[mock_i + 1];
+ const char *value = test->given[mock_i + 1];
- FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, valu))
+ FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, value))
}
/********
* TEST *
@@ -1163,10 +1163,10 @@ test_hrb_node_set(void)
mock_i = 0;
while (test->expected[mock_i] != NULL && node != NULL) {
const char *name = test->expected[mock_i];
- const char *valu = test->expected[mock_i + 1];
+ const char *value = test->expected[mock_i + 1];
JSVERIFY_STR(name, node->name, NULL)
- JSVERIFY_STR(valu, node->value, NULL)
+ JSVERIFY_STR(value, node->value, NULL)
mock_i += 2;
node = node->next;
@@ -1559,7 +1559,7 @@ test_parse_url(void)
NULL,
"a=b&d=b",
},
- "QUERY with implict PATH",
+ "QUERY with implicit PATH",
},
{
"http://[5]/path?a=b&d=b",
@@ -2014,7 +2014,7 @@ error:
* H5FD_s3comms_s3r_read << called by getsize(), multiple times working
* H5FD_s3comms_s3r_close
*
- * Shows most basic curl interation.
+ * Shows most basic curl iteration.
*
* Programmer: Jacob Smith
* 2017-10-06
@@ -2610,7 +2610,7 @@ main(void)
bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL");
if (bucket_url_env == NULL || bucket_url_env[0] == '\0') {
- HDprintf("WARNING: S3 bucket url is not defined in enviornment "
+ HDprintf("WARNING: S3 bucket url is not defined in environment "
"variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n");
}
else {
@@ -2618,7 +2618,7 @@ main(void)
s3_test_bucket_defined = TRUE;
}
- /* tests ordered rougly by dependence */
+ /* tests ordered roughly by dependence */
nerrors += test_macro_format_credential() < 0 ? 1 : 0;
nerrors += test_trim() < 0 ? 1 : 0;
nerrors += test_nlowercase() < 0 ? 1 : 0;
diff --git a/test/set_extent.c b/test/set_extent.c
index 3078c2d..44bf8cb 100644
--- a/test/set_extent.c
+++ b/test/set_extent.c
@@ -115,12 +115,12 @@ main(void)
unsigned chunk_cache; /* Whether to enable chunk caching */
int nerrors = 0;
const char *env_h5_drvr; /* File Driver value from environment */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
env_h5_drvr = HDgetenv("HDF5_DRIVER");
if (env_h5_drvr == NULL)
env_h5_drvr = "nomatch";
- /* Current VFD that does not support contigous address space */
+ /* Current VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Initialize random number seed */
@@ -649,7 +649,7 @@ test_rank1(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
HDprintf("\n");
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++)
if (buf_s[i] != buf_o[i]) {
HDprintf("buf_s[%d] = %d\n", i, buf_s[i]);
@@ -1055,7 +1055,7 @@ test_rank2(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
}
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
if (buf_s[i][j] != buf_o[i][j]) {
@@ -1536,7 +1536,7 @@ test_rank3(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f
HDprintf("\n");
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
for (k = 0; k < (int)dims_r[2]; k++) {
@@ -1734,7 +1734,7 @@ test_external(hid_t fapl)
hsize_t dims_r[RANK2]; /* read dimensions */
hsize_t maxdims[RANK2] = {DIME0, DIM1}; /* only the first dimension can be extendible */
int buf_o[DIM0][DIM1]; /* original buffer, for writing */
- int buf_s[DIMS0][DIMS1]; /* shrinked buffer, for reading */
+ int buf_s[DIMS0][DIMS1]; /* shrunk buffer, for reading */
int buf_e[DIME0][DIM1]; /* extended buffer, for writing, dimension 1 is the original */
int buf_ro[DIM0][DIM1]; /* original buffer for reading */
int i, j;
@@ -1946,7 +1946,7 @@ test_external(hid_t fapl)
}
#endif
- /* compare the read array with the shrinked array */
+ /* compare the read array with the shrunk array */
for (i = 0; i < (int)dims_r[0]; i++) {
for (j = 0; j < (int)dims_r[1]; j++) {
if (buf_s[i][j] != buf_o[i][j]) {
diff --git a/test/stab.c b/test/stab.c
index 798619d..b11abd5 100644
--- a/test/stab.c
+++ b/test/stab.c
@@ -125,7 +125,7 @@ test_misc(hid_t fcpl, hid_t fapl, hbool_t new_format)
if (H5Gclose(g3) < 0)
TEST_ERROR
- /* Open all groups with absolute names to check for exsistence */
+ /* Open all groups with absolute names to check for existence */
if ((g1 = H5Gopen2(fid, "/test_1a", H5P_DEFAULT)) < 0)
TEST_ERROR
if ((g2 = H5Gopen2(fid, "/test_1a/sub_1", H5P_DEFAULT)) < 0)
@@ -1416,7 +1416,7 @@ main(void)
hid_t fcpl, fcpl2; /* File creation property list ID */
unsigned new_format; /* Whether to use the new format or not */
const char *env_h5_drvr; /* File Driver value from environment */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
int nerrors = 0;
/* Get the VFD to use */
@@ -1424,7 +1424,7 @@ main(void)
if (env_h5_drvr == NULL)
env_h5_drvr = "nomatch";
- /* VFD that does not support contigous address space */
+ /* VFD that does not support contiguous address space */
contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0);
/* Reset library */
diff --git a/test/swmr.c b/test/swmr.c
index b609fb6..5294267 100644
--- a/test/swmr.c
+++ b/test/swmr.c
@@ -3666,7 +3666,7 @@ test_append_flush_dataset_chunked(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dcpl = -1; /* A copy of dataset creation property */
hid_t dapl = -1; /* A copy of dataset access property */
@@ -3889,7 +3889,7 @@ test_append_flush_dataset_fixed(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dapl = -1; /* A copy of dataset access property */
hid_t ddapl = -1; /* The dataset access property of the opened dataset */
@@ -4107,7 +4107,7 @@ test_append_flush_dataset_multiple(hid_t in_fapl)
{
hid_t fid = -1; /* file ID */
hid_t fapl = -1; /* A copy of file access property */
- hid_t did1 = -1, did2 = -1; /* The datset ID */
+ hid_t did1 = -1, did2 = -1; /* The dataset ID */
hid_t sid = -1; /* The dataspace ID */
hid_t dcpl = -1; /* A copy of dataset creation property */
hid_t dapl1 = -1; /* A copy of dataset access property */
@@ -7412,7 +7412,7 @@ main(void)
/*
* Modify the following routines to test for files:
* H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format)
- * --both result in v3 superblock and latest version suppport
+ * --both result in v3 superblock and latest version support
*/
nerrors += test_start_swmr_write(fapl, TRUE);
nerrors += test_start_swmr_write(fapl, FALSE);
@@ -7446,7 +7446,7 @@ main(void)
*
* Modify the following 2 routines to test for files:
* H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format)
- * --both result in v3 superblock and latest version suppport
+ * --both result in v3 superblock and latest version support
*/
nerrors += test_file_lock_swmr_same(fapl);
nerrors += test_file_lock_swmr_concur(fapl);
diff --git a/test/swmr_generator.c b/test/swmr_generator.c
index a455f3f..93cfa0f 100644
--- a/test/swmr_generator.c
+++ b/test/swmr_generator.c
@@ -221,7 +221,7 @@ gen_skeleton(const char *filename, hbool_t verbose, hbool_t swmr_write, int comp
if (verbose)
HDfprintf(stderr, "Closing objects\n");
- /* Close everythign */
+ /* Close everything */
if (H5Pclose(dcpl) < 0)
return -1;
if (H5Sclose(sid) < 0)
diff --git a/test/swmr_start_write.c b/test/swmr_start_write.c
index 3e84dcc..d046b40 100644
--- a/test/swmr_start_write.c
+++ b/test/swmr_start_write.c
@@ -125,7 +125,7 @@ create_file(const char *filename, hbool_t verbose, FILE *verbose_file, unsigned
*
* Parameters:
* fid: file ID for the SWMR test file
- * comp_level: the compresssion level
+ * comp_level: the compression level
* index_type: The chunk index type (b1 | b2 | ea | fa)
* verbose: whether verbose console output is desired.
* verbose_file: file pointer for verbose output
diff --git a/test/tattr.c b/test/tattr.c
index e75eeb5..974367c 100644
--- a/test/tattr.c
+++ b/test/tattr.c
@@ -5858,7 +5858,7 @@ test_attr_info_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for query on non-existant attribute */
+ /* Check for query on non-existent attribute */
ret = H5Aget_info_by_idx(my_dataset, ".", H5_INDEX_CRT_ORDER, H5_ITER_INC, (hsize_t)0, &ainfo,
H5P_DEFAULT);
VERIFY(ret, FAIL, "H5Aget_info_by_idx");
@@ -6346,7 +6346,7 @@ test_attr_delete_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for deleting non-existant attribute */
+ /* Check for deleting non-existent attribute */
ret = H5Adelete_by_idx(my_dataset, ".", idx_type, order, (hsize_t)0, H5P_DEFAULT);
VERIFY(ret, FAIL, "H5Adelete_by_idx");
@@ -8001,7 +8001,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for opening a non-existant attribute on an object with no attributes */
+ /* Check for opening a non-existent attribute on an object with no attributes */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -8040,7 +8040,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
is_dense = H5O__is_attr_dense_test(my_dataset);
VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test");
- /* Check for opening a non-existant attribute on an object with compact attribute storage */
+ /* Check for opening a non-existent attribute on an object with compact attribute storage */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -8121,7 +8121,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl)
VERIFY(name_count, (max_compact * 2), "H5O__attr_dense_info_test");
} /* end if */
- /* Check for opening a non-existant attribute on an object with dense attribute storage */
+ /* Check for opening a non-existent attribute on an object with dense attribute storage */
ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT);
VERIFY(ret_id, FAIL, "H5Aopen");
@@ -10989,7 +10989,7 @@ test_attr_bug8(hid_t fcpl, hid_t fapl)
** chunk. Do this twice: once with only attributes in the
** object header chunk and once with a (small) soft link in
** the chunk as well. In both cases, the shrunk chunk will
-** initally be too small and a new NULL message must be
+** initially be too small and a new NULL message must be
** created.
**
****************************************************************/
diff --git a/test/tcoords.c b/test/tcoords.c
index fef8689..b2d68d3 100644
--- a/test/tcoords.c
+++ b/test/tcoords.c
@@ -17,7 +17,7 @@
*
* Test the element coordinates for dataspace selection. For
* chunked dataset, when the hyperslab selection of some
- * dimensions is full, the library optimize it by "flattenning"
+ * dimensions is full, the library optimize it by "flattening"
* the fully selected dimensions. This program tests if the
* coordinates of selected elements are correctly calculated.
*
diff --git a/test/testmeta.c b/test/testmeta.c
index 7cb3000..a3b5160 100644
--- a/test/testmeta.c
+++ b/test/testmeta.c
@@ -15,7 +15,7 @@
* This program illustrates assertion errors when linked
* to HDF5 1.4.1-post2 or 1.4.2-pre3 debug library.
*
- * If the assertion errors are ignored, the program eventially causes
+ * If the assertion errors are ignored, the program eventually causes
* an error in H5Gcreate2 when writing object 83381.
*
* When writing in single file mode, the assertion errors still occur
diff --git a/test/tfile.c b/test/tfile.c
index 8c3f9b0..d49a64c 100644
--- a/test/tfile.c
+++ b/test/tfile.c
@@ -225,7 +225,7 @@ test_file_create(void)
/* First ensure the file does not exist */
HDremove(FILE1);
- /* Try opening a non-existant file */
+ /* Try opening a non-existent file */
fid1 = H5Fopen(FILE1, H5F_ACC_RDWR, H5P_DEFAULT);
VERIFY(fid1, FAIL, "H5Fopen");
@@ -1774,7 +1774,7 @@ test_file_is_accessible(const char *env_h5_drvr)
**
** test_file_ishdf5(): low-level file test routine.
** This test checks whether the H5Fis_hdf5() routine is working
-** correctly in variuous situations.
+** correctly in various situations.
**
*****************************************************************/
#ifndef H5_NO_DEPRECATED_SYMBOLS
@@ -1998,7 +1998,7 @@ test_file_delete(hid_t fapl_id)
**
** test_file_open_dot(): low-level file test routine.
** This test checks whether opening objects with "." for a name
-** works correctly in variuous situations.
+** works correctly in various situations.
**
*****************************************************************/
static void
@@ -2380,7 +2380,7 @@ test_file_double_root_open(void)
grp2_id = H5Gopen2(file2_id, "/", H5P_DEFAULT);
CHECK(grp2_id, FAIL, "H5Gopen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Gclose(grp1_id);
CHECK(ret, FAIL, "H5Gclose");
ret = H5Gclose(grp2_id);
@@ -2419,7 +2419,7 @@ test_file_double_group_open(void)
grp2_id = H5Gopen2(file2_id, GRP_NAME, H5P_DEFAULT);
CHECK(grp2_id, FAIL, "H5Gopen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Gclose(grp1_id);
CHECK(ret, FAIL, "H5Gclose");
ret = H5Gclose(grp2_id);
@@ -2468,7 +2468,7 @@ test_file_double_dataset_open(void)
ret = H5Sclose(space_id);
CHECK(ret, FAIL, "H5Sclose");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Dclose(dset1_id);
CHECK(ret, FAIL, "H5Dclose");
ret = H5Dclose(dset2_id);
@@ -2869,7 +2869,7 @@ test_file_double_datatype_open(void)
type2_id = H5Topen2(file2_id, TYPE_NAME, H5P_DEFAULT);
CHECK(type2_id, FAIL, "H5Topen2");
- /* Note "assymetric" close order */
+ /* Note "asymmetric" close order */
ret = H5Tclose(type1_id);
CHECK(ret, FAIL, "H5Tclose");
ret = H5Tclose(type2_id);
@@ -2914,7 +2914,7 @@ test_userblock_file_size(void)
ret = H5Pset_userblock(fcpl2_id, USERBLOCK_SIZE);
CHECK(ret, FAIL, "H5Pset_userblock");
- /* Create files. Onyl file2 with have a userblock. */
+ /* Create files. Only file2 with have a userblock. */
file1_id = H5Fcreate(FILE1, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
CHECK(file1_id, FAIL, "H5Fcreate");
file2_id = H5Fcreate(FILE2, H5F_ACC_TRUNC, fcpl2_id, H5P_DEFAULT);
@@ -3835,7 +3835,7 @@ test_userblock_alignment_paged(void)
* Strategy is H5F_FILE_SPACE_NONE; fsp_size = 1024
* H5Pset_alignment() is 16
* Outcome:
- * Should succed:
+ * Should succeed:
* userblock (512) is integral multiple of alignment (16)
*/
/* Create file creation property list with user block */
@@ -3874,7 +3874,7 @@ test_userblock_alignment_paged(void)
* H5Pset_alignment() is 3
* Reopen the file; H5Pset_alignment() is 1024
* Outcome:
- * Should succed:
+ * Should succeed:
* Userblock (512) is the same as alignment (512);
* The H5Pset_alignment() calls have no effect
*/
@@ -3943,7 +3943,7 @@ test_filespace_info(const char *env_h5_drvr)
hsize_t fs_threshold; /* Free-space section threshold--iteration variable */
hsize_t fsp_size; /* File space page size */
char filename[FILENAME_LEN]; /* Filename to use */
- hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */
+ hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */
herr_t ret; /* Return value */
/* Output message about test being performed */
@@ -5185,7 +5185,7 @@ test_filespace_1_10_0_compatible(void)
** Verify that the trunk can open, read and modify these files--
** 1) They are initially created (via gen_filespace.c) in the trunk
** with combinations of file space strategies, default/non-default
-** threshold, and file spacing paging enabled/disbled.
+** threshold, and file spacing paging enabled/disabled.
** The library creates the file space info message with
** "mark if unknown" in these files.
** 2) They are copied to the 1.8 branch, and are opened/read/modified
@@ -6345,7 +6345,7 @@ test_libver_bounds_dataset(hid_t fapl)
VERIFY(dset->shared->dcpl_cache.fill.version, H5O_fill_ver_bounds[low], "H5O_fill_ver_bounds");
}
- /* Verify filter pipleline message version */
+ /* Verify filter pipeline message version */
VERIFY(dset->shared->dcpl_cache.pline.version, H5O_pline_ver_bounds[low], "H5O_pline_ver_bounds");
/* Close the dataset */
@@ -6398,7 +6398,7 @@ test_libver_bounds_dataset(hid_t fapl)
ret = H5Sclose(sid);
CHECK(ret, FAIL, "H5Sclose");
- /* Close the datset creation property list */
+ /* Close the dataset creation property list */
ret = H5Pclose(dcpl);
CHECK(ret, FAIL, "H5Pclose");
@@ -6418,7 +6418,7 @@ test_libver_bounds_dataset(hid_t fapl)
/* Loop through all the combinations of low/high bounds in new_fapl */
/* Open the file with the fapl and create the chunked dataset */
- /* Verify the dataset's layout, fill value and filter pipleline message versions */
+ /* Verify the dataset's layout, fill value and filter pipeline message versions */
for (low = H5F_LIBVER_EARLIEST; low < H5F_LIBVER_NBOUNDS; low++) {
for (high = H5F_LIBVER_EARLIEST; high < H5F_LIBVER_NBOUNDS; high++) {
H5E_BEGIN_TRY
@@ -6523,7 +6523,7 @@ test_libver_bounds_dataset(hid_t fapl)
** new fapl:
** --Open the same file in (a) with the fapl
** --Create a chunked dataset, a compact dataset and
-** a contigous dataset
+** a contiguous dataset
** --Verify the dataspace message version for these
** three datasets
** --Delete the three datasets and the dataspaces
@@ -6872,7 +6872,7 @@ static void
test_libver_bounds_datatype_check(hid_t fapl, hid_t tid)
{
hid_t fid = H5I_INVALID_HID; /* File ID */
- hid_t new_fapl = H5I_INVALID_HID; /* File acess property list */
+ hid_t new_fapl = H5I_INVALID_HID; /* File access property list */
hid_t dcpl = H5I_INVALID_HID; /* Dataset creation property list */
hid_t dtid = H5I_INVALID_HID; /* Datatype ID for the dataset */
hid_t str_tid = H5I_INVALID_HID; /* String datatype ID */
@@ -7132,7 +7132,7 @@ test_libver_bounds_attributes(hid_t fapl)
fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, H5P_DEFAULT, fapl);
CHECK(fid, H5I_INVALID_HID, "H5Fcreate");
- /* Integer datatpye */
+ /* Integer datatype */
tid = H5Tcopy(H5T_NATIVE_INT);
CHECK(tid, H5I_INVALID_HID, "H5Tcopy");
@@ -7241,7 +7241,7 @@ test_libver_bounds_attributes(hid_t fapl)
fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, fcpl, fapl);
CHECK(fid, H5I_INVALID_HID, "H5Fcreate");
- /* Create an integer datatye */
+ /* Create an integer datatype */
tid = H5Tcopy(H5T_NATIVE_INT);
CHECK(tid, H5I_INVALID_HID, "H5Tcopy");
@@ -8037,7 +8037,7 @@ test_file(void)
test_libver_macros(); /* Test the macros for library version comparison */
test_libver_macros2(); /* Test the macros for library version comparison */
test_incr_filesize(); /* Test H5Fincrement_filesize() and H5Fget_eoa() */
- test_min_dset_ohdr(); /* Test datset object header minimization */
+ test_min_dset_ohdr(); /* Test dataset object header minimization */
#ifndef H5_NO_DEPRECATED_SYMBOLS
test_file_ishdf5(env_h5_drvr); /* Test detecting HDF5 files correctly */
test_deprec(); /* Test deprecated routines */
diff --git a/test/tgenprop.c b/test/tgenprop.c
index 55e5a9b..f54fbbc 100644
--- a/test/tgenprop.c
+++ b/test/tgenprop.c
@@ -107,7 +107,7 @@ test_genprop_basic_class(void)
ret = H5Pequal(cid2, H5P_ROOT);
VERIFY(ret, 1, "H5Pequal");
- /* Make certain false postives aren't being returned */
+ /* Make certain false positives aren't being returned */
ret = H5Pequal(cid2, H5P_FILE_CREATE);
VERIFY(ret, 0, "H5Pequal");
@@ -185,7 +185,7 @@ test_genprop_basic_class_prop(void)
CHECK_I(ret, "H5Pget_nprops");
VERIFY(nprops, 0, "H5Pget_nprops");
- /* Check the existance of the first property (should fail) */
+ /* Check the existence of the first property (should fail) */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 0, "H5Pexist");
@@ -199,7 +199,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister2");
- /* Check the existance of the first property */
+ /* Check the existence of the first property */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -223,7 +223,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister2");
- /* Check the existance of the second property */
+ /* Check the existence of the second property */
ret = H5Pexist(cid1, PROP2_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -242,7 +242,7 @@ test_genprop_basic_class_prop(void)
H5Pregister2(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL);
CHECK_I(ret, "H5Pregister2");
- /* Check the existance of the third property */
+ /* Check the existence of the third property */
ret = H5Pexist(cid1, PROP3_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1942,7 +1942,7 @@ test_genprop_deprec_class(void)
CHECK_I(ret, "H5Pget_nprops");
VERIFY(nprops, 0, "H5Pget_nprops");
- /* Check the existance of the first property (should fail) */
+ /* Check the existence of the first property (should fail) */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 0, "H5Pexist");
@@ -1954,7 +1954,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister1");
- /* Check the existance of the first property */
+ /* Check the existence of the first property */
ret = H5Pexist(cid1, PROP1_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1976,7 +1976,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
VERIFY(ret, FAIL, "H5Pregister1");
- /* Check the existance of the second property */
+ /* Check the existence of the second property */
ret = H5Pexist(cid1, PROP2_NAME);
VERIFY(ret, 1, "H5Pexist");
@@ -1994,7 +1994,7 @@ test_genprop_deprec_class(void)
ret = H5Pregister1(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL);
CHECK_I(ret, "H5Pregister1");
- /* Check the existance of the third property */
+ /* Check the existence of the third property */
ret = H5Pexist(cid1, PROP3_NAME);
VERIFY(ret, 1, "H5Pexist");
diff --git a/test/th5o.c b/test/th5o.c
index ebb6c6e..1afbe17 100644
--- a/test/th5o.c
+++ b/test/th5o.c
@@ -594,7 +594,7 @@ test_h5o_refcount(void)
CHECK(ret, FAIL, "H5Oget_info_by_name3");
VERIFY(oinfo.rc, 1, "reference count in H5Oget_info_by_name3");
- /* Increment the reference counts and then close the file to make sure the increment is permanant */
+ /* Increment the reference counts and then close the file to make sure the increment is permanent */
ret = H5Oincr_refcount(grp);
CHECK(ret, FAIL, "H5Oincr_refcount");
ret = H5Oincr_refcount(dtype);
diff --git a/test/th5s.c b/test/th5s.c
index 51216dc..1a6f56e 100644
--- a/test/th5s.c
+++ b/test/th5s.c
@@ -2022,7 +2022,7 @@ test_h5s_encode_points(H5F_libver_t low, H5F_libver_t high)
{
hid_t fapl = -1; /* File access property list ID */
hid_t sid; /* Dataspace ID */
- hsize_t numparticles = 8388608; /* Used to calculate dimenion size */
+ hsize_t numparticles = 8388608; /* Used to calculate dimension size */
unsigned num_dsets = 513; /* used to calculate dimension size */
hsize_t total_particles = numparticles * num_dsets;
hsize_t vdsdims[1] = {total_particles}; /* Dimension size */
diff --git a/test/tid.c b/test/tid.c
index c7e40a0..ab06840 100644
--- a/test/tid.c
+++ b/test/tid.c
@@ -41,7 +41,7 @@ basic_id_test(void)
int num_ref;
hsize_t num_members;
- /* Try to register an ID with ficticious types */
+ /* Try to register an ID with fictitious types */
H5E_BEGIN_TRY
arrayID = H5Iregister((H5I_type_t)420, testObj);
H5E_END_TRY
@@ -58,7 +58,7 @@ basic_id_test(void)
if (arrayID != H5I_INVALID_HID)
goto out;
- /* Try to access IDs with ficticious types */
+ /* Try to access IDs with fictitious types */
H5E_BEGIN_TRY
testPtr = H5Iobject_verify((hid_t)100, (H5I_type_t)0);
H5E_END_TRY
diff --git a/test/tmisc.c b/test/tmisc.c
index 3655bed..9d7d180 100644
--- a/test/tmisc.c
+++ b/test/tmisc.c
@@ -1280,7 +1280,7 @@ test_misc8(void)
for (v = 0; v < MISC8_DIM1; v++)
*tdata++ = (int)(((u * MISC8_DIM1) + v) % 13);
- /* Create a file acccess property list */
+ /* Create a file access property list */
fapl = H5Pcreate(H5P_FILE_ACCESS);
CHECK(fapl, FAIL, "H5Pcreate");
@@ -3892,7 +3892,7 @@ test_misc21(void)
/****************************************************************
**
-** test_misc22(): Test SZIP bits-per-pixel paramter.
+** test_misc22(): Test SZIP bits-per-pixel parameter.
** This should be set according to the datatype.
** Tests for precision and offset combo's.
**
diff --git a/test/trefer.c b/test/trefer.c
index 9b6c415..0e25ec2 100644
--- a/test/trefer.c
+++ b/test/trefer.c
@@ -2732,7 +2732,7 @@ test_reference_compat_conv(void)
hid_t sid1, sid2, sid3; /* Dataspace IDs */
hid_t tid1, tid2; /* Datatype ID */
hsize_t dims1[] = {SPACE1_DIM1}, dims2[] = {SPACE2_DIM1, SPACE2_DIM2},
- dims3[] = {SPACE1_DIM1}; /* Purposedly set dimension larger to test NULL references */
+ dims3[] = {SPACE1_DIM1}; /* Purposely set dimension larger to test NULL references */
hsize_t start[SPACE2_RANK]; /* Starting location of hyperslab */
hsize_t stride[SPACE2_RANK]; /* Stride of hyperslab */
hsize_t count[SPACE2_RANK]; /* Element count of hyperslab */
diff --git a/test/tselect.c b/test/tselect.c
index 31bb9c4..7d32407 100644
--- a/test/tselect.c
+++ b/test/tselect.c
@@ -1879,7 +1879,7 @@ verify_select_hyper_contig_dr__run_test(const uint16_t *cube_buf, size_t H5_ATTR
uint16_t expected_value; /* Expected value in dataset */
unsigned i, j, k, l, m; /* Local index variables */
size_t s; /* Local index variable */
- hbool_t mis_match; /* Flag to indicate mis-match in expected value */
+ hbool_t mis_match; /* Flag to indicate mismatch in expected value */
HDassert(cube_buf);
HDassert(cube_size > 0);
@@ -2357,7 +2357,7 @@ test_select_hyper_contig_dr__run_test(int test_num, const uint16_t *cube_buf, co
/* Now write the contents of the in memory small cube to slices of
* the on disk cube. After each write, read the on disk cube
- * into memeory, and verify that it contains the expected
+ * into memory, and verify that it contains the expected
* data. Verify that H5Sselect_shape_same() returns true on
* the memory and file selections.
*/
@@ -2742,7 +2742,7 @@ test_select_hyper_checker_board_dr__select_checker_board(hid_t tgt_n_cube_sid, u
i++;
} while ((i <= 1) && (0 >= sel_offset));
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Some how, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -3401,7 +3401,7 @@ test_select_hyper_checker_board_dr__run_test(int test_num, const uint16_t *cube_
/* Now write checker board selections of the entries in memory
* small cube to slices of the on disk cube. After each write,
- * read the on disk large cube * into memeory, and verify that
+ * read the on disk large cube * into memory, and verify that
* it contains the expected * data. Verify that
* H5Sselect_shape_same() returns true on the memory and file
* selections.
@@ -5256,7 +5256,7 @@ test_select_hyper_union_stagger(void)
error = H5Fclose(file_id);
CHECK(error, FAIL, "H5Fclose");
- /* Initialize intput buffer */
+ /* Initialize input buffer */
HDmemset(data_out, 0, 7 * 7 * sizeof(int));
/* Open file */
@@ -11254,7 +11254,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i
CHECK(ret, FAIL, "H5Sselect_hyperslab");
} /* end if */
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Some how, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -11349,7 +11349,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i
CHECK(ret, FAIL, "H5Sselect_hyperslab");
} /* end if */
- /* Wierdness alert:
+ /* Weirdness alert:
*
* Again, it seems that selections can extend beyond the
* boundaries of the target dataspace -- hence the following
@@ -12130,7 +12130,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 2-D hyperslab selection\n"));
/* Create 2-D dataspace */
@@ -12222,7 +12222,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 3-D hyperslab selection\n"));
@@ -12324,7 +12324,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 4-D hyperslab selection\n"));
@@ -12443,7 +12443,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
MESSAGE(7, ("Testing functionality to rebuild 5-D hyperslab selection\n"));
@@ -12572,7 +12572,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* We use 5-D to test a special case with
rebuilding routine TRUE, FALSE and TRUE */
@@ -12619,7 +12619,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* Adding some selections to make it real irregular */
start5[3] = 1;
@@ -12647,7 +12647,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
/* Add more selections to make it regular again */
start5[3] = 5;
@@ -12675,7 +12675,7 @@ test_space_rebuild(void)
ret = FAIL;
CHECK(ret, FAIL, "H5S_hyper_rebuild");
}
- /* No need to do shape comparision */
+ /* No need to do shape comparison */
H5Sclose(sid_reg1);
CHECK(ret, FAIL, "H5Sclose");
@@ -12981,7 +12981,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_update_diminfo");
} /* end if */
- /* Add block parially overlapping first, with OR */
+ /* Add block partially overlapping first, with OR */
start1[0] = 4;
count1[0] = 1;
block1[0] = 2;
@@ -13026,7 +13026,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_update_diminfo");
} /* end if */
- /* Add equally sized block parially overlapping current, with XOR */
+ /* Add equally sized block partially overlapping current, with XOR */
start1[0] = 3;
count1[0] = 1;
block1[0] = 5;
@@ -13060,7 +13060,7 @@ test_space_update_diminfo(void)
CHECK(ret, FAIL, "H5S_hyper_rebuild");
} /* end if */
- /* Add differently sized block parially overlapping current, with XOR */
+ /* Add differently sized block partially overlapping current, with XOR */
start1[0] = 4;
count1[0] = 1;
block1[0] = 5;
@@ -14236,7 +14236,7 @@ test_hyper_regular(void)
**
****************************************************************/
static void
-test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enblocks, hsize_t *eblock1,
+test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t endpoints, hssize_t enblocks, hsize_t *eblock1,
hsize_t *eblock2)
{
hid_t lim_sid;
@@ -14261,7 +14261,7 @@ test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enb
/* Check number of elements */
npoints = H5Sget_select_npoints(lim_sid);
CHECK(npoints, FAIL, "H5Sget_select_npoints");
- VERIFY(npoints, enpoints, "H5Sget_select_npoints");
+ VERIFY(npoints, endpoints, "H5Sget_select_npoints");
/* Get selection type */
sel_type = H5Sget_select_type(lim_sid);
diff --git a/test/tsohm.c b/test/tsohm.c
index b4ece0b..9678f46 100644
--- a/test/tsohm.c
+++ b/test/tsohm.c
@@ -636,7 +636,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_
if (GetTestExpress() > 1)
test_file_closing = FALSE;
- /* Intialize wdata */
+ /* Initialize wdata */
HDmemset(&wdata, 0, sizeof(wdata));
wdata.i1 = 11;
HDstrcpy(wdata.str, "string");
@@ -649,7 +649,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_
wdata.i8 = 88;
wdata.f1 = 0.0F;
- /* Intialize rdata */
+ /* Initialize rdata */
HDmemset(&rdata, 0, sizeof(rdata));
dtype1_id = make_dtype_1();
@@ -762,7 +762,7 @@ error:
* Function: getsize_testsize1
*
* Purpose: Creates a test file, populates it, and returns its file size.
- * Oject header information from the "first" dataset in the file
+ * Object header information from the "first" dataset in the file
* is stored in pointer `oinfo`.
*
* Programmer: Jacob Smith
@@ -1551,7 +1551,7 @@ size2_dump_struct(const char *name, size2_helper_struct *sizes)
/*-------------------------------------------------------------------------
* Function: size2_helper
*
- * Purpose: A helper functon for test_sohm_size2.
+ * Purpose: A helper function for test_sohm_size2.
*
* Creates a file using the given fcpl, then creates lots
* of different kinds of messages within the file and
@@ -1931,7 +1931,7 @@ size2_helper(hid_t fcpl_id, int test_file_closing, size2_helper_struct *ret_size
/*-------------------------------------------------------------------------
* Function: size2_verify
*
- * Purpose: A helper functon to verify the file created by size2_helper.
+ * Purpose: A helper function to verify the file created by size2_helper.
*
* Runs various tests (not exhaustive) to ensure that the
* file FILENAME actually has the structure that size2_helper
diff --git a/test/ttsafe_attr_vlen.c b/test/ttsafe_attr_vlen.c
index ecd0aa6..9fafaca 100644
--- a/test/ttsafe_attr_vlen.c
+++ b/test/ttsafe_attr_vlen.c
@@ -136,7 +136,7 @@ tts_attr_vlen_thread(void H5_ATTR_UNUSED *client_data)
gid = H5Gopen2(fid, "/", H5P_DEFAULT);
CHECK(gid, H5I_INVALID_HID, "H5Gopen");
- /* Open the attribte */
+ /* Open the attribute */
aid = H5Aopen(gid, "root_attr", H5P_DEFAULT);
CHECK(aid, H5I_INVALID_HID, "H5Aopen");
diff --git a/test/ttsafe_rec_rw_lock.c b/test/ttsafe_rec_rw_lock.c
index 0597578..a5033f7 100644
--- a/test/ttsafe_rec_rw_lock.c
+++ b/test/ttsafe_rec_rw_lock.c
@@ -136,7 +136,7 @@ void *tts_rw_lock_smoke_check_test_thread(void *_udata);
*
* 8) Drop the read lock a second time.
*
- * 9) Verify the exptected stats, and then reset them.
+ * 9) Verify the expected stats, and then reset them.
*
* 10) Obtain a write lock.
*
@@ -170,7 +170,7 @@ void *tts_rw_lock_smoke_check_test_thread(void *_udata);
*
* 25) Shut down the recursive R/W lock.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -242,7 +242,7 @@ tts_rec_rw_lock_smoke_check_1(void)
result = H5TS_rw_unlock(&rec_rw_lock);
CHECK_I(result, "H5TS_rw_unlock -- 3");
- /* 9) Verify the exptected stats, and then reset them. */
+ /* 9) Verify the expected stats, and then reset them. */
result = H5TS_rw_lock_get_stats(&rec_rw_lock, &stats);
CHECK_I(result, "H5TS_rw_lock_get_stats -- 2");
@@ -442,7 +442,7 @@ cleanup_rec_rw_lock_smoke_check_1(void)
* number of recursive lock and unlock calls is random, as is the
* order of read and write locks if both are enabled.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -593,7 +593,7 @@ tts_rw_lock_smoke_check_test_thread(void *_udata)
* repeats the process until the spcified number of read locks have
* been acquired and dropped.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -839,7 +839,7 @@ cleanup_rec_rw_lock_smoke_check_2(void)
* repeats the process until the spcified number of write locks have
* been acquired and dropped.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
@@ -1085,7 +1085,7 @@ cleanup_rec_rw_lock_smoke_check_3(void)
* repeats the process until the spcified number of locks have
* been acquired and dropped.
*
- * Creted Sept. 3. 2020.
+ * Created Sept. 3. 2020.
*
* Programmer: John Mainzer
*
diff --git a/test/tvltypes.c b/test/tvltypes.c
index 03a8ad3..0c4cf68 100644
--- a/test/tvltypes.c
+++ b/test/tvltypes.c
@@ -131,7 +131,7 @@ test_vltypes_dataset_create(void)
dcpl = H5Pcreate(H5P_DATASET_CREATE);
CHECK(dcpl, FAIL, "H5Pcreate");
- /* Set fill value writting time to be NEVER */
+ /* Set fill value writing time to be NEVER */
ret = H5Pset_fill_time(dcpl, H5D_FILL_TIME_NEVER);
CHECK(ret, FAIL, "H5Pset_fill_time");
diff --git a/test/twriteorder.c b/test/twriteorder.c
index 9cff00a..c58b030 100644
--- a/test/twriteorder.c
+++ b/test/twriteorder.c
@@ -54,7 +54,7 @@
*
* If the system, in which the writer and reader processes run, the readers
* will always get all chain-linked blocks correctly. If the order of write
- * is not maintained, some reader processes may found unexpect block data.
+ * is not maintained, some reader processes may found unexpected block data.
*
*************************************************************/
diff --git a/test/unlink.c b/test/unlink.c
index 94bd035..19b7b37 100644
--- a/test/unlink.c
+++ b/test/unlink.c
@@ -617,7 +617,7 @@ test_filespace(hid_t fapl)
/* Create common objects for datasets */
- /* Create dataset creation property list for contigous storage */
+ /* Create dataset creation property list for contiguous storage */
if ((contig_dcpl = H5Pcreate(H5P_DATASET_CREATE)) < 0)
TEST_ERROR
@@ -1099,7 +1099,7 @@ test_filespace(hid_t fapl)
PASSED();
- /* Create simple group hiearchy, remove it & verify file size */
+ /* Create simple group hierarchy, remove it & verify file size */
TESTING(" simple group hierarchy");
/* Create file */
@@ -1138,7 +1138,7 @@ test_filespace(hid_t fapl)
PASSED();
- /* Create complex group hiearchy, remove it & verify file size */
+ /* Create complex group hierarchy, remove it & verify file size */
TESTING(" complex group hierarchy");
/* Create file */
diff --git a/test/use_common.c b/test/use_common.c
index 8f78f3b..7b5df53 100644
--- a/test/use_common.c
+++ b/test/use_common.c
@@ -298,7 +298,7 @@ write_uc_file(hbool_t tosend, hid_t file_id, options_t *opts)
return -1;
}
- /* verify chunk_dims against set paramenters */
+ /* verify chunk_dims against set parameters */
if (chunk_dims[0] != opts->chunkdims[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) {
HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n",
(unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1],
@@ -533,7 +533,7 @@ read_uc_file(hbool_t towait, options_t *opts)
/* quit when all nplanes have been read */
loops_waiting_for_plane = 0;
while (nplanes_seen < opts->nplanes) {
- /* print progress message according to if new planes are availalbe */
+ /* print progress message according to if new planes are available */
if (nplanes_seen < dims[0]) {
if (loops_waiting_for_plane) {
/* end the previous message */
diff --git a/test/use_disable_mdc_flushes.c b/test/use_disable_mdc_flushes.c
index f28e574..d40e56d 100644
--- a/test/use_disable_mdc_flushes.c
+++ b/test/use_disable_mdc_flushes.c
@@ -360,7 +360,7 @@ write_file(void)
return -1;
}
- /* verify chunk_dims against set paramenters */
+ /* verify chunk_dims against set parameters */
if (chunk_dims[0] != chunkdims_g[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) {
HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n",
(unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1],
diff --git a/test/vds.c b/test/vds.c
index ac9bb5d..79424ea 100644
--- a/test/vds.c
+++ b/test/vds.c
@@ -196,7 +196,7 @@ vds_select_equal(hid_t space1, hid_t space2)
if (nblocks1 != nblocks2)
return FALSE;
- /* Allocate block lists. Do not return directly afer
+ /* Allocate block lists. Do not return directly after
* allocating, to make sure buffers are freed. */
if (NULL ==
(buf1 = (hsize_t *)HDmalloc((size_t)2 * (size_t)rank1 * (size_t)nblocks1 * sizeof(*buf1))))
@@ -479,7 +479,7 @@ test_api_get_ex_dcpl(test_api_config_t config, hid_t fapl, hid_t dcpl, hid_t *ex
/* Verify examination DCPL is equal to original DCPL. Do not compare the
* plist to itself, and do not do the comparison if we reopened the file,
- * because in that case the extent of the source dset will not be corrent.
+ * because in that case the extent of the source dset will not be current.
*/
if ((*ex_dcpl != dcpl) && (config != TEST_API_REOPEN_FILE)) {
if ((tri_ret = H5Pequal(dcpl, *ex_dcpl)) < 0)
@@ -1044,7 +1044,7 @@ test_api(test_api_config_t config, hid_t fapl, H5F_libver_t low)
if ((vspace[i] = H5Screate_simple(2, dims, NULL)) < 0)
TEST_ERROR
- /* Select row in virual dataspace */
+ /* Select row in virtual dataspace */
start[0] = (hsize_t)i;
if (H5Sselect_hyperslab(vspace[i], H5S_SELECT_SET, start, NULL, count, block) < 0)
TEST_ERROR
@@ -1182,7 +1182,7 @@ test_vds_prefix_first(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
int buf[10][26]; /* Write and expected read buffer */
@@ -1461,7 +1461,7 @@ test_basic_io(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
hsize_t start[4]; /* Hyperslab start */
@@ -4389,7 +4389,7 @@ test_unlim(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {10, 10}; /* Data space current size */
hsize_t mdims[2] = {10, 20}; /* Data space maximum size */
@@ -7403,7 +7403,7 @@ test_printf(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[2] = {-1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {10, 0}; /* Data space current size */
hsize_t mdims[2] = {10, 20}; /* Data space maximum size */
@@ -11111,7 +11111,7 @@ test_all(unsigned config, hid_t vds_fapl, hid_t src_fapl)
hid_t vspace[3] = {-1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
hid_t filespace = -1; /* File dataspace */
- hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[2] = {6, 6}; /* Data space current size */
hsize_t mdims[2] = {10, 10}; /* Data space maximum size */
diff --git a/test/vds_env.c b/test/vds_env.c
index 1f00444..db54ea0 100644
--- a/test/vds_env.c
+++ b/test/vds_env.c
@@ -58,7 +58,7 @@ test_vds_prefix_second(unsigned config, hid_t fapl)
hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */
hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */
hid_t memspace = -1; /* Memory dataspace */
- hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */
+ hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */
hid_t vdset = -1; /* Virtual dataset */
hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */
int buf[10][26]; /* Write and expected read buffer */
diff --git a/test/vfd.c b/test/vfd.c
index e90a589..b0f37c3 100644
--- a/test/vfd.c
+++ b/test/vfd.c
@@ -186,7 +186,7 @@ test_sec2(void)
if (os_file_handle == NULL)
FAIL_PUTS_ERROR("NULL os-specific vfd/file handle was returned from H5Fget_vfd_handle");
- /* There is no garantee the size of metadata in file is constant.
+ /* There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* Currently it should be around 2 KB.
@@ -382,7 +382,7 @@ test_core(void)
if (os_file_handle == NULL)
FAIL_PUTS_ERROR("NULL os-specific vfd/file handle was returned from H5Fget_vfd_handle");
- /* There is no garantee the size of metadata in file is constant.
+ /* There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* TODO: Needs justification of why is this is a reasonable size.
@@ -529,7 +529,7 @@ test_core(void)
} /* end if */
/* Check file size API.
- * There is no garantee the size of metadata in file is constant.
+ * There is no guarantee the size of metadata in file is constant.
* Just try to check if it's reasonable.
*
* TODO: Needs justification of why is this is a reasonable size.
@@ -1101,7 +1101,7 @@ error:
* See if we can open files created with v1.6 library.
* The source file was created by the test/file_handle.c
* of the v1.6 library. Then tools/misc/h5repart.c was
- * used to concantenated. The command was "h5repart -m 5k
+ * used to concatenated. The command was "h5repart -m 5k
* family_file%05d.h5 family_v16_%05d.h5".
*
* Return: Success: 0
@@ -1369,7 +1369,7 @@ H5_GCC_DIAG_ON("format-nonliteral")
/*-------------------------------------------------------------------------
* Function: test_multi
*
- * Purpose: Tests the file handle interface for MUTLI driver
+ * Purpose: Tests the file handle interface for MULTI driver
*
* Return: SUCCEED/FAIL
*
@@ -2524,7 +2524,7 @@ done:
* Function: driver_is_splitter_compatible
*
* Purpose: Determine whether the driver set in the FAPL ID is compatible
- * with the Splitter VFD -- specificially, Write-Only channel.
+ * with the Splitter VFD -- specifically, Write-Only channel.
*
* Return: Success: 0
* Failure: -1
@@ -3097,7 +3097,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3117,7 +3117,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3148,7 +3148,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
@@ -3179,7 +3179,7 @@ splitter_tentative_open_test(hid_t child_fapl_id)
SPLITTER_TEST_FAULT("can't close file ID\n");
}
if (!file_exists(filename_rw, child_fapl_id)) {
- SPLITTER_TEST_FAULT("R/W file mysteriously disappared\n");
+ SPLITTER_TEST_FAULT("R/W file mysteriously disappeared\n");
}
if (!file_exists(vfd_config->wo_path, child_fapl_id)) {
SPLITTER_TEST_FAULT("W/O file mysteriously disappeared\n");
diff --git a/testpar/t_2Gio.c b/testpar/t_2Gio.c
index b1c6e78..911be2c 100644
--- a/testpar/t_2Gio.c
+++ b/testpar/t_2Gio.c
@@ -33,7 +33,7 @@
#include "mpi.h"
/* For this test, we don't want to inherit the RANK definition
- * from testphdf5.h. We'll define MAX_RANK to accomodate 3D arrays
+ * from testphdf5.h. We'll define MAX_RANK to accommodate 3D arrays
* and use that definition rather than RANK.
*/
#ifndef MAX_RANK
@@ -3323,7 +3323,7 @@ none_selection_chunk(void)
* Simple independent I/O. This tests that the defaults are properly set.
*
* TEST_ACTUAL_IO_RESET:
- * Perfroms collective and then independent I/O wit hthe same dxpl to
+ * Performs collective and then independent I/O with hthe same dxpl to
* make sure the peroperty is correctly reset to the default on each use.
* Specifically, this test runs TEST_ACTUAL_IO_MULTI_CHUNK_NO_OPT_MIX_DISAGREE
* (The most complex case that works on all builds) and then performs
@@ -3332,7 +3332,7 @@ none_selection_chunk(void)
* Note: DIRECT_MULTI_CHUNK_MIX and DIRECT_MULTI_CHUNK_MIX_DISAGREE
* is not needed as they are covered by DIRECT_CHUNK_MIX and
* MULTI_CHUNK_MIX_DISAGREE cases. _DIRECT_ cases are only for testing
- * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO insted of num-threshold.
+ * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO instead of num-threshold.
*
* Modification:
* - Refctore to remove multi-chunk-without-opimization test and update for
@@ -3520,7 +3520,7 @@ test_actual_io_mode(int selection_mode)
break;
/* RESET tests that the properties are properly reset to defaults each time I/O is
- * performed. To acheive this, we have RESET perform collective I/O (which would change
+ * performed. To achieve this, we have RESET perform collective I/O (which would change
* the values from the defaults) followed by independent I/O (which should report the
* default values). RESET doesn't need to have a unique selection, so we reuse
* MULTI_CHUMK_MIX_DISAGREE, which was chosen because it is a complex case that works
@@ -3536,7 +3536,7 @@ test_actual_io_mode(int selection_mode)
* assgigned collective I/O, while each other process gets independent I/O.
* Since the root process with only access the first chunk, it will report
* collective I/O. The subsequent processes will access the first chunk
- * collectively, and their other chunk indpendently, reporting mixed I/O.
+ * collectively, and their other chunk independently, reporting mixed I/O.
*/
if (mpi_rank == 0) {
@@ -3651,7 +3651,7 @@ test_actual_io_mode(int selection_mode)
ret = H5Pset_dxpl_mpio_chunk_opt_num(dxpl_write, (unsigned)mpi_size * 2);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_num succeeded");
- /* set this to manipulate testing senario about allocating processes
+ /* set this to manipulate testing scenario about allocating processes
* to chunks */
ret = H5Pset_dxpl_mpio_chunk_opt_ratio(dxpl_write, (unsigned)99);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_ratio succeeded");
@@ -3675,12 +3675,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dwrite() dataset multichunk write succeeded");
- /* Retreive Actual io valuess */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode suceeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Read */
ret = H5Dread(dataset, data_type, mem_space, file_space, dxpl_read, buffer);
@@ -3688,12 +3688,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dread() dataset multichunk read succeeded");
- /* Retreive Actual io values */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Check write vs read */
VRFY((actual_io_mode_read == actual_io_mode_write),
@@ -3714,7 +3714,7 @@ test_actual_io_mode(int selection_mode)
actual_io_mode_write);
}
- /* To test that the property is succesfully reset to the default, we perform some
+ /* To test that the property is successfully reset to the default, we perform some
* independent I/O after the collective I/O
*/
if (selection_mode == TEST_ACTUAL_IO_RESET) {
@@ -3731,9 +3731,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_write == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset write (independent)");
@@ -3746,9 +3746,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_read == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset read (independent)");
@@ -3794,11 +3794,11 @@ actual_io_mode_tests(void)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_IND);
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_COL);
- /* The Multi Chunk Mixed test requires atleast three processes. */
+ /* The Multi Chunk Mixed test requires at least three processes. */
if (mpi_size > 2)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX);
else
- HDfprintf(stdout, "Multi Chunk Mixed test requires 3 proceses minimum\n");
+ HDfprintf(stdout, "Multi Chunk Mixed test requires 3 processes minimum\n");
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX_DISAGREE);
@@ -3842,7 +3842,7 @@ actual_io_mode_tests(void)
* Test for Data Type Conversion as the cause of breaking collective I/O.
*
* TEST_DATA_TRANSFORMS:
- * Test for Data Transfrom feature as the cause of breaking collective I/O.
+ * Test for Data Transform feature as the cause of breaking collective I/O.
*
* TEST_NOT_SIMPLE_OR_SCALAR_DATASPACES:
* Test for NULL dataspace as the cause of breaking collective I/O.
@@ -3980,7 +3980,7 @@ test_no_collective_cause_mode(int selection_mode)
}
if (selection_mode & TEST_DATA_TRANSFORMS) {
- test_name = "Broken Collective I/O - DATA Transfroms";
+ test_name = "Broken Collective I/O - DATA Transforms";
no_collective_cause_local_expected |= H5D_MPIO_DATA_TRANSFORMS;
no_collective_cause_global_expected |= H5D_MPIO_DATA_TRANSFORMS;
}
@@ -4070,7 +4070,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_write, &no_collective_cause_local_write,
&no_collective_cause_global_write);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/*---------------------
* Test Read access
@@ -4090,7 +4090,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_read, &no_collective_cause_local_read,
&no_collective_cause_global_read);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/* Check write vs read */
VRFY((no_collective_cause_local_read == no_collective_cause_local_write),
@@ -4274,7 +4274,7 @@ dataset_atomicity(void)
/* file locking allows only one file open (serial) for writing */
if (MAINPROCESS) {
fid = H5Fopen(filename, H5F_ACC_RDWR, H5P_DEFAULT);
- VRFY((fid >= 0), "H5Fopen succeeed");
+ VRFY((fid >= 0), "H5Fopen succeeded");
}
/* should fail */
diff --git a/testpar/t_bigio.c b/testpar/t_bigio.c
index ed99fc4..c21d233 100644
--- a/testpar/t_bigio.c
+++ b/testpar/t_bigio.c
@@ -1331,7 +1331,7 @@ coll_chunk2(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = space_dim1*mpi_size
@@ -1840,7 +1840,7 @@ main(int argc, char **argv)
/* Set the bigio processing limit to be 'newsize' bytes */
hsize_t oldsize = H5_mpi_set_bigio_count(newsize);
- /* Having set the bigio handling to a size that is managable,
+ /* Having set the bigio handling to a size that is manageable,
* we'll set our 'bigcount' variable to be 2X that limit so
* that we try to ensure that our bigio handling is actually
* envoked and tested.
diff --git a/testpar/t_cache.c b/testpar/t_cache.c
index a5c8039..3ce5729 100644
--- a/testpar/t_cache.c
+++ b/testpar/t_cache.c
@@ -126,7 +126,7 @@ int total_writes = 0;
*
* local_pinned: Boolean flag that is set to true iff the entry
* has been pinned in the local cache, but probably not all
- * caches. Such pins will typically not be consistant across
+ * caches. Such pins will typically not be consistent across
* processes, and thus cannot be marked as dirty unless they
* happen to overlap some collective operation.
*
@@ -205,7 +205,7 @@ struct datum data[NUM_DATA_ENTRIES];
* even divisor of NUM_DATA_ENTRIES. So far, all tests have been with
* powers of 10 that meet these criteria.
*
- * Further, this value must be consistant across all processes.
+ * Further, this value must be consistent across all processes.
*/
#define STD_VIRT_NUM_DATA_ENTRIES NUM_DATA_ENTRIES
@@ -252,7 +252,7 @@ int data_index[NUM_DATA_ENTRIES];
*
* Strangely, at least on Phoenix, the first solution runs faster by a
* rather large margin. However, I can imagine this changing with
- * different OS's and MPI implementatins.
+ * different OS's and MPI implementations.
*
* Thus I have left code supporting the second solution in place.
*
@@ -1280,7 +1280,7 @@ reset_server_counters(void)
* Function: server_main()
*
* Purpose: Main function for the server process. This process exists
- * to provide an independant view of the data array.
+ * to provide an independent view of the data array.
*
* The function handles request from the other processes in
* the test until the count of done messages received equals
@@ -3956,7 +3956,7 @@ setup_cache_for_test(hid_t *fid_ptr, H5F_t **file_ptr_ptr, H5C_t **cache_ptr_ptr
* Purpose: Verify that the indicated entries have been written exactly
* once each, and that the indicated total number of writes
* has been processed by the server process. Flag an error if
- * discrepency is noted. Finally reset the counters maintained
+ * discrepancy is noted. Finally reset the counters maintained
* by the server process.
*
* This function should only be called by the metadata cache
@@ -4044,7 +4044,7 @@ verify_writes(unsigned num_writes, haddr_t *written_entries_tbl)
/* final barrier to ensure that all processes think that the server
* counters have been reset before we leave the sync point. This
- * barrier is probaby not necessary at this point in time (5/9/10),
+ * barrier is probably not necessary at this point in time (5/9/10),
* but I can think of at least one likely change to the metadata write
* strategies that will require it -- hence its insertion now.
*/
@@ -4788,7 +4788,7 @@ server_smoke_check(void)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -5135,7 +5135,7 @@ smoke_check_1(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -5299,7 +5299,7 @@ smoke_check_2(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -5447,7 +5447,7 @@ smoke_check_2(int metadata_write_strategy)
*
* Purpose: Third smoke check for the parallel cache.
*
- * Use random reads to vary the loads on the diffferent
+ * Use random reads to vary the loads on the different
* processors. Also force different cache size adjustments.
*
* In this test, load process 0 heavily, and the other
@@ -5504,7 +5504,7 @@ smoke_check_3(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -5732,7 +5732,7 @@ smoke_check_3(int metadata_write_strategy)
*
* Purpose: Fourth smoke check for the parallel cache.
*
- * Use random reads to vary the loads on the diffferent
+ * Use random reads to vary the loads on the different
* processors. Also force different cache size adjustments.
*
* In this test, load process 0 lightly, and the other
@@ -5789,7 +5789,7 @@ smoke_check_4(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -6064,7 +6064,7 @@ smoke_check_5(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -6349,7 +6349,7 @@ trace_file_check(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose)
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
@@ -6656,7 +6656,7 @@ smoke_check_6(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC);
diff --git a/testpar/t_cache_image.c b/testpar/t_cache_image.c
index 642cace..4522572 100644
--- a/testpar/t_cache_image.c
+++ b/testpar/t_cache_image.c
@@ -452,7 +452,7 @@ create_data_sets(hid_t file_id, int min_dset, int max_dset)
}
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 10X10 element chunks.
+ * to be partitioned into 10X10 element chunks.
*/
if (pass) {
@@ -798,7 +798,7 @@ delete_data_sets(hid_t file_id, int min_dset, int max_dset)
* Set pass to FALSE and issue a suitable failure
* message if either the file contains a metadata cache image
* superblock extension and mdci_sbem_expected is TRUE, or
- * vise versa.
+ * vice versa.
*
* If mdci_sbem_expected is TRUE, also verify that the metadata
* cache has been advised of this.
@@ -810,7 +810,7 @@ delete_data_sets(hid_t file_id, int min_dset, int max_dset)
* FAPL entry when opening the file, and verify that the
* metadata cache is notified.
*
- * If config_fsm is TRUE, setup the persistant free space
+ * If config_fsm is TRUE, setup the persistent free space
* manager. Note that this flag may only be set if
* create_file is also TRUE.
*
@@ -887,7 +887,7 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons
if (show_progress)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
- /* create a file access propertly list. */
+ /* create a file access property list. */
if (pass) {
fapl_id = H5Pcreate(H5P_FILE_ACCESS);
@@ -959,7 +959,7 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons
if (show_progress)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
- /* setup the persistant free space manager if indicated */
+ /* setup the persistent free space manager if indicated */
if ((pass) && (config_fsm)) {
fcpl_id = H5Pcreate(H5P_FILE_CREATE);
@@ -1141,13 +1141,13 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons
if ((file_ptr->shared->page_buf) && ((!enable_page_buffer) || (l_facc_type == FACC_MPIO))) {
pass = FALSE;
- failure_mssg = "page buffer unexepectedly enabled.";
+ failure_mssg = "page buffer unexpectedly enabled.";
}
else if ((file_ptr->shared->page_buf != NULL) &&
((enable_page_buffer) || (l_facc_type != FACC_MPIO))) {
pass = FALSE;
- failure_mssg = "page buffer unexepectedly disabled.";
+ failure_mssg = "page buffer unexpectedly disabled.";
}
}
@@ -1363,7 +1363,7 @@ par_create_dataset(int dset_num, hid_t file_id, int mpi_rank, int mpi_size)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
/* set the dataset creation plist to specify that the raw data is
- * to be partioned into 1X10X10 element chunks.
+ * to be partitioned into 1X10X10 element chunks.
*/
if (pass) {
@@ -3016,7 +3016,7 @@ verify_cache_image_RO(int file_name_id, int md_write_strat, int mpi_rank)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3287,7 +3287,7 @@ verify_cache_image_RW(int file_name_id, int md_write_strat, int mpi_rank)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3613,7 +3613,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3708,7 +3708,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size)
/* Verify that all other processes receive the cache image block
* from process 0.
*
- * Since we have alread verified that only process 0 has read the
+ * Since we have already verified that only process 0 has read the
* image, it is sufficient to verify that the image was loaded on
* all processes.
*/
@@ -3753,9 +3753,9 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size)
HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass);
/* 13) Get the size of the file. Verify that it is less
- * than 20 KB. Without deletions and persistant free
+ * than 20 KB. Without deletions and persistent free
* space managers, size size is about 30 MB, so this
- * is sufficient to verify that the persistant free
+ * is sufficient to verify that the persistent free
* space managers are more or less doing their job.
*
* Note that this test will have to change if we use
diff --git a/testpar/t_coll_chunk.c b/testpar/t_coll_chunk.c
index 651a392..104460a 100644
--- a/testpar/t_coll_chunk.c
+++ b/testpar/t_coll_chunk.c
@@ -155,7 +155,7 @@ coll_chunk2(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -209,7 +209,7 @@ coll_chunk3(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -261,7 +261,7 @@ coll_chunk4(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -315,7 +315,7 @@ coll_chunk5(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -367,7 +367,7 @@ coll_chunk6(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -419,7 +419,7 @@ coll_chunk7(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -471,7 +471,7 @@ coll_chunk8(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
@@ -523,7 +523,7 @@ coll_chunk9(void)
*/
/* ------------------------------------------------------------------------
- * Descriptions for the selection: one singular selection accross many chunks
+ * Descriptions for the selection: one singular selection across many chunks
* Two dimensions, Num of chunks = 2* mpi_size
*
* dim1 = SPACE_DIM1*mpi_size
diff --git a/testpar/t_dset.c b/testpar/t_dset.c
index d6be619..8616bef 100644
--- a/testpar/t_dset.c
+++ b/testpar/t_dset.c
@@ -2881,7 +2881,7 @@ none_selection_chunk(void)
* Simple independent I/O. This tests that the defaults are properly set.
*
* TEST_ACTUAL_IO_RESET:
- * Perfroms collective and then independent I/O wit hthe same dxpl to
+ * Performs collective and then independent I/O with hthe same dxpl to
* make sure the peroperty is correctly reset to the default on each use.
* Specifically, this test runs TEST_ACTUAL_IO_MULTI_CHUNK_NO_OPT_MIX_DISAGREE
* (The most complex case that works on all builds) and then performs
@@ -2890,7 +2890,7 @@ none_selection_chunk(void)
* Note: DIRECT_MULTI_CHUNK_MIX and DIRECT_MULTI_CHUNK_MIX_DISAGREE
* is not needed as they are covered by DIRECT_CHUNK_MIX and
* MULTI_CHUNK_MIX_DISAGREE cases. _DIRECT_ cases are only for testing
- * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO insted of num-threshold.
+ * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO instead of num-threshold.
*
* Modification:
* - Refctore to remove multi-chunk-without-opimization test and update for
@@ -3078,7 +3078,7 @@ test_actual_io_mode(int selection_mode)
break;
/* RESET tests that the properties are properly reset to defaults each time I/O is
- * performed. To acheive this, we have RESET perform collective I/O (which would change
+ * performed. To achieve this, we have RESET perform collective I/O (which would change
* the values from the defaults) followed by independent I/O (which should report the
* default values). RESET doesn't need to have a unique selection, so we reuse
* MULTI_CHUMK_MIX_DISAGREE, which was chosen because it is a complex case that works
@@ -3094,7 +3094,7 @@ test_actual_io_mode(int selection_mode)
* assgigned collective I/O, while each other process gets independent I/O.
* Since the root process with only access the first chunk, it will report
* collective I/O. The subsequent processes will access the first chunk
- * collectively, and their other chunk indpendently, reporting mixed I/O.
+ * collectively, and their other chunk independently, reporting mixed I/O.
*/
if (mpi_rank == 0) {
@@ -3209,7 +3209,7 @@ test_actual_io_mode(int selection_mode)
ret = H5Pset_dxpl_mpio_chunk_opt_num(dxpl_write, (unsigned)mpi_size * 2);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_num succeeded");
- /* set this to manipulate testing senario about allocating processes
+ /* set this to manipulate testing scenario about allocating processes
* to chunks */
ret = H5Pset_dxpl_mpio_chunk_opt_ratio(dxpl_write, (unsigned)99);
VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_ratio succeeded");
@@ -3233,12 +3233,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dwrite() dataset multichunk write succeeded");
- /* Retreive Actual io valuess */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode suceeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Read */
ret = H5Dread(dataset, data_type, mem_space, file_space, dxpl_read, buffer);
@@ -3246,12 +3246,12 @@ test_actual_io_mode(int selection_mode)
H5Eprint2(H5E_DEFAULT, stdout);
VRFY((ret >= 0), "H5Dread() dataset multichunk read succeeded");
- /* Retreive Actual io values */
+ /* Retrieve Actual io values */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
/* Check write vs read */
VRFY((actual_io_mode_read == actual_io_mode_write),
@@ -3272,7 +3272,7 @@ test_actual_io_mode(int selection_mode)
actual_io_mode_write);
}
- /* To test that the property is succesfully reset to the default, we perform some
+ /* To test that the property is successfully reset to the default, we perform some
* independent I/O after the collective I/O
*/
if (selection_mode == TEST_ACTUAL_IO_RESET) {
@@ -3289,9 +3289,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_write == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset write (independent)");
@@ -3304,9 +3304,9 @@ test_actual_io_mode(int selection_mode)
/* Check Properties */
ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read);
- VRFY((ret >= 0), "retriving actual io mode succeeded");
+ VRFY((ret >= 0), "retrieving actual io mode succeeded");
ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read);
- VRFY((ret >= 0), "retriving actual chunk opt mode succeeded");
+ VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded");
VRFY(actual_chunk_opt_mode_read == H5D_MPIO_NO_CHUNK_OPTIMIZATION,
"actual_chunk_opt_mode has correct value for reset read (independent)");
@@ -3359,11 +3359,11 @@ actual_io_mode_tests(void)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_IND);
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_COL);
- /* The Multi Chunk Mixed test requires atleast three processes. */
+ /* The Multi Chunk Mixed test requires at least three processes. */
if (mpi_size > 2)
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX);
else
- HDfprintf(stdout, "Multi Chunk Mixed test requires 3 proceses minimum\n");
+ HDfprintf(stdout, "Multi Chunk Mixed test requires 3 processes minimum\n");
test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX_DISAGREE);
@@ -3407,7 +3407,7 @@ actual_io_mode_tests(void)
* Test for Data Type Conversion as the cause of breaking collective I/O.
*
* TEST_DATA_TRANSFORMS:
- * Test for Data Transfrom feature as the cause of breaking collective I/O.
+ * Test for Data Transform feature as the cause of breaking collective I/O.
*
* TEST_NOT_SIMPLE_OR_SCALAR_DATASPACES:
* Test for NULL dataspace as the cause of breaking collective I/O.
@@ -3547,7 +3547,7 @@ test_no_collective_cause_mode(int selection_mode)
}
if (selection_mode & TEST_DATA_TRANSFORMS) {
- test_name = "Broken Collective I/O - DATA Transfroms";
+ test_name = "Broken Collective I/O - DATA Transforms";
no_collective_cause_local_expected |= H5D_MPIO_DATA_TRANSFORMS;
no_collective_cause_global_expected |= H5D_MPIO_DATA_TRANSFORMS;
}
@@ -3637,7 +3637,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_write, &no_collective_cause_local_write,
&no_collective_cause_global_write);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/*---------------------
* Test Read access
@@ -3657,7 +3657,7 @@ test_no_collective_cause_mode(int selection_mode)
/* Get the cause of broken collective I/O */
ret = H5Pget_mpio_no_collective_cause(dxpl_read, &no_collective_cause_local_read,
&no_collective_cause_global_read);
- VRFY((ret >= 0), "retriving no collective cause succeeded");
+ VRFY((ret >= 0), "retrieving no collective cause succeeded");
/* Check write vs read */
VRFY((no_collective_cause_local_read == no_collective_cause_local_write),
@@ -3843,7 +3843,7 @@ dataset_atomicity(void)
/* file locking allows only one file open (serial) for writing */
if (MAINPROCESS) {
fid = H5Fopen(filename, H5F_ACC_RDWR, H5P_DEFAULT);
- VRFY((fid >= 0), "H5Fopen succeeed");
+ VRFY((fid >= 0), "H5Fopen succeeded");
/* should fail */
H5E_BEGIN_TRY
diff --git a/testpar/t_file.c b/testpar/t_file.c
index 52cae6d..557a699 100644
--- a/testpar/t_file.c
+++ b/testpar/t_file.c
@@ -179,7 +179,7 @@ test_page_buffer_access(void)
data = (int *)HDmalloc(sizeof(int) * (size_t)num_elements);
- /* intialize all the elements to have a value of -1 */
+ /* initialize all the elements to have a value of -1 */
for (i = 0; i < num_elements; i++)
data[i] = -1;
if (MAINPROCESS) {
diff --git a/testpar/t_mdset.c b/testpar/t_mdset.c
index 3041e77..97d5966 100644
--- a/testpar/t_mdset.c
+++ b/testpar/t_mdset.c
@@ -1721,7 +1721,7 @@ io_mode_confusion(void)
* Open the each of the data sets in turn. If all opens are successful,
* the test passes. Otherwise the test fails.
*
- * Note that this test will probably become irrelevent shortly, when we
+ * Note that this test will probably become irrelevant shortly, when we
* land the journaling modifications on the trunk -- at which point all
* cache clients will have to construct on disk images on demand.
*
@@ -1782,7 +1782,7 @@ rr_obj_hdr_flush_confusion(void)
mrc = MPI_Comm_split(MPI_COMM_WORLD, is_reader, mpi_rank, &comm);
VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split");
- /* The reader proocesses branches off to do reading
+ /* The reader processes branches off to do reading
* while the writer processes continues to do writing
* Whenever writers finish one writing step, including a H5Fflush,
* they inform the readers, via MPI_COMM_WORLD, to verify.
@@ -2063,7 +2063,7 @@ rr_obj_hdr_flush_confusion_writer(MPI_Comm comm)
/*
* flush the metadata cache yet again to clean the object headers.
*
- * This is an attempt to crate a situation where we have dirty
+ * This is an attempt to create a situation where we have dirty
* object header continuation chunks, but clean opject headers
* to verify a speculative bug fix -- it doesn't seem to work,
* but I will leave the code in anyway, as the object header
diff --git a/testpar/t_mpi.c b/testpar/t_mpi.c
index a883f55..bb56f35 100644
--- a/testpar/t_mpi.c
+++ b/testpar/t_mpi.c
@@ -215,7 +215,7 @@ test_mpio_gb_file(char *filename)
* sizes.
*/
if (MAINPROCESS) { /* only process 0 needs to check it*/
- HDprintf("MPI_Offset is %s %d bytes integeral type\n", is_signed ? "signed" : "unsigned",
+ HDprintf("MPI_Offset is %s %d bytes integral type\n", is_signed ? "signed" : "unsigned",
(int)sizeof(MPI_Offset));
if (sizeof_mpi_offset <= 4 && is_signed) {
HDprintf("Skipped 2GB range test "
@@ -400,7 +400,7 @@ finish:
* MPI-IO Test: One writes, Many reads.
* Verify if only one process writes some data and then all other
* processes can read them back correctly. This tests if the
- * underlaying parallel I/O and file system supports parallel I/O
+ * underlying parallel I/O and file system supports parallel I/O
* correctly.
*
* Algorithm: Only one process (e.g., process 0) writes some data.
diff --git a/testpar/t_ph5basic.c b/testpar/t_ph5basic.c
index 05f2c77..59f14cb 100644
--- a/testpar/t_ph5basic.c
+++ b/testpar/t_ph5basic.c
@@ -56,7 +56,7 @@ test_fapl_mpio_dup(void)
HDprintf("rank/size of MPI_COMM_WORLD are %d/%d\n", mpi_rank, mpi_size);
/* Create a new communicator that has the same processes as MPI_COMM_WORLD.
- * Use MPI_Comm_split because it is simplier than MPI_Comm_create
+ * Use MPI_Comm_split because it is simpler than MPI_Comm_create
*/
mrc = MPI_Comm_split(MPI_COMM_WORLD, 0, 0, &comm);
VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split");
diff --git a/testpar/t_pread.c b/testpar/t_pread.c
index cf91b31..97f6d99 100644
--- a/testpar/t_pread.c
+++ b/testpar/t_pread.c
@@ -68,7 +68,7 @@ static char *test_argv0 = NULL;
* more or less in half. Each sub group will operate
* collectively on their assigned file. This split into
* subgroups validates that parallel groups can successfully
- * open and read data independantly from the other parallel
+ * open and read data independently from the other parallel
* operations taking place.
*
* Return: Success: 0
@@ -132,7 +132,7 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id)
}
/* Setup the file names
- * The test specfic filenames are stored as consecutive
+ * The test specific filenames are stored as consecutive
* array entries in the global 'FILENAMES' array above.
* Here, we simply decide on the starting index for
* file construction. The reading portion of the test
@@ -1043,11 +1043,11 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id)
* Purpose: To implement a parallel test which validates whether the
* new superblock lookup functionality is working correctly.
*
- * The test consists of creating two seperate HDF datasets
+ * The test consists of creating two separate HDF datasets
* in which random text is inserted at the start of each
* file using the 'j5jam' application. This forces the
* HDF5 file superblock to a non-zero offset.
- * Having created the two independant files, we create two
+ * Having created the two independent files, we create two
* non-overlapping MPI groups, each of which is then tasked
* with the opening and validation of the data contained
* therein.
@@ -1152,7 +1152,7 @@ main(int argc, char **argv)
goto finish;
}
- /* Now read the generated test file (stil using MPI_COMM_WORLD) */
+ /* Now read the generated test file (still using MPI_COMM_WORLD) */
nerrs += test_parallel_read(MPI_COMM_WORLD, mpi_rank, mpi_size, which_group);
if (nerrs > 0) {
diff --git a/testpar/t_shapesame.c b/testpar/t_shapesame.c
index 2dd867a..0a18781 100644
--- a/testpar/t_shapesame.c
+++ b/testpar/t_shapesame.c
@@ -11,7 +11,7 @@
* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */
/*
- This program will test independant and collective reads and writes between
+ This program will test independent and collective reads and writes between
selections of different rank that non-the-less are deemed as having the
same shape by H5Sselect_shape_same().
*/
@@ -415,11 +415,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
/* setup selections for writing initial data to the small data set */
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
if (MAINPROCESS) { /* add an additional slice to the selections */
@@ -427,11 +427,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, or) succeeded");
}
/* write the initial value of the small data set to file */
@@ -475,21 +475,21 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) succeeded");
/* In passing, setup the process slice dataspaces as well */
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_process_slice_sid, H5S_SELECT_SET, tv_ptr->start,
tv_ptr->stride, tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_process_slice_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_process_slice_sid, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_process_slice_sid, H5S_SELECT_SET, tv_ptr->start,
tv_ptr->stride, tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_process_slice_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_process_slice_sid, set) succeeded");
if (MAINPROCESS) { /* add an additional slice to the selections */
@@ -497,11 +497,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, or) succeeded");
}
/* write the initial value of the large data set to file */
@@ -933,7 +933,7 @@ contig_hs_dr_pio_test__d2m_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
#if CONTIG_HS_DR_PIO_TEST__D2M_S2L__DEBUG
HDfprintf(stdout, "%s reading slices of on disk small data set into slices of big data set.\n", fcnName);
@@ -1185,11 +1185,11 @@ contig_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
/* set up start, stride, count, and block -- note that we will
* change start[] so as to read slices of the large cube.
@@ -1462,7 +1462,7 @@ contig_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
/* set up start, stride, count, and block -- note that we will
* change start[] so as to write slices of the small data set to
@@ -1572,7 +1572,7 @@ contig_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Dwrite(tv_ptr->large_dataset, H5T_NATIVE_UINT32, tv_ptr->large_ds_slice_sid,
tv_ptr->file_large_ds_process_slice_sid, tv_ptr->xfer_plist,
tv_ptr->large_ds_buf_2);
- VRFY((ret != FAIL), "H5Dwrite() to zero large ds suceeded");
+ VRFY((ret != FAIL), "H5Dwrite() to zero large ds succeeded");
/* select the portion of the in memory large cube to which we
* are going to write data.
@@ -2008,8 +2008,8 @@ contig_hs_dr_pio_test(ShapeSameTestMethods sstest_type)
** edge_size, and a checker_edge_size, select a checker
** board selection of a sel_rank (sel_rank < tgt_rank)
** dimensional slice through the dataspace parallel to the
-** sel_rank fastest changing indicies, with origin (in the
-** higher indicies) as indicated by the start array.
+** sel_rank fastest changing indices, with origin (in the
+** higher indices) as indicated by the start array.
**
** Note that this function, like all its relatives, is
** hard coded to presume a maximum dataspace rank of 5.
@@ -2318,7 +2318,7 @@ ckrbrd_hs_dr_pio_test__slct_ckrbrd(const int mpi_rank, const hid_t tgt_sid, cons
** or writing a checker board selection of an m (1 <= m <
** rank) dimensional slice through this processes slice
** of the target data set. Also, this slice must be parallel
-** to the fastest changing indicies.
+** to the fastest changing indices.
**
** It is further presumed that the buffer was zeroed before
** the read/write, and that the full target data set (i.e.
@@ -2353,7 +2353,7 @@ ckrbrd_hs_dr_pio_test__slct_ckrbrd(const int mpi_rank, const hid_t tgt_sid, cons
** of the buffer resides either at the origin of either
** a selected or an unselected checker. (Translation:
** if partial checkers appear in the buffer, they will
-** intersect the edges of the n-cube oposite the origin.)
+** intersect the edges of the n-cube opposite the origin.)
**
****************************************************************/
@@ -3091,11 +3091,11 @@ ckrbrd_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
sel_start[0] = sel_start[1] = sel_start[2] = sel_start[3] = sel_start[4] = 0;
sel_start[tv_ptr->small_ds_offset] = (hsize_t)(tv_ptr->mpi_rank);
@@ -3384,11 +3384,11 @@ ckrbrd_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) succeeded");
ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride,
tv_ptr->count, tv_ptr->block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, set) succeeded");
/* setup a checkerboard selection of the slice of the in memory small
* data set associated with the process's mpi rank.
@@ -3496,7 +3496,7 @@ ckrbrd_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr)
*/
ret = H5Dwrite(tv_ptr->large_dataset, H5T_NATIVE_UINT32, tv_ptr->mem_large_ds_sid,
tv_ptr->file_large_ds_sid_0, tv_ptr->xfer_plist, tv_ptr->large_ds_buf_2);
- VRFY((ret != FAIL), "H5Dwrite() to zero large ds suceeded");
+ VRFY((ret != FAIL), "H5Dwrite() to zero large ds succeeded");
/* select the portion of the in memory large cube to which we
* are going to write data.
@@ -4210,42 +4210,42 @@ create_faccess_plist(MPI_Comm comm, MPI_Info info, int l_facc_type)
return (ret_pl);
}
-/* Shape Same test using contigous hyperslab using independent IO on contigous datasets */
+/* Shape Same test using contiguous hyperslab using independent IO on contiguous datasets */
static void
sscontig1(void)
{
contig_hs_dr_pio_test(IND_CONTIG);
}
-/* Shape Same test using contigous hyperslab using collective IO on contigous datasets */
+/* Shape Same test using contiguous hyperslab using collective IO on contiguous datasets */
static void
sscontig2(void)
{
contig_hs_dr_pio_test(COL_CONTIG);
}
-/* Shape Same test using contigous hyperslab using independent IO on chunked datasets */
+/* Shape Same test using contiguous hyperslab using independent IO on chunked datasets */
static void
sscontig3(void)
{
contig_hs_dr_pio_test(IND_CHUNKED);
}
-/* Shape Same test using contigous hyperslab using collective IO on chunked datasets */
+/* Shape Same test using contiguous hyperslab using collective IO on chunked datasets */
static void
sscontig4(void)
{
contig_hs_dr_pio_test(COL_CHUNKED);
}
-/* Shape Same test using checker hyperslab using independent IO on contigous datasets */
+/* Shape Same test using checker hyperslab using independent IO on contiguous datasets */
static void
sschecker1(void)
{
ckrbrd_hs_dr_pio_test(IND_CONTIG);
}
-/* Shape Same test using checker hyperslab using collective IO on contigous datasets */
+/* Shape Same test using checker hyperslab using collective IO on contiguous datasets */
static void
sschecker2(void)
{
@@ -4305,7 +4305,7 @@ main(int argc, char **argv)
/* Initialize testing framework */
TestInit(argv[0], usage, parse_options);
- /* Shape Same tests using contigous hyperslab */
+ /* Shape Same tests using contiguous hyperslab */
AddTest("sscontig1", sscontig1, NULL, "Cntg hslab, ind IO, cntg dsets", PARATESTFILE);
AddTest("sscontig2", sscontig2, NULL, "Cntg hslab, col IO, cntg dsets", PARATESTFILE);
AddTest("sscontig3", sscontig3, NULL, "Cntg hslab, ind IO, chnk dsets", PARATESTFILE);
diff --git a/testpar/t_span_tree.c b/testpar/t_span_tree.c
index 7c2763c..2db8a38 100644
--- a/testpar/t_span_tree.c
+++ b/testpar/t_span_tree.c
@@ -422,7 +422,7 @@ coll_write_test(int chunk_factor)
the correctedness of collective write compared with
independent write,
- In order to throughly test this feature, we choose
+ In order to thoroughly test this feature, we choose
a different selection set for reading the data out.
@@ -888,8 +888,8 @@ coll_read_test(void)
** edge_size, and a checker_edge_size, select a checker
** board selection of a sel_rank (sel_rank < tgt_rank)
** dimensional slice through the dataspace parallel to the
-** sel_rank fastest changing indicies, with origin (in the
-** higher indicies) as indicated by the start array.
+** sel_rank fastest changing indices, with origin (in the
+** higher indices) as indicated by the start array.
**
** Note that this function, is hard coded to presume a
** maximum dataspace rank of 5.
@@ -1228,7 +1228,7 @@ lower_dim_size_comp_test__select_checker_board(const int mpi_rank, const hid_t t
** or writing a checker board selection of an m (1 <= m <
** rank) dimensional slice through this processes slice
** of the target data set. Also, this slice must be parallel
-** to the fastest changing indicies.
+** to the fastest changing indices.
**
** It is further presumed that the buffer was zeroed before
** the read/write, and that the full target data set (i.e.
@@ -1263,7 +1263,7 @@ lower_dim_size_comp_test__select_checker_board(const int mpi_rank, const hid_t t
** of the buffer resides either at the origin of either
** a selected or an unselected checker. (Translation:
** if partial checkers appear in the buffer, they will
-** intersect the edges of the n-cube oposite the origin.)
+** intersect the edges of the n-cube opposite the origin.)
**
****************************************************************/
@@ -1725,7 +1725,7 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
if (!use_collective_io) {
ret = H5Pset_dxpl_mpio_collective_opt(xfer_plist, H5FD_MPIO_INDIVIDUAL_IO);
- VRFY((ret >= 0), "H5Pset_dxpl_mpio_collective_opt() suceeded");
+ VRFY((ret >= 0), "H5Pset_dxpl_mpio_collective_opt() succeeded");
}
/* setup selection to write initial data to the small data sets */
@@ -1757,10 +1757,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
/* setup selections for writing initial data to the small data set */
ret = H5Sselect_hyperslab(mem_small_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(file_small_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, set) succeeded");
if (MAINPROCESS) { /* add an additional slice to the selections */
@@ -1781,10 +1781,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
#endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */
ret = H5Sselect_hyperslab(mem_small_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(file_small_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, or) succeeded");
}
check = H5Sselect_valid(mem_small_ds_sid);
@@ -1863,10 +1863,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
#endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */
ret = H5Sselect_hyperslab(mem_large_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) succeeded");
ret = H5Sselect_hyperslab(file_large_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, set) succeeded");
#if LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG
if (mpi_rank == LOWER_DIM_SIZE_COMP_TEST_DEBUG_TARGET_RANK) {
@@ -1896,10 +1896,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_
#endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */
ret = H5Sselect_hyperslab(mem_large_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) succeeded");
ret = H5Sselect_hyperslab(file_large_ds_sid, H5S_SELECT_OR, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, or) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, or) succeeded");
#if LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG
if (mpi_rank == LOWER_DIM_SIZE_COMP_TEST_DEBUG_TARGET_RANK) {
@@ -2426,7 +2426,7 @@ link_chunk_collective_io_test(void)
/* select the file and mem spaces */
start[0] = (hsize_t)(mpi_rank * LINK_CHUNK_COLLECTIVE_IO_TEST_CHUNK_SIZE);
ret = H5Sselect_hyperslab(file_ds_sid, H5S_SELECT_SET, start, stride, count, block);
- VRFY((ret >= 0), "H5Sselect_hyperslab(file_ds_sid, set) suceeded");
+ VRFY((ret >= 0), "H5Sselect_hyperslab(file_ds_sid, set) succeeded");
ret = H5Sselect_all(write_mem_ds_sid);
VRFY((ret != FAIL), "H5Sselect_all(mem_ds_sid) succeeded");
diff --git a/testpar/testphdf5.h b/testpar/testphdf5.h
index f015005..cf41f9b 100644
--- a/testpar/testphdf5.h
+++ b/testpar/testphdf5.h
@@ -214,8 +214,8 @@ typedef int DATATYPE;
/* Shape Same Tests Definitions */
typedef enum {
- IND_CONTIG, /* Independent IO on contigous datasets */
- COL_CONTIG, /* Collective IO on contigous datasets */
+ IND_CONTIG, /* Independent IO on contiguous datasets */
+ COL_CONTIG, /* Collective IO on contiguous datasets */
IND_CHUNKED, /* Independent IO on chunked datasets */
COL_CHUNKED /* Collective IO on chunked datasets */
} ShapeSameTestMethods;
diff --git a/tools/lib/h5diff.c b/tools/lib/h5diff.c
index 0ad319e..2275c72 100644
--- a/tools/lib/h5diff.c
+++ b/tools/lib/h5diff.c
@@ -1228,7 +1228,7 @@ diff_match(hid_t file1_id, const char *grp1, trav_info_t *info1, hid_t file2_id,
H5TOOLS_DEBUG("Beginning of big else block");
/* We're in parallel mode */
/* Since the data type of diff value is hsize_t which can
- * be arbitary large such that there is no MPI type that
+ * be arbitrary large such that there is no MPI type that
* matches it, the value is passed between processes as
* an array of bytes in order to be portable. But this
* may not work in non-homogeneous MPI environments.
diff --git a/tools/src/h5dump/h5dump_ddl.c b/tools/src/h5dump/h5dump_ddl.c
index bb495f2..868b75d 100644
--- a/tools/src/h5dump/h5dump_ddl.c
+++ b/tools/src/h5dump/h5dump_ddl.c
@@ -1959,7 +1959,7 @@ handle_datatypes(hid_t fid, const char *type, void H5_ATTR_UNUSED *data, int pe,
const char *real_name = display_name ? display_name : type;
if ((type_id = H5Topen2(fid, type, H5P_DEFAULT)) < 0) {
- /* check if type is unamed datatype */
+ /* check if type is unnamed datatype */
unsigned idx = 0;
while (idx < type_table->nobjs) {
diff --git a/tools/src/h5dump/h5dump_xml.c b/tools/src/h5dump/h5dump_xml.c
index ccf4156..5f64d6c 100644
--- a/tools/src/h5dump/h5dump_xml.c
+++ b/tools/src/h5dump/h5dump_xml.c
@@ -23,7 +23,7 @@
const char *xmlnsprefix = "hdf5:";
/*
- * Alternative formating for data dumped to XML
+ * Alternative formatting for data dumped to XML
* In general, the numbers are the same, but separators
* except spaces are not used.
*
@@ -2795,7 +2795,7 @@ xml_dump_group(hid_t gid, const char *name)
if (isRoot && unamedtype) {
unsigned u;
- /* Very special case: dump unamed type in root group */
+ /* Very special case: dump unnamed type in root group */
for (u = 0; u < type_table->nobjs; u++) {
if (!type_table->objs[u].recorded) {
char *obj_tok_str = NULL;
@@ -2888,7 +2888,7 @@ xml_dump_group(hid_t gid, const char *name)
if (isRoot && unamedtype) {
unsigned u;
- /* Very special case: dump unamed type in root group */
+ /* Very special case: dump unnamed type in root group */
for (u = 0; u < type_table->nobjs; u++) {
if (!type_table->objs[u].recorded) {
char *obj_tok_str = NULL;
@@ -2940,7 +2940,7 @@ xml_dump_group(hid_t gid, const char *name)
/*-------------------------------------------------------------------------
* Function: xml_print_refs
*
- * Purpose: Print a path to the objects referenced by HDF5 Referneces.
+ * Purpose: Print a path to the objects referenced by HDF5 References.
*
* Return: void
*
diff --git a/tools/src/h5import/h5import.c b/tools/src/h5import/h5import.c
index ddd34ce..317db81 100644
--- a/tools/src/h5import/h5import.c
+++ b/tools/src/h5import/h5import.c
@@ -3725,7 +3725,7 @@ getCompressionParameter(struct Input *in, FILE *strm)
int ival;
const char *err1 = "Unable to get integer value.\n";
- const char *err2 = "Invalid value for compression paramter.\n";
+ const char *err2 = "Invalid value for compression parameter.\n";
const char *err3 = "Unsupported Compression Type.\n";
switch (in->compressionType) {
diff --git a/tools/src/h5perf/pio_engine.c b/tools/src/h5perf/pio_engine.c
index e35f86e..90d63c8 100644
--- a/tools/src/h5perf/pio_engine.c
+++ b/tools/src/h5perf/pio_engine.c
@@ -246,7 +246,7 @@ do_pio(parameters param)
if (((size_t)(snbytes / pio_mpi_nprocs_g) % buf_size) != 0) {
HDfprintf(stderr,
"Dataset size/process (%" H5_PRINTF_LL_WIDTH "d) must be a multiple of the "
- "trasfer buffer size (%zu)\n",
+ "transfer buffer size (%zu)\n",
(long long)(snbytes / pio_mpi_nprocs_g), buf_size);
GOTOERROR(FAIL);
}
@@ -255,7 +255,7 @@ do_pio(parameters param)
if (((size_t)snbytes % buf_size) != 0) {
HDfprintf(stderr,
"Dataset side size (%" H5_PRINTF_LL_WIDTH "d) must be a multiple of the "
- "trasfer buffer size (%zu)\n",
+ "transfer buffer size (%zu)\n",
(long long)snbytes, buf_size);
GOTOERROR(FAIL);
}
diff --git a/tools/src/h5perf/sio_engine.c b/tools/src/h5perf/sio_engine.c
index dc2c68e..dda5bd2 100644
--- a/tools/src/h5perf/sio_engine.c
+++ b/tools/src/h5perf/sio_engine.c
@@ -181,7 +181,7 @@ do_sio(parameters param, results *res)
if ((param.dset_size[i] % param.buf_size[i]) != 0) {
HDfprintf(stderr,
"Dataset size[%d] (%" H5_PRINTF_LL_WIDTH "d) must be a multiple of the "
- "trasfer buffer size[%d] (%zu)\n",
+ "transfer buffer size[%d] (%zu)\n",
param.rank, (long long)param.dset_size[i], param.rank, param.buf_size[i]);
GOTOERROR(FAIL);
}
diff --git a/tools/src/h5perf/sio_perf.c b/tools/src/h5perf/sio_perf.c
index b2dd765..f299c04 100644
--- a/tools/src/h5perf/sio_perf.c
+++ b/tools/src/h5perf/sio_perf.c
@@ -1278,7 +1278,7 @@ usage(const char *prog)
HDprintf("\n");
HDprintf(" Example: 2K,2K,3K\n");
HDprintf("\n");
- HDprintf(" The example defines an object (dataset, tranfer buffer) with three\n");
+ HDprintf(" The example defines an object (dataset, transfer buffer) with three\n");
HDprintf(" dimensions. Be aware that as the number of dimensions increases, the\n");
HDprintf(" the total size of the object increases exponentially.\n");
HDprintf("\n");
diff --git a/tools/src/h5repack/h5repack.h b/tools/src/h5repack/h5repack.h
index b242a71..65e5d40 100644
--- a/tools/src/h5repack/h5repack.h
+++ b/tools/src/h5repack/h5repack.h
@@ -124,7 +124,7 @@ typedef struct {
int grp_compact; /* Set the maximum number of links to store as header messages in the group */
int grp_indexed; /* Set the minimum number of links to store in the indexed format */
int msg_size[8]; /* Minimum size of shared messages: dataspace,
- datatype, fill value, filter pipleline, attribute */
+ datatype, fill value, filter pipeline, attribute */
const char * ublock_filename; /* user block file name */
hsize_t ublock_size; /* user block size */
hsize_t meta_block_size; /* metadata aggregation block size (for H5Pset_meta_block_size) */
diff --git a/tools/src/h5repack/h5repack_copy.c b/tools/src/h5repack/h5repack_copy.c
index 65843d4..772c3a7 100644
--- a/tools/src/h5repack/h5repack_copy.c
+++ b/tools/src/h5repack/h5repack_copy.c
@@ -67,7 +67,7 @@ copy_objects(const char *fnamein, const char *fnameout, pack_opt_t *options)
hid_t fcpl = H5P_DEFAULT; /* file creation property list ID */
trav_table_t * travt = NULL;
hsize_t ub_size = 0; /* size of user block */
- H5F_fspace_strategy_t set_strategy; /* Strategy to be set in outupt file */
+ H5F_fspace_strategy_t set_strategy; /* Strategy to be set in output file */
hbool_t set_persist; /* Persist free-space status to be set in output file */
hsize_t set_threshold; /* Free-space section threshold to be set in output file */
hsize_t set_pagesize; /* File space page size to be set in output file */
@@ -188,7 +188,7 @@ copy_objects(const char *fnamein, const char *fnameout, pack_opt_t *options)
H5TOOLS_GOTO_ERROR((-1), "H5Pset_shared_mesg_nindexes failed to set the number of shared "
"object header message indexes");
- /* msg_size[0]=dataspace, 1=datatype, 2=file value, 3=filter pipleline, 4=attribute */
+ /* msg_size[0]=dataspace, 1=datatype, 2=file value, 3=filter pipeline, 4=attribute */
for (i = 0; i < (nindex - 1); i++)
if (H5Pset_shared_mesg_index(fcpl, i, mesg_type_flags[i], min_mesg_sizes[i]) < 0)
H5TOOLS_GOTO_ERROR((-1), "H5Pset_shared_mesg_index failed to configure the specified "
@@ -402,18 +402,18 @@ done:
/*-------------------------------------------------------------------------
* Function: get_hyperslab
*
- * Purpose: Calulate a hyperslab from a dataset for higher performance.
- * The size of hyperslab is limitted by H5TOOLS_BUFSIZE.
- * Return the hyperslab dimentions and size in byte.
+ * Purpose: Calculate a hyperslab from a dataset for higher performance.
+ * The size of hyperslab is limited by H5TOOLS_BUFSIZE.
+ * Return the hyperslab dimensions and size in byte.
*
* Return: 0 - SUCCEED, -1 FAILED
*
* Parameters:
* dcpl_id : [IN] dataset creation property.
* rank_dset : [IN] dataset rank
- * dims_dset[] : [IN] dataset dimentions
+ * dims_dset[] : [IN] dataset dimensions
* size_datum : [IN] size of a data element in byte
- * dims_hslab[] : [OUT] calculated hyperslab dimentions
+ * dims_hslab[] : [OUT] calculated hyperslab dimensions
* * hslab_nbytes_p : [OUT] total byte of the hyperslab
*
* Update:
@@ -428,7 +428,7 @@ done:
* 3. If not chunked, each data element would be a unit of collection and
* the boundary would be dataset's dims.
*
- * The calulation starts from the last dimention (h5dump dims output).
+ * The calculation starts from the last dimension (h5dump dims output).
*-----------------------------------------*/
int
@@ -442,8 +442,8 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
hsize_t size_chunk = 1;
hsize_t nchunk_fit; /* number of chunks that fits in hyperslab buffer (H5TOOLS_BUFSIZE) */
hsize_t ndatum_fit; /* number of dataum that fits in hyperslab buffer (H5TOOLS_BUFSIZE) */
- hsize_t chunk_dims_map[H5S_MAX_RANK]; /* mapped chunk dimentions */
- hsize_t hs_dims_map[H5S_MAX_RANK]; /* mapped hyperslab dimentions */
+ hsize_t chunk_dims_map[H5S_MAX_RANK]; /* mapped chunk dimensions */
+ hsize_t hs_dims_map[H5S_MAX_RANK]; /* mapped hyperslab dimensions */
hsize_t hslab_nbytes; /* size of hyperslab in byte */
int ret_value = 0;
@@ -468,18 +468,18 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
/* 1. if a chunk fit in hyperslab buffer */
if (nchunk_fit >= 1) {
- /* Calulate a hyperslab that contains as many chunks that can fit
+ /* Calculate a hyperslab that contains as many chunks that can fit
* in hyperslab buffer. Hyperslab will be increased starting from
- * the last dimention of the dataset (see h5dump's dims output).
+ * the last dimension of the dataset (see h5dump's dims output).
* The calculation boundary is dataset dims.
* In the loop, used mapping from a datum to a chunk to figure out
* chunk based hyperslab.
*/
for (k = rank_dset; k > 0; --k) {
- /* map dataset dimentions with a chunk dims */
+ /* map dataset dimensions with a chunk dims */
chunk_dims_map[k - 1] = dims_dset[k - 1] / dims_chunk[k - 1];
- /* if reminder exist, increse by 1 to cover partial edge chunks */
+ /* if reminder exist, increase by 1 to cover partial edge chunks */
if (dims_dset[k - 1] % dims_chunk[k - 1] > 0)
chunk_dims_map[k - 1]++;
@@ -492,7 +492,7 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
if (nchunk_fit == 0)
nchunk_fit = 1;
- /* get hyperslab dimentions as unmapping to actual size */
+ /* get hyperslab dimensions as unmapping to actual size */
dims_hslab[k - 1] = MIN((hs_dims_map[k - 1] * dims_chunk[k - 1]), dims_dset[k - 1]);
/* calculate total size for the hyperslab */
@@ -501,9 +501,9 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
}
/* 2. if a chunk is bigger than hyperslab buffer */
else {
- /* Calulate a hyperslab that contains as many data elements that
+ /* Calculate a hyperslab that contains as many data elements that
* can fit in hyperslab buffer. Hyperslab will be increased
- * starting from the last dimention of the chunk (see h5dump's dims
+ * starting from the last dimension of the chunk (see h5dump's dims
* output).
* The calculation boundary is a chunk dims.
*/
@@ -513,7 +513,7 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
/* if a datum is bigger than rest of buffer */
if (ndatum_fit == 0)
ndatum_fit = 1;
- /* get hyperslab dimentions within a chunk boundary */
+ /* get hyperslab dimensions within a chunk boundary */
dims_hslab[k - 1] = MIN(dims_chunk[k - 1], ndatum_fit);
/* calculate total size for the hyperslab */
@@ -526,9 +526,9 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
}
/* 3. if dataset is not chunked */
else {
- /* Calulate a hyperslab that contains as many data elements that can
+ /* Calculate a hyperslab that contains as many data elements that can
* fit in hyperslab buffer. Hyperslab will be increased starting from
- * the last dimention of the dataset (see h5dump's dims output).
+ * the last dimension of the dataset (see h5dump's dims output).
* The calculation boundary is dataset dims.
*/
for (k = rank_dset; k > 0; --k) {
@@ -537,7 +537,7 @@ get_hyperslab(hid_t dcpl_id, int rank_dset, const hsize_t dims_dset[], size_t si
/* if a datum is bigger than rest of buffer */
if (ndatum_fit == 0)
ndatum_fit = 1;
- /* get hyperslab dimentions within dataset boundary */
+ /* get hyperslab dimensions within dataset boundary */
dims_hslab[k - 1] = MIN(dims_dset[k - 1], ndatum_fit);
/* calculate total size for the hyperslab */
@@ -601,7 +601,7 @@ done:
* in (2) is that, when using the strip mine size, it assures that the "remaining" part
* of the dataset that does not fill an entire strip mine is processed.
*
- * 1. figure out a hyperslab (dimentions) and size (refer to get_hyperslab()).
+ * 1. figure out a hyperslab (dimensions) and size (refer to get_hyperslab()).
* 2. Calculate the hyperslab selections as the selection is moving forward.
* Selection would be same as the hyperslab except for the remaining edge portion
* of the dataset. The code take care of the remaining portion if exist.
@@ -1621,7 +1621,7 @@ copy_user_block(const char *infile, const char *outfile, hsize_t size)
/* Copy the userblock from the input file to the output file */
while (size > 0) {
- ssize_t nread, nbytes; /* # of bytes transfered, etc. */
+ ssize_t nread, nbytes; /* # of bytes transferred, etc. */
char rbuf[USERBLOCK_XFER_SIZE]; /* Buffer for reading */
const char *wbuf; /* Pointer into buffer, for writing */
diff --git a/tools/src/h5repack/h5repack_filters.c b/tools/src/h5repack/h5repack_filters.c
index 32476ce..76bac5e 100644
--- a/tools/src/h5repack/h5repack_filters.c
+++ b/tools/src/h5repack/h5repack_filters.c
@@ -231,7 +231,7 @@ aux_assign_obj(const char * name, /* object name from traverse list */
* do extra checking in the case of SZIP; delete all filters in the case
* of H5Z_FILTER_NONE present in the PACK_INFO_T filter array
*
- * Return: 0 success, -1 an error occured
+ * Return: 0 success, -1 an error occurred
*-------------------------------------------------------------------------
*/
@@ -258,7 +258,7 @@ apply_filters(const char * name, /* object name from traverse list */
H5TOOLS_GOTO_DONE(0);
/*-------------------------------------------------------------------------
- * initialize the assigment object
+ * initialize the assignment object
*-------------------------------------------------------------------------
*/
init_packobject(&obj);
diff --git a/tools/src/h5repack/h5repack_main.c b/tools/src/h5repack/h5repack_main.c
index 98d2a73..87a83bc 100644
--- a/tools/src/h5repack/h5repack_main.c
+++ b/tools/src/h5repack/h5repack_main.c
@@ -39,13 +39,13 @@ static struct long_options l_opts[] = {{"alignment", require_arg, 'a'},
{"file", require_arg, 'e'},
{"filter", require_arg, 'f'},
{"help", no_arg, 'h'},
- {"infile", require_arg, 'i'}, /* for backward compability */
+ {"infile", require_arg, 'i'}, /* for backward compatibility */
{"low", require_arg, 'j'},
{"high", require_arg, 'k'},
{"layout", require_arg, 'l'},
{"minimum", require_arg, 'm'},
{"native", no_arg, 'n'},
- {"outfile", require_arg, 'o'}, /* for backward compability */
+ {"outfile", require_arg, 'o'}, /* for backward compatibility */
{"sort_by", require_arg, 'q'},
{"ssize", require_arg, 's'},
{"threshold", require_arg, 't'},
diff --git a/tools/src/h5repack/h5repack_refs.c b/tools/src/h5repack/h5repack_refs.c
index 6d62e52..3259b89 100644
--- a/tools/src/h5repack/h5repack_refs.c
+++ b/tools/src/h5repack/h5repack_refs.c
@@ -539,7 +539,7 @@ copy_refs_attr(hid_t loc_in, hid_t loc_out, trav_table_t *travt, hid_t fidout) /
}
/* This line below needs to be moved in this loop instead of inserting outside. Otherwise,
ref_comp_field_n may be >0 for the next attribute, which may not be
- the reference type and will be accidently treated as the reference type.
+ the reference type and will be accidentally treated as the reference type.
It will then cause the H5Acreate2 failed since that attribute is already created.
KY 2020-02-07
*/
diff --git a/tools/src/h5repack/h5repack_verify.c b/tools/src/h5repack/h5repack_verify.c
index 0542d03..98883b2 100644
--- a/tools/src/h5repack/h5repack_verify.c
+++ b/tools/src/h5repack/h5repack_verify.c
@@ -397,7 +397,7 @@ h5repack_cmp_pl(const char *fname1, hid_t fname1_fapl, const char *fname2, hid_t
H5TOOLS_GOTO_ERROR((-1), "h5trav_gettable failed");
/*-------------------------------------------------------------------------
- * traverse the suppplied object list
+ * traverse the supplied object list
*-------------------------------------------------------------------------
*/
for (i = 0; i < trav->nobjs; i++) {
diff --git a/tools/test/h5copy/h5copygentest.c b/tools/test/h5copy/h5copygentest.c
index 429ad42..2b40470 100644
--- a/tools/test/h5copy/h5copygentest.c
+++ b/tools/test/h5copy/h5copygentest.c
@@ -618,7 +618,7 @@ gen_region_ref(hid_t loc_id)
goto out;
}
- /* create normal dataset which is refered */
+ /* create normal dataset which is referred */
oid2 = H5Dcreate2(loc_id, REG_REF_DS2, H5T_STD_I8LE, sid, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
if (oid2 < 0) {
HDfprintf(stderr, "Error: %s %d> H5Dcreate2 failed.\n", FUNC, __LINE__);
diff --git a/tools/test/h5copy/testh5copy.sh.in b/tools/test/h5copy/testh5copy.sh.in
index 79b6fd5..6349ef4 100644
--- a/tools/test/h5copy/testh5copy.sh.in
+++ b/tools/test/h5copy/testh5copy.sh.in
@@ -88,7 +88,7 @@ test -d $TESTDIR || mkdir $TESTDIR
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
diff --git a/tools/test/h5diff/h5diff_plugin.sh.in b/tools/test/h5diff/h5diff_plugin.sh.in
index 0ab169a..28f723e 100644
--- a/tools/test/h5diff/h5diff_plugin.sh.in
+++ b/tools/test/h5diff/h5diff_plugin.sh.in
@@ -158,7 +158,7 @@ done
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -236,7 +236,7 @@ TOOLTEST() {
# don't add exit code check in pmode, as it causes failure. (exit code
# is from mpirun not tool)
# if any problem occurs relate to an exit code, it will be caught in
- # serial mode, so the test is fullfilled.
+ # serial mode, so the test is fulfilled.
if test $h5haveexitcode = 'yes' -a -z "$pmode"; then
echo "EXIT CODE: $EXIT_CODE" >> $actual
fi
@@ -327,7 +327,7 @@ TOOLTEST_ERR() {
# don't add exit code check in pmode, as it causes failure. (exit code
# is from mpirun not tool)
# if any problem occurs relate to an exit code, it will be caught in
- # serial mode, so the test is fullfilled.
+ # serial mode, so the test is fulfilled.
if test $h5haveexitcode = 'yes' -a -z "$pmode"; then
echo "EXIT CODE: $EXIT_CODE" >> $actual
fi
diff --git a/tools/test/h5diff/h5diffgentest.c b/tools/test/h5diff/h5diffgentest.c
index f7e5c52..6abd265 100644
--- a/tools/test/h5diff/h5diffgentest.c
+++ b/tools/test/h5diff/h5diffgentest.c
@@ -82,7 +82,7 @@ size_t H5TOOLS_MALLOCSIZE = (128 * 1024 * 1024);
#define EXCLUDE_FILE3_2 "h5diff_exclude3-2.h5"
/* compound type with multiple vlen string types */
#define COMP_VL_STRS_FILE "h5diff_comp_vl_strs.h5"
-/* attribute compre with verbose level */
+/* attribute compare with verbose level */
#define ATTR_VERBOSE_LEVEL_FILE1 "h5diff_attr_v_level1.h5"
#define ATTR_VERBOSE_LEVEL_FILE2 "h5diff_attr_v_level2.h5"
/* file containing valid/invalid enum value mix */
@@ -2904,7 +2904,7 @@ test_group_recurse2(void)
* FILE 2-3
*------------------------------------------------------------------------*/
- /* crate target file */
+ /* create target file */
fileid4 = H5Fcreate(GRP_RECURSE2_EXT3, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
/*-----------------------------------------------
@@ -2938,7 +2938,7 @@ test_group_recurse2(void)
* FILE 2-2
*------------------------------------------------------------------------*/
- /* crate target file */
+ /* create target file */
fileid3 = H5Fcreate(GRP_RECURSE2_EXT2, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
/*-----------------------------------------------
@@ -2990,7 +2990,7 @@ test_group_recurse2(void)
* FILE 2-1
*------------------------------------------------------------------------*/
- /* crate target file */
+ /* create target file */
fileid2 = H5Fcreate(GRP_RECURSE2_EXT1, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
/*-----------------------------------------------
@@ -4856,7 +4856,7 @@ test_data_nocomparables(const char *fname, int make_diffs)
/* attr1 - non-compatible : different type */
write_attr(did2, 1, dims1_1, "attr1", tid_attr1, attr_data_ptr2);
- /* attr2 - non-compatible : same rank, different dimention */
+ /* attr2 - non-compatible : same rank, different dimension */
write_attr(did2, 1, (hsize_t *)attr2_dim_ptr, "attr2", H5T_NATIVE_INT, data3);
/* attr3 - non-compatible : different rank */
@@ -7905,7 +7905,7 @@ gen_datareg(hid_t fid, int make_diffs /* flag to modify data buffers */)
* Function: test_hyperslab
*
* Purpose: test diff by hyperslabs. create a dataset with 1GB dimensions
- * by iterating trough 1KB hyperslabs
+ * by iterating through 1KB hyperslabs
*
*-------------------------------------------------------------------------
*/
diff --git a/tools/test/h5diff/testh5diff.sh.in b/tools/test/h5diff/testh5diff.sh.in
index bd27dd0..6871463 100644
--- a/tools/test/h5diff/testh5diff.sh.in
+++ b/tools/test/h5diff/testh5diff.sh.in
@@ -455,7 +455,7 @@ done
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -518,7 +518,7 @@ TOOLTEST() {
# don't add exit code check in pmode, as it causes failure. (exit code
# is from mpirun not tool)
# if any problem occurs relate to an exit code, it will be caught in
- # serial mode, so the test is fullfilled.
+ # serial mode, so the test is fulfilled.
if test $h5haveexitcode = 'yes' -a -z "$pmode"; then
echo "EXIT CODE: $EXIT_CODE" >> $actual
fi
@@ -926,7 +926,7 @@ TOOLTEST h5diff_207.txt -c h5diff_basic2.h5 h5diff_basic2.h5 g2/dset8 g2/dset9
TOOLTEST h5diff_208.txt -c h5diff_dset_zero_dim_size1.h5 h5diff_dset_zero_dim_size2.h5
# non-comparable dataset with comparable attribute, and other comparable datasets.
-# Also test non-compatible attributes with different type, dimention, rank.
+# Also test non-compatible attributes with different type, dimension, rank.
# All the comparables should display differences.
TOOLTEST h5diff_220.txt -c non_comparables1.h5 non_comparables2.h5 /g1
@@ -1089,7 +1089,7 @@ TOOLTEST h5diff_474.txt -v h5diff_danglelinks1.h5 h5diff_danglelinks2.h5 /ext_li
TOOLTEST h5diff_475.txt -v h5diff_danglelinks1.h5 h5diff_danglelinks2.h5 /ext_link1
# ##############################################################################
-# # test for group diff recursivly
+# # test for group diff recursively
# ##############################################################################
# root
TOOLTEST h5diff_500.txt -v h5diff_grp_recurse1.h5 h5diff_grp_recurse2.h5 / /
diff --git a/tools/test/h5dump/h5dumpgentest.c b/tools/test/h5dump/h5dumpgentest.c
index 261fb09..e271511 100644
--- a/tools/test/h5dump/h5dumpgentest.c
+++ b/tools/test/h5dump/h5dumpgentest.c
@@ -236,7 +236,7 @@ typedef struct s1_t {
#define ARRAY3_DIM2 3
/* VL string datatype name */
-/* TODO remove complier error not used, remove the link when everything is OK */
+/* TODO remove compiler error not used, remove the link when everything is OK */
/* #define VLSTR_TYPE "vl_string_type" */
/* "File 41" macros */
@@ -1128,7 +1128,7 @@ gent_compound_dt(void)
H5Dclose(dataset);
H5Gclose(group);
- /* unamed data type */
+ /* unnamed data type */
group = H5Gcreate2(fid, "/group2", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
type = H5Tcreate(H5T_COMPOUND, sizeof(dset5_t));
@@ -1312,7 +1312,7 @@ gent_compound_dt2(void)
H5Dclose(dataset);
H5Gclose(group);
- /* unamed data type */
+ /* unnamed data type */
group = H5Gcreate2(fid, "/group2", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
type = H5Tcreate(H5T_COMPOUND, sizeof(dset5_t));
@@ -1522,11 +1522,11 @@ gent_loop2(void)
fid = H5Fcreate(FILE11, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT);
- /* create group object g1 and implcit path from root object */
+ /* create group object g1 and implicit path from root object */
group = H5Gcreate2(fid, "/g1", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
H5Gclose(group);
- /* create group object g2 and implcit path from root object */
+ /* create group object g2 and implicit path from root object */
group = H5Gcreate2(fid, "/g2", H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
H5Gclose(group);
@@ -2468,7 +2468,7 @@ gent_attrreg(void)
static void
gent_nestcomp(void)
{
- /* Compound memeber of the compound datatype*/
+ /* Compound member of the compound datatype*/
typedef struct cmp_t {
char a;
float b[2];
@@ -2541,7 +2541,7 @@ gent_nestcomp(void)
H5Tinsert(s2_tid, "c_name", HOFFSET(s2_t, c), H5T_NATIVE_DOUBLE);
H5Tinsert(s2_tid, "b_name", HOFFSET(s2_t, b), H5T_NATIVE_FLOAT);
- /* Insert compound memeber created above */
+ /* Insert compound member created above */
H5Tinsert(s2_tid, "d_name", HOFFSET(s2_t, d), cmp_tid);
/*
@@ -5218,7 +5218,7 @@ gent_named_dtype_attr(void)
sid = H5Screate(H5S_SCALAR);
HDassert(sid > 0);
- /* Create attribute on commited datatype */
+ /* Create attribute on committed datatype */
aid = H5Acreate2(tid, F42_ATTRNAME, H5T_STD_I32LE, sid, H5P_DEFAULT, H5P_DEFAULT);
HDassert(aid > 0);
@@ -5735,7 +5735,7 @@ gent_filters(void)
HDassert(ret >= 0);
if (supports_comments) {
- ret = H5Oset_comment_by_name(fid, "mytype", "This is a commited datatype", H5P_DEFAULT);
+ ret = H5Oset_comment_by_name(fid, "mytype", "This is a committed datatype", H5P_DEFAULT);
HDassert(ret >= 0);
}
@@ -6803,7 +6803,7 @@ gent_attr_creation_order(void)
goto out;
/*-------------------------------------------------------------------------
- * create a dataset with creation order tracked for attributes and atributes in it
+ * create a dataset with creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6827,7 +6827,7 @@ gent_attr_creation_order(void)
did = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a dataset without creation order tracked for attributes and atributes in it
+ * create a dataset without creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6851,7 +6851,7 @@ gent_attr_creation_order(void)
did = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a group with creation order tracked for attributes and atributes in it
+ * create a group with creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6874,7 +6874,7 @@ gent_attr_creation_order(void)
gid = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a group without creation order tracked for attributes and atributes in it
+ * create a group without creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6897,7 +6897,7 @@ gent_attr_creation_order(void)
gid = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a named datatype with creation order tracked for attributes and atributes in it
+ * create a named datatype with creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
@@ -6923,7 +6923,7 @@ gent_attr_creation_order(void)
tid = H5I_INVALID_HID;
/*-------------------------------------------------------------------------
- * create a named datatype without creation order tracked for attributes and atributes in it
+ * create a named datatype without creation order tracked for attributes and attributes in it
*-------------------------------------------------------------------------
*/
diff --git a/tools/test/h5dump/testh5dump.sh.in b/tools/test/h5dump/testh5dump.sh.in
index 9464302..f9f3734 100644
--- a/tools/test/h5dump/testh5dump.sh.in
+++ b/tools/test/h5dump/testh5dump.sh.in
@@ -1171,7 +1171,7 @@ TOOLTEST tcomp-1.ddl --enable-error-stack tcompound.h5
# test for named data types
TOOLTEST tcomp-2.ddl --enable-error-stack -t /type1 --datatype /type2 --datatype=/group1/type3 tcompound.h5
TOOLTEST tcomp-2.ddl --enable-error-stack -N /type1 --any_path /type2 --any_path=/group1/type3 tcompound.h5
-# test for unamed type
+# test for unnamed type
TOOLTEST4 tcomp-3.ddl --enable-error-stack -t /#6632 -g /group2 tcompound.h5
# test complicated compound datatype
TOOLTEST tcomp-4.ddl --enable-error-stack tcompound_complex.h5
diff --git a/tools/test/h5dump/testh5dumppbits.sh.in b/tools/test/h5dump/testh5dumppbits.sh.in
index 6689749..eaf91fd 100644
--- a/tools/test/h5dump/testh5dumppbits.sh.in
+++ b/tools/test/h5dump/testh5dumppbits.sh.in
@@ -516,7 +516,7 @@ TOOLTEST tnofilename-with-packed-bits.ddl --enable-error-stack
# Limits:
# Maximum number of packed bits is 8 (for now).
# Maximum integer size is 64 (for now).
-# Maximun Offset is 63 (Maximum size - 1).
+# Maximum Offset is 63 (Maximum size - 1).
# Maximum Offset+Length is 64 (Maximum size).
# Tests:
# Normal operation on both signed and unsigned int datasets.
diff --git a/tools/test/h5format_convert/h5fc_gentest.c b/tools/test/h5format_convert/h5fc_gentest.c
index 09cca45..ccb64b5 100644
--- a/tools/test/h5format_convert/h5fc_gentest.c
+++ b/tools/test/h5format_convert/h5fc_gentest.c
@@ -381,7 +381,7 @@ error:
/*
* Function: gen_err_level()
*
- * Generate a file to test the situtation described in HDFFV-9434:
+ * Generate a file to test the situation described in HDFFV-9434:
* Exceed the limit of v1-btree level
*
* Create a file with H5Pset_istore_k(fcpl, 1).
diff --git a/tools/test/h5format_convert/testh5fc.sh.in b/tools/test/h5format_convert/testh5fc.sh.in
index 762f43d..df083d1 100644
--- a/tools/test/h5format_convert/testh5fc.sh.in
+++ b/tools/test/h5format_convert/testh5fc.sh.in
@@ -313,7 +313,7 @@ TOOLTEST_MASK_OUT() {
$RM $actual $actual_err $actual_sav $actual_err_sav
fi
}
-# Same as TOOLTEST_OUT except only compares error outout
+# Same as TOOLTEST_OUT except only compares error output
TOOLTEST_ERR() {
# Prepare expected and actual output
expect="$TESTDIR/$1"
diff --git a/tools/test/h5jam/testh5jam.sh.in b/tools/test/h5jam/testh5jam.sh.in
index caa3416..191fef0 100644
--- a/tools/test/h5jam/testh5jam.sh.in
+++ b/tools/test/h5jam/testh5jam.sh.in
@@ -345,7 +345,7 @@ CHECK_NOUB() {
# JAMTEST user_block file.h5 [--clobber] [ofile.h5]
#
# Test the 'jam' tool:
-# 1. figure out the input and output, and the comparision
+# 1. figure out the input and output, and the comparison
# that will be done.
# 2. call 'jam' with the appropriate arguments
# 3. check the user block is correct in the output (Check_UB)
diff --git a/tools/test/h5ls/h5ls_plugin.sh.in b/tools/test/h5ls/h5ls_plugin.sh.in
index 29e982d..0077b07 100644
--- a/tools/test/h5ls/h5ls_plugin.sh.in
+++ b/tools/test/h5ls/h5ls_plugin.sh.in
@@ -166,7 +166,7 @@ VERIFY() {
# the `nerrors' global variable and (if $verbose is set) display up to $NLINS
# lines of the actual output from the tool test. The actual output is not
# removed if $HDF5_NOCLEANUP has a non-zero value.
-# Arguemnts:
+# Arguments:
# $1 -- actual output filename to use
# $2 and on -- argument for the h5ls tool
TOOLTEST() {
diff --git a/tools/test/h5ls/testh5ls.sh.in b/tools/test/h5ls/testh5ls.sh.in
index ee0dad7..36583da 100644
--- a/tools/test/h5ls/testh5ls.sh.in
+++ b/tools/test/h5ls/testh5ls.sh.in
@@ -164,7 +164,7 @@ $SRC_H5LS_TESTFILES/tdset_idx.ls
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -230,7 +230,7 @@ TESTING() {
# the `nerrors' global variable and (if $verbose is set) display up to $NLINS
# lines of the actual output from the tool test. The actual output is not
# removed if $HDF5_NOCLEANUP has a non-zero value.
-# Arguemnts:
+# Arguments:
# $1 -- actual output filename to use
# $2 and on -- argument for the h5ls tool
TOOLTEST() {
diff --git a/tools/test/h5ls/testh5lsvds.sh.in b/tools/test/h5ls/testh5lsvds.sh.in
index 4814f85..ab2d137 100644
--- a/tools/test/h5ls/testh5lsvds.sh.in
+++ b/tools/test/h5ls/testh5lsvds.sh.in
@@ -103,7 +103,7 @@ $SRC_H5LS_TESTFILES/vds/tvds-5.ls
# RUNSERIAL is used. Check if it can return exit code from executalbe correctly.
if [ -n "$RUNSERIAL_NOEXITCODE" ]; then
- echo "***Warning*** Serial Exit Code is not passed back to shell corretly."
+ echo "***Warning*** Serial Exit Code is not passed back to shell correctly."
echo "***Warning*** Exit code checking is skipped."
h5haveexitcode=no
fi
@@ -169,7 +169,7 @@ TESTING() {
# the `nerrors' global variable and (if $verbose is set) display up to $NLINS
# lines of the actual output from the tool test. The actual output is not
# removed if $HDF5_NOCLEANUP has a non-zero value.
-# Arguemnts:
+# Arguments:
# $1 -- actual output filename to use
# $2 and on -- argument for the h5ls tool
TOOLTEST() {
diff --git a/tools/test/h5repack/dynlib_rpk.c b/tools/test/h5repack/dynlib_rpk.c
index b228344..bef032a 100644
--- a/tools/test/h5repack/dynlib_rpk.c
+++ b/tools/test/h5repack/dynlib_rpk.c
@@ -83,7 +83,7 @@ H5Z_filter_dynlib1(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
add_on = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) { /*read*/
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= add_on;
buf_left -= sizeof(int);
diff --git a/tools/test/h5repack/dynlib_vrpk.c b/tools/test/h5repack/dynlib_vrpk.c
index 5eac4f2..a45de65 100644
--- a/tools/test/h5repack/dynlib_vrpk.c
+++ b/tools/test/h5repack/dynlib_vrpk.c
@@ -91,7 +91,7 @@ H5Z_filter_dynlib4(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_
add_on = (int)cd_values[0];
if (flags & H5Z_FLAG_REVERSE) { /*read*/
- /* Substract the "add on" value to all the data values */
+ /* Subtract the "add on" value to all the data values */
while (buf_left > 0) {
*int_ptr++ -= add_on;
buf_left -= sizeof(int);
diff --git a/tools/test/h5repack/h5repack.sh.in b/tools/test/h5repack/h5repack.sh.in
index 1e54670..a241ea8 100644
--- a/tools/test/h5repack/h5repack.sh.in
+++ b/tools/test/h5repack/h5repack.sh.in
@@ -525,7 +525,7 @@ VERIFY_LAYOUT_ALL()
#---------------------------------
# check the layout from a dataset
- # check if the other layouts still exsit
+ # check if the other layouts still exist
VERIFY "layouts"
(
cd $TESTDIR
@@ -1713,10 +1713,10 @@ TOOLTEST1 family tfamily%05d.h5
TOOLTEST bug1814 h5repack_refs.h5
# test attribute with various references (bug1797 / HDFFV-5932)
-# the references in attribute of compund or vlen datatype
+# the references in attribute of compound or vlen datatype
TOOLTEST HDFFV-5932 h5repack_attr_refs.h5
-# Add test for memory leak in attirbute. This test is verified by CTEST.
+# Add test for memory leak in attribute. This test is verified by CTEST.
# 1. leak from vlen string
# 2. leak from compound type without reference member
# (HDFFV-7840, )
diff --git a/tools/test/h5repack/h5repackgentest.c b/tools/test/h5repack/h5repackgentest.c
index a22af3d..c025891 100644
--- a/tools/test/h5repack/h5repackgentest.c
+++ b/tools/test/h5repack/h5repackgentest.c
@@ -27,9 +27,9 @@
* There is no restriction on the name, number, or structure of datasets and
* groups in HDF5 file.
*
- * The inluded datatypes should be more than adequate to verify the correctness
+ * The included datatypes should be more than adequate to verify the correctness
* of the behavior -- if one type can be consolidated from external storage,
- * then thay all can.
+ * then they all can.
*/
#include "hdf5.h"
diff --git a/tools/test/h5repack/h5repacktst.c b/tools/test/h5repack/h5repacktst.c
index 1627b1b..4b8ba56 100644
--- a/tools/test/h5repack/h5repacktst.c
+++ b/tools/test/h5repack/h5repacktst.c
@@ -624,7 +624,7 @@ main(void)
SKIPPED();
#endif
- TESTING(" addding shuffle filter");
+ TESTING(" adding shuffle filter");
/*-------------------------------------------------------------------------
* test an individual object option
@@ -653,7 +653,7 @@ main(void)
*-------------------------------------------------------------------------
*/
- TESTING(" addding shuffle filter to all");
+ TESTING(" adding shuffle filter to all");
if (h5repack_init(&pack_options, 0, FALSE) < 0)
GOERROR;
@@ -1592,17 +1592,17 @@ main(void)
#endif
/*-------------------------------------------------------------------------
- * test file with aligment
+ * test file with alignment
*-------------------------------------------------------------------------
*/
- TESTING(" file with aligment");
+ TESTING(" file with alignment");
#ifdef H5_HAVE_FILTER_DEFLATE
if (h5repack_init(&pack_options, 0, FALSE) < 0)
GOERROR;
- /* add the options for aligment */
+ /* add the options for alignment */
pack_options.alignment = 1;
pack_options.threshold = 1;
@@ -1613,7 +1613,7 @@ main(void)
if (h5repack_verify(FNAME8, FNAME8OUT, &pack_options) <= 0)
GOERROR;
- /* verify aligment */
+ /* verify alignment */
{
hsize_t threshold;
hsize_t alignment;
@@ -6244,7 +6244,7 @@ gen_refered_objs(hid_t loc_id)
goto out;
}
- /* create normal dataset which is refered */
+ /* create normal dataset which is referred */
did2 = H5Dcreate2(loc_id, NAME_OBJ_DS2, H5T_STD_I8LE, sid2, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT);
if (did2 < 0) {
HDfprintf(stderr, "Error: %s %d> H5Dcreate2 failed.\n", FUNC, __LINE__);
@@ -6284,7 +6284,7 @@ out:
* Generate object references to objects (dataset,group and named datatype)
*
* Note:
- * copied from h5copygentest.c and upate to create named datatype
+ * copied from h5copygentest.c and update to create named datatype
*
* Programmer: Jonathan Kim (March 18, 2010)
*------------------------------------------------------------------------*/
diff --git a/tools/test/h5repack/testfiles/README b/tools/test/h5repack/testfiles/README
index 4096dee..4cb66f4 100644
--- a/tools/test/h5repack/testfiles/README
+++ b/tools/test/h5repack/testfiles/README
@@ -1,5 +1,5 @@
h5repack_nested_8bit_enum_deflated.h5:
h5repack_nested_8bit_enum.h5:
- enuberated 8bit type nested in compount type. Original file provided
+ enuberated 8bit type nested in compound type. Original file provided
by a user (HDFFV-8667) as a test file. Used h5copy to extract only the
Compound type dataset. The non-deflated version is produced by h5repack.
diff --git a/tools/test/h5stat/testh5stat.sh.in b/tools/test/h5stat/testh5stat.sh.in
index 647f06a..1b4e504 100644
--- a/tools/test/h5stat/testh5stat.sh.in
+++ b/tools/test/h5stat/testh5stat.sh.in
@@ -13,7 +13,7 @@
#
# Tests for the h5stat tool
#
-# Modifcations:
+# Modifications:
# Vailin Choi; July 2013
# Add tests for -l, -m, -a options
#
diff --git a/tools/test/misc/h5clear_gentest.c b/tools/test/misc/h5clear_gentest.c
index 999611a..04a2379 100644
--- a/tools/test/misc/h5clear_gentest.c
+++ b/tools/test/misc/h5clear_gentest.c
@@ -16,7 +16,7 @@
/* The HDF5 test files */
const char *FILENAME[] = {
"h5clear_sec2_v3.h5", /* 0 -- sec2 file with superblock version 3 */
- "h5clear_log_v3.h5", /* 1 -- log file with superblock veresion 3 */
+ "h5clear_log_v3.h5", /* 1 -- log file with superblock version 3 */
"h5clear_sec2_v0.h5", /* 2 -- sec2 file with superblock version 0 */
"h5clear_sec2_v2.h5" /* 3 -- sec2 file with superblock version 2 */
};
diff --git a/tools/test/misc/h5perf_gentest.c b/tools/test/misc/h5perf_gentest.c
index 332e49c..248fd92 100644
--- a/tools/test/misc/h5perf_gentest.c
+++ b/tools/test/misc/h5perf_gentest.c
@@ -255,9 +255,9 @@ create_perf_test_file(const char *fname, int ngrps, int ndsets, int nattrs, hsiz
buf_vlen_s = (char **)calloc(dim0, sizeof(char *));
buf_str = malloc(dim0 * sizeof(*buf_str));
- /* allocate array of doulbe pointers */
+ /* allocate array of double pointers */
buf_double2d = (double **)calloc(dims2d[0], sizeof(double *));
- /* allocate a contigous chunk of memory for the data */
+ /* allocate a contiguous chunk of memory for the data */
buf_double2d[0] = (double *)calloc(dims2d[0] * dims2d[1], sizeof(double));
/* assign memory city to pointer array */
for (i = 1; i < dims2d[0]; i++)
@@ -396,7 +396,7 @@ create_perf_test_file(const char *fname, int ngrps, int ndsets, int nattrs, hsiz
H5Dclose(did);
/* 6 add compound data */
- HDsprintf(name, "%05d compund data", j);
+ HDsprintf(name, "%05d compound data", j);
did = H5Dcreate(gid1, name, cmp_tid, sid_1d, H5P_DEFAULT, dcpl, H5P_DEFAULT);
H5Dwrite(did, cmp_tid, H5S_ALL, H5S_ALL, H5P_DEFAULT, buf_comp);
if (!j)
diff --git a/tools/test/perform/direct_write_perf.c b/tools/test/perform/direct_write_perf.c
index 762cd61..6fd9688 100644
--- a/tools/test/perform/direct_write_perf.c
+++ b/tools/test/perform/direct_write_perf.c
@@ -630,7 +630,7 @@ test_unix_write(void)
TEST_ERROR;
}
else if (op_size == 0) {
- printf(" unable to write sufficent data to file because %s \n", strerror(errno));
+ printf(" unable to write sufficient data to file because %s \n", strerror(errno));
TEST_ERROR;
}
}
diff --git a/tools/test/perform/zip_perf.c b/tools/test/perform/zip_perf.c
index 4c7da15..040e774 100644
--- a/tools/test/perform/zip_perf.c
+++ b/tools/test/perform/zip_perf.c
@@ -267,7 +267,7 @@ usage(void)
HDfprintf(stdout, " data you want [default: 0]");
HDfprintf(stdout, " -s S, --file-size=S Maximum size of uncompressed file [default: 64M]\n");
HDfprintf(stdout, " -B S, --max-buffer_size=S Maximum size of buffer [default: 1M]\n");
- HDfprintf(stdout, " -b S, --min-buffer_size=S Minumum size of buffer [default: 128K]\n");
+ HDfprintf(stdout, " -b S, --min-buffer_size=S Minimum size of buffer [default: 128K]\n");
HDfprintf(stdout, " -p D, --prefix=D The directory prefix to place the file\n");
HDfprintf(stdout, " -r, --random-test Use random data to write to the file\n");
HDfprintf(stdout, " [default: no]\n");
diff --git a/utils/mirror_vfd/mirror_server.c b/utils/mirror_vfd/mirror_server.c
index d00e443..5381d95 100644
--- a/utils/mirror_vfd/mirror_server.c
+++ b/utils/mirror_vfd/mirror_server.c
@@ -80,7 +80,7 @@
* line.
*
* `log_prepend_type` (int)
- * Flag that the logging messages should have the assocaited verbosity
+ * Flag that the logging messages should have the associated verbosity
* level present in the line (e.g., "WARN", "ERROR", or "INFO").
*
* `log_path` (char *)
diff --git a/utils/mirror_vfd/mirror_server_stop.c b/utils/mirror_vfd/mirror_server_stop.c
index bf306d9..024b33a 100644
--- a/utils/mirror_vfd/mirror_server_stop.c
+++ b/utils/mirror_vfd/mirror_server_stop.c
@@ -15,7 +15,7 @@
* Exists for cross-platform, optionally remote shutdown.
*/
-#include "H5private.h" /* System compatability call-wrapper macros */
+#include "H5private.h" /* System compatibility call-wrapper macros */
#ifdef H5_HAVE_MIRROR_VFD
@@ -66,7 +66,7 @@ usage(void)
"\n"
"Options:\n"
" -h | --help Print this usage message and exit.\n"
- " --ip=ADDR IP Address of remote server (defaut %s)\n"
+ " --ip=ADDR IP Address of remote server (default %s)\n"
" --port=PORT Handshake port of remote server (default %d)\n",
MSHS_DEFAULT_IP, MSHS_DEFAULT_PORTNO);
} /* end usage() */