diff options
author | Allen Byrne <50328838+byrnHDF@users.noreply.github.com> | 2023-02-16 13:23:09 (GMT) |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-02-16 13:23:09 (GMT) |
commit | af9659d8475123d25f2235ef09b39567c4e2dde0 (patch) | |
tree | b94ef18a88c9c0a18f1cd027cf19f7d1487440da | |
parent | 5b79bb2f0f0c9722da2236deef2875291bdb7bde (diff) | |
download | hdf5-af9659d8475123d25f2235ef09b39567c4e2dde0.zip hdf5-af9659d8475123d25f2235ef09b39567c4e2dde0.tar.gz hdf5-af9659d8475123d25f2235ef09b39567c4e2dde0.tar.bz2 |
1.10: Add spelling check to CI and fix errors. (#2442)
* Update CI and support files to match latest from 1.14
* Correct spelling
* More spelling corrections
* spelling fixes in testpar
* Fix spelling errors in tools
* More tools spelling fixes
* Spelling fixes for rest of tools and some src
* Fix spelling errors in src files, pt 2
* Fix spelling in src pt3
* Fix spelling errors pt4
* Fix spelling errors pt5
* Spelling fix pt6
* fix spelling error examples
* fix spelling in tests
* fix spelling errors in test pt2
* Fix spelling errors in test pt3
* fix spelling in test pt4
* Fix spelling errors in hl
* fix spelling errors in c++
* Spelling fixes for fortran
* spelling fixes for bin and java
* Add relative path
* Change spelling action to use a file for list of ignore words
* Fix spelling ignore list
* remove unused file
367 files changed, 1345 insertions, 1305 deletions
diff --git a/.github/workflows/codespell.yml b/.github/workflows/codespell.yml new file mode 100644 index 0000000..2292a6a --- /dev/null +++ b/.github/workflows/codespell.yml @@ -0,0 +1,15 @@ +# GitHub Action to automate the identification of common misspellings in text files +# https://github.com/codespell-project/codespell +# https://github.com/codespell-project/actions-codespell +name: codespell +on: [push, pull_request] +jobs: + codespell: + name: Check for spelling errors + runs-on: ubuntu-latest + steps: + - uses: actions/checkout@v3 + - uses: codespell-project/actions-codespell@master + with: + skip: ./.github/workflows/codespell.yml,./bin/trace,./hl/tools/h5watch/h5watch.c,./tools/test/h5jam/tellub.c,./config/sanitizer/LICENSE,./tools/test/h5repack/testfiles/*.dat,./c++/src/*.html + ignore_words_list: ot,isnt,inout,nd,parms,parm,ba,offsetP,ser,ois,had,fiter,fo,clude,refere,minnum,offsetp,creat,eiter,lastr,ans,ifset,sur,trun,dne,tthe,hda,filname,te,htmp,minnum,ake,gord,numer,ro,oce diff --git a/.github/workflows/main.yml b/.github/workflows/main.yml index b60ee97..4a528bc 100644 --- a/.github/workflows/main.yml +++ b/.github/workflows/main.yml @@ -91,7 +91,7 @@ jobs: parallel: OFF mirror_vfd: ON direct_vfd: ON - toolchain: "config/toolchain/GCC.cmake" + toolchain: "config/toolchain/gcc.cmake" generator: "-G Ninja" run_tests: true diff --git a/CMakeLists.txt b/CMakeLists.txt index 2c4f88e..33b5922 100644 --- a/CMakeLists.txt +++ b/CMakeLists.txt @@ -12,6 +12,11 @@ if (POLICY CMP0083) cmake_policy (SET CMP0083 NEW) endif () +# Avoid warning about DOWNLOAD_EXTRACT_TIMESTAMP in CMake 3.24: +if (CMAKE_VERSION VERSION_GREATER_EQUAL "3.24.0") + cmake_policy(SET CMP0135 NEW) +endif() + #----------------------------------------------------------------------------- # Instructions for use : Normal Build # @@ -1,10 +1,10 @@ #!/bin/sh # -# Copyright by The HDF Group. -# All rights reserved. -# -# This file is part of HDF5. The full HDF5 copyright notice, including -# terms governing use, modification, and redistribution, is contained in +# Copyright by The HDF Group. +# All rights reserved. +# +# This file is part of HDF5. The full HDF5 copyright notice, including +# terms governing use, modification, and redistribution, is contained in # the COPYING file, which can be found at the root of the source code # distribution tree, or in https://www.hdfgroup.org/licenses. # If you do not have access to either file, you may request a copy from diff --git a/bin/batch/ctest.qsub.in.cmake b/bin/batch/ctest.qsub.in.cmake index 702fca7..da3ae2f 100644 --- a/bin/batch/ctest.qsub.in.cmake +++ b/bin/batch/ctest.qsub.in.cmake @@ -11,11 +11,13 @@ CTEST_CMD=`which ctest` cd @HDF5_BINARY_DIR@ if [[ $SUMMARY_FILE == *"ctestS"* ]]; then - CMD="${CTEST_CMD} . -E MPI_TEST_ -C Release -j 32 -T test" + CMD="${CTEST_CMD} -S ctest_serial.cmake" + qsub -t 60 -n 1 -q debug-flat-quad -A ${ACCOUNT_ID} ${CMD} >& ${SUMMARY_FILE} + echo "Done running ctest serial command." + touch ctestS.done else - CMD="${CTEST_CMD} . -R MPI_TEST_ ${SKIP_TESTS} -C Release -T test" + CMD="${CTEST_CMD} -S ctest_parallel.cmake" + qsub -t 60 -n 1 -q debug-flat-quad -A ${ACCOUNT_ID} ${CMD} >& ${SUMMARY_FILE} + echo "Done running ctest parallel command." + touch ctestP.done fi - -qsub -t 60 -n 1 -q debug-flat-quad -A ${ACCOUNT_ID} ${CMD} >& ${SUMMARY_FILE} - -echo "Done running ctest parallel command." diff --git a/bin/batch/ctestP.lsf.in.cmake b/bin/batch/ctestP.lsf.in.cmake index 3fdd5ce..8b89c9c 100644 --- a/bin/batch/ctestP.lsf.in.cmake +++ b/bin/batch/ctestP.lsf.in.cmake @@ -14,6 +14,7 @@ cd @HDF5_BINARY_DIR@ echo "Run parallel test command. Test output will be in build/ctestP.out" -ctest . -R MPI_TEST_ -C Release -T test >& ctestP.out +ctest -S ctest_parallel.cmake >& ctestP.out echo "Done running ctest parallel command." +touch ctestP.done diff --git a/bin/batch/ctestP.sl.in.cmake b/bin/batch/ctestP.sl.in.cmake index 1069fa9..599fee8 100644 --- a/bin/batch/ctestP.sl.in.cmake +++ b/bin/batch/ctestP.sl.in.cmake @@ -8,7 +8,8 @@ #SBATCH --job-name=h5_ctestP cd @HDF5_BINARY_DIR@ -ctest . -R MPI_TEST_ -C Release -T test >& ctestP.out - -echo "Done running ctestP.sl" +echo "Run parallel test command. Test output will be in build/ctestP.out" +ctest -S ctest_parallel.cmake >& ctestP.out +echo "Done running ctest parallel command." +touch ctestP.done diff --git a/bin/batch/ctestS.lsf.in.cmake b/bin/batch/ctestS.lsf.in.cmake index a01d39b..73b17c0 100644 --- a/bin/batch/ctestS.lsf.in.cmake +++ b/bin/batch/ctestS.lsf.in.cmake @@ -11,8 +11,7 @@ cd @HDF5_BINARY_DIR@ echo "Run command. Test output will be in build/ctestS.out" -ctest . -E MPI_TEST_ -C Release -j 32 -T test >& ctestS.out +ctest -S ctest_serial.cmake >& ctestS.out -##$CMD >& ctestS.out echo "Done running command." - +touch ctestS.done diff --git a/bin/batch/ctestS.sl.in.cmake b/bin/batch/ctestS.sl.in.cmake index 4f96bb9..bf0d6ae 100644 --- a/bin/batch/ctestS.sl.in.cmake +++ b/bin/batch/ctestS.sl.in.cmake @@ -8,8 +8,8 @@ #SBATCH --job-name=h5_ctestS cd @HDF5_BINARY_DIR@ -CMD="ctest . -E MPI_TEST_ -C Release -j 32 -T test" +echo "Run command. Test output will be in build/ctestS.out" +ctest -S ctest_serial.cmake >& ctestS.out -echo "Run $CMD. Test output will be in build/ctestS.out" -$CMD >& ctestS.out -echo "Done running $CMD" +echo "Done running command." +touch ctestS.done diff --git a/bin/batch/ctest_parallel.cmake.in b/bin/batch/ctest_parallel.cmake.in new file mode 100644 index 0000000..f35a772 --- /dev/null +++ b/bin/batch/ctest_parallel.cmake.in @@ -0,0 +1,12 @@ +if(NOT "$ENV{CI_SITE_NAME}" STREQUAL "") + set(CTEST_SITE "$ENV{CI_SITE_NAME}") +endif() +if(NOT "$ENV{CI_BUILD_NAME}" STREQUAL "") + set(CTEST_BUILD_NAME "$ENV{CI_BUILD_NAME}") +endif() + +ctest_start ("$ENV{CI_MODEL}" "@HDF5_SOURCE_DIR@" "@HDF5_BINARY_DIR@" APPEND) +ctest_test (BUILD "@HDF5_BINARY_DIR@" APPEND INCLUDE MPI_TEST_ RETURN_VALUE res) +if (${res} LESS 0 OR ${res} GREATER 0) + file (APPEND ${CTEST_SCRIPT_DIRECTORY}/FailedCTest.txt "Failed Tests: ${res}\n") + endif () diff --git a/bin/batch/ctest_serial.cmake.in b/bin/batch/ctest_serial.cmake.in new file mode 100644 index 0000000..e54421b --- /dev/null +++ b/bin/batch/ctest_serial.cmake.in @@ -0,0 +1,12 @@ +if(NOT "$ENV{CI_SITE_NAME}" STREQUAL "") + set(CTEST_SITE "$ENV{CI_SITE_NAME}") +endif() +if(NOT "$ENV{CI_BUILD_NAME}" STREQUAL "") + set(CTEST_BUILD_NAME "$ENV{CI_BUILD_NAME}") +endif() + +ctest_start ("$ENV{CI_MODEL}" "@HDF5_SOURCE_DIR@" "@HDF5_BINARY_DIR@" APPEND) +ctest_test (BUILD "@HDF5_BINARY_DIR@" APPEND EXCLUDE MPI_TEST_ PARALLEL_LEVEL 32 RETURN_VALUE res) +if (${res} LESS 0 OR ${res} GREATER 0) + file (APPEND ${CTEST_SCRIPT_DIRECTORY}/FailedCTest.txt "Failed Tests: ${res}\n") + endif () diff --git a/bin/batch/knl_ctestP.sl.in.cmake b/bin/batch/knl_ctestP.sl.in.cmake index f985fbb..97bd2ad 100644 --- a/bin/batch/knl_ctestP.sl.in.cmake +++ b/bin/batch/knl_ctestP.sl.in.cmake @@ -9,8 +9,8 @@ #SBATCH --job-name=h5_ctestP cd @HDF5_BINARY_DIR@ -#run parallel tests except t_cache_image test -ctest . -R MPI_TEST_ -C Release -T test >& ctestP.out - -echo "Done running $CMD" +echo "Run parallel test command. Test output will be in build/ctestP.out" +ctest -S ctest_parallel.cmake >& ctestP.out +echo "Done running ctest parallel command." +touch ctestP.done diff --git a/bin/batch/knl_ctestS.sl.in.cmake b/bin/batch/knl_ctestS.sl.in.cmake index af6353b..87c4a48 100644 --- a/bin/batch/knl_ctestS.sl.in.cmake +++ b/bin/batch/knl_ctestS.sl.in.cmake @@ -9,9 +9,8 @@ #SBATCH --job-name=h5_ctestS cd @HDF5_BINARY_DIR@ -CMD="ctest . -E MPI_TEST_ -C Release -j 32 -T test" - -echo "Run $CMD. Test output will be in build/ctestS.out" -$CMD >& ctestS.out -echo "Done running $CMD" +echo "Run command. Test output will be in build/ctestS.out" +ctest -S ctest_serial.cmake >& ctestS.out +echo "Done running command." +touch ctestS.done diff --git a/bin/batch/ray_ctestP.lsf.in.cmake b/bin/batch/ray_ctestP.lsf.in.cmake index 7067a65..51425b5 100644 --- a/bin/batch/ray_ctestP.lsf.in.cmake +++ b/bin/batch/ray_ctestP.lsf.in.cmake @@ -15,6 +15,8 @@ cd @HDF5_BINARY_DIR@ echo "Run parallel test command. Test output will be in build/ctestP.out" -ctest . -R 'MPI_TEST_' -C Release -T test >& ctestP.out +ctest -S ctest_parallel.cmake >& ctestP.out echo "Done running ctest parallel command." +touch ctestP.done +~ diff --git a/bin/batch/ray_ctestS.lsf.in.cmake b/bin/batch/ray_ctestS.lsf.in.cmake index da20438..7f585b4 100644 --- a/bin/batch/ray_ctestS.lsf.in.cmake +++ b/bin/batch/ray_ctestS.lsf.in.cmake @@ -11,8 +11,7 @@ cd @HDF5_BINARY_DIR@ echo "Run command. Test output will be in build/ctestS.out" -ctest . -E 'MPI_TEST_' -C Release -j 32 -T test >& ctestS.out +ctest -S ctest_serial.cmake >& ctestS.out -##$CMD >& ctestS.out echo "Done running command." - +touch ctestS.done diff --git a/bin/warnhist b/bin/warnhist index 0150138..669cbcc 100755 --- a/bin/warnhist +++ b/bin/warnhist @@ -382,7 +382,7 @@ while (<>) { $warning =~ s/"[A-Za-z_0-9]*"/"-"/g; } - # Genericize [GCC?] C/C++ warning text about suggessted attribute + # Genericize [GCC?] C/C++ warning text about suggested attribute if($warning =~ /attribute=[A-Za-z_0-9]*\]/) { $warning =~ s/=[A-Za-z_0-9]*\]/=-\]/g; } diff --git a/c++/src/H5AbstractDs.cpp b/c++/src/H5AbstractDs.cpp index eeb0155..70d8531 100644 --- a/c++/src/H5AbstractDs.cpp +++ b/c++/src/H5AbstractDs.cpp @@ -302,7 +302,7 @@ AbstractDs::getStrType() const //-------------------------------------------------------------------------- // Function: AbstractDs::getVarLenType -///\brief Returns the floating-point datatype of this abstract dataset, +///\brief Returns the variable length datatype of this abstract dataset, /// which can be a dataset or an attribute. ///\return VarLenType instance ///\exception H5::DataTypeIException diff --git a/c++/src/H5Attribute.cpp b/c++/src/H5Attribute.cpp index 865e97c..06285bc 100644 --- a/c++/src/H5Attribute.cpp +++ b/c++/src/H5Attribute.cpp @@ -562,7 +562,7 @@ Attribute::p_read_variable_len(const DataType &mem_type, H5std_string &strg) con ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description: -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. // Programmer Binh-Minh Ribler - 2000 diff --git a/c++/src/H5DataSet.cpp b/c++/src/H5DataSet.cpp index b376e57..984713a 100644 --- a/c++/src/H5DataSet.cpp +++ b/c++/src/H5DataSet.cpp @@ -770,7 +770,7 @@ DataSet::p_read_variable_len(const hid_t mem_type_id, const hid_t mem_space_id, ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description: -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. //-------------------------------------------------------------------------- diff --git a/c++/src/H5DataSpace.cpp b/c++/src/H5DataSpace.cpp index 0485530..be64a17 100644 --- a/c++/src/H5DataSpace.cpp +++ b/c++/src/H5DataSpace.cpp @@ -631,7 +631,7 @@ DataSpace::getId() const ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description: -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. // Programmer Binh-Minh Ribler - 2000 diff --git a/c++/src/H5DataType.cpp b/c++/src/H5DataType.cpp index 3228dcb..07a1463 100644 --- a/c++/src/H5DataType.cpp +++ b/c++/src/H5DataType.cpp @@ -932,7 +932,7 @@ DataType::p_opentype(const H5Location &loc, const char *dtype_name) const ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description: -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. // Programmer Binh-Minh Ribler - 2000 @@ -988,7 +988,7 @@ DataType::close() // - Replaced decRefCount with close() to let the C library // handle the reference counting - BMR, Jun 1, 2006 // - Added the use of H5CPP_EXITED to terminate the HDF5 library -// and elimiate previous memory leaks. See comments in the +// and eliminate previous memory leaks. See comments in the // header file "H5PredType.h" for details. - BMR, Mar 30, 2012 // - Major re-implementation of the global constants was done // to avoid relying on the order of the creation and deletion diff --git a/c++/src/H5DcreatProp.cpp b/c++/src/H5DcreatProp.cpp index 4c71d5e..5d07943 100644 --- a/c++/src/H5DcreatProp.cpp +++ b/c++/src/H5DcreatProp.cpp @@ -553,7 +553,7 @@ DSetCreatPropList::setShuffle() const ///\exception H5::PropListIException ///\par Description /// The values of space allocation time can be one of the -/// followings: +/// following: /// \li \c H5D_ALLOC_TIME_DEFAULT /// \li \c H5D_ALLOC_TIME_EARLY /// \li \c H5D_ALLOC_TIME_LATE diff --git a/c++/src/H5FaccProp.cpp b/c++/src/H5FaccProp.cpp index 3eaf23a..e0118e0 100644 --- a/c++/src/H5FaccProp.cpp +++ b/c++/src/H5FaccProp.cpp @@ -157,7 +157,7 @@ FileAccPropList::getDriver() const // Function: FileAccPropList::setDriver ///\brief Set file driver for this property list. ///\param new_driver_id - IN: File driver -///\param new_driver_info - IN: Struct containing the driver-specific properites +///\param new_driver_info - IN: Struct containing the driver-specific properties ///\exception H5::PropListIException ///\par Description /// For information, please refer to the H5Pset_driver API in diff --git a/c++/src/H5File.cpp b/c++/src/H5File.cpp index c7f0f95..b5f6c40 100644 --- a/c++/src/H5File.cpp +++ b/c++/src/H5File.cpp @@ -603,7 +603,7 @@ H5File::getLocId() const ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. // December 2000 diff --git a/c++/src/H5Group.cpp b/c++/src/H5Group.cpp index fe791da..b017f4f 100644 --- a/c++/src/H5Group.cpp +++ b/c++/src/H5Group.cpp @@ -208,7 +208,7 @@ Group::getId() const ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description: -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. // Programmer Binh-Minh Ribler - 2000 diff --git a/c++/src/H5IntType.h b/c++/src/H5IntType.h index 185c8f0..ff7bb64 100644 --- a/c++/src/H5IntType.h +++ b/c++/src/H5IntType.h @@ -41,7 +41,7 @@ class H5_DLLCPP IntType : public AtomType { // Retrieves the sign type for an integer type H5T_sign_t getSign() const; - // Sets the sign proprety for an integer type. + // Sets the sign property for an integer type. void setSign(H5T_sign_t sign) const; ///\brief Returns this class name. diff --git a/c++/src/H5Location.cpp b/c++/src/H5Location.cpp index c1204ab..6277860 100644 --- a/c++/src/H5Location.cpp +++ b/c++/src/H5Location.cpp @@ -1626,7 +1626,7 @@ H5Location::getObjinfo(const H5std_string &name, hbool_t follow_link, H5G_stat_t // Function: H5Location::getObjinfo ///\brief This is an overloaded member function, provided for convenience. /// It differs from the above functions in that it doesn't have -/// the paramemter \a follow_link. +/// the parameter \a follow_link. // Nov, 2005 //-------------------------------------------------------------------------- void diff --git a/c++/src/H5Location.h b/c++/src/H5Location.h index 67f14d5..d2fe3f9 100644 --- a/c++/src/H5Location.h +++ b/c++/src/H5Location.h @@ -15,7 +15,7 @@ #ifndef H5Location_H #define H5Location_H -#include "H5Classes.h" // constains forward class declarations +#include "H5Classes.h" // contains forward class declarations namespace H5 { diff --git a/c++/src/H5PredType.cpp b/c++/src/H5PredType.cpp index 9e67968..86d721f 100644 --- a/c++/src/H5PredType.cpp +++ b/c++/src/H5PredType.cpp @@ -916,7 +916,7 @@ September 2015: classes, such as PropList, PredType, DataSpace, etc... Previously, these global constants were declared statically and the C++ library used a constant, called PredType::AtExit, to detect when all the global - contants are destroyed then close the C library (H5close). This method + constants are destroyed then close the C library (H5close). This method relied on the order of the constants being created and destroyed and that PredType constants be the last to be destroyed. In September 2015, it was recognized that the order in which the global constants were diff --git a/c++/src/H5PropList.cpp b/c++/src/H5PropList.cpp index f2c1d70..9757f8c 100644 --- a/c++/src/H5PropList.cpp +++ b/c++/src/H5PropList.cpp @@ -303,7 +303,7 @@ PropList::getId() const ///\exception H5::IdComponentException when the attempt to close the HDF5 /// object fails // Description: -// The underlaying reference counting in the C library ensures +// The underlying reference counting in the C library ensures // that the current valid id of this object is properly closed. // Then the object's id is reset to the new id. // Programmer Binh-Minh Ribler - 2000 diff --git a/c++/test/tobject.cpp b/c++/test/tobject.cpp index cb12573..96389bf 100644 --- a/c++/test/tobject.cpp +++ b/c++/test/tobject.cpp @@ -732,7 +732,7 @@ test_object() MESSAGE(5, ("Testing Object Functions\n")); test_get_objname(); // Test get object name from groups/datasets - test_existance(); // Test check for object existance + test_existance(); // Test check for object existence test_get_objname_ontypes(); // Test get object name from types test_get_objtype(); // Test get object type test_open_object_header(); // Test object header functions (H5O) diff --git a/c++/test/trefer.cpp b/c++/test/trefer.cpp index f4eab6f..36e58b4 100644 --- a/c++/test/trefer.cpp +++ b/c++/test/trefer.cpp @@ -446,7 +446,7 @@ test_reference_group() file1->close(); /* - * Re-open the file and test deferencing group + * Re-open the file and test dereferencing group */ // Re-open file diff --git a/c++/test/ttypes.cpp b/c++/test/ttypes.cpp index e50f61e..5241efc 100644 --- a/c++/test/ttypes.cpp +++ b/c++/test/ttypes.cpp @@ -28,7 +28,7 @@ using namespace H5; #include "h5cpputil.h" // C++ utilility header file /* - * Offset from alinged memory returned by malloc(). This can be used to test + * Offset from aligned memory returned by malloc(). This can be used to test * that type conversions handle non-aligned buffers correctly. */ #define ALIGNMENT 1 diff --git a/config/cmake/HDFFortranCompilerFlags.cmake b/config/cmake/HDFFortranCompilerFlags.cmake index 84b3ebe..599ec30 100644 --- a/config/cmake/HDFFortranCompilerFlags.cmake +++ b/config/cmake/HDFFortranCompilerFlags.cmake @@ -62,7 +62,7 @@ if (NOT MSVC AND NOT MINGW) # General flags if (CMAKE_Fortran_COMPILER_ID STREQUAL "Intel") ADD_H5_FLAGS (HDF5_CMAKE_Fortran_FLAGS "${HDF5_SOURCE_DIR}/config/intel-warnings/ifort-general") - list (APPEND HDF5_CMAKE_Fortran_FLAGS "-stand f03" "-free") + list (APPEND HDF5_CMAKE_Fortran_FLAGS "-stand:f03" "-free") elseif (CMAKE_Fortran_COMPILER_ID STREQUAL "GNU") ADD_H5_FLAGS (HDF5_CMAKE_Fortran_FLAGS "${HDF5_SOURCE_DIR}/config/gnu-warnings/gfort-general") list (APPEND HDF5_CMAKE_Fortran_FLAGS "-ffree-form" "-fimplicit-none") diff --git a/config/cmake_ext_mod/HDFUseCXX.cmake b/config/cmake_ext_mod/HDFUseCXX.cmake index fcd3ba6..6788773 100644 --- a/config/cmake_ext_mod/HDFUseCXX.cmake +++ b/config/cmake_ext_mod/HDFUseCXX.cmake @@ -17,7 +17,7 @@ ENABLE_LANGUAGE (CXX) set (HDF_PREFIX "H5") #------------------------------------------------------------------------------- -# Fix CXX flags if we are compiling staticly on Windows using +# Fix CXX flags if we are compiling statically on Windows using # Windows_MT.cmake from config/cmake/UserMacros #------------------------------------------------------------------------------- if (BUILD_STATIC_CRT_LIBS) diff --git a/config/ibm-aix b/config/ibm-aix index f48f9d8..8720db2 100644 --- a/config/ibm-aix +++ b/config/ibm-aix @@ -58,7 +58,7 @@ if test "X-" = "X-$f9x_flags_set"; then H5_FCFLAGS="$H5_FCFLAGS ${F9XSUFFIXFLAG}" FSEARCH_DIRS="-I./ -I../src" - # Produciton + # Production PROD_FCFLAGS= # Debug @@ -109,7 +109,7 @@ case $CC_BASENAME in # Use very generic flags H5_CFLAGS="$H5_CFLAGS -ansi" - # Produciton + # Production PROD_CFLAGS= # Debug diff --git a/config/ibm-flags b/config/ibm-flags index ab4b8ff..f3f247c 100644 --- a/config/ibm-flags +++ b/config/ibm-flags @@ -62,7 +62,7 @@ if test "XL" = "$cc_vendor"; then # concerning the portability of __inline__. H5_CFLAGS="-qlanglvl=stdc99 -qflag=w:w $H5_CFLAGS" - # Produciton + # Production PROD_CFLAGS= # Debug diff --git a/config/toolchain/aarch64.cmake b/config/toolchain/aarch64.cmake new file mode 100644 index 0000000..adb8639 --- /dev/null +++ b/config/toolchain/aarch64.cmake @@ -0,0 +1,17 @@ +set(TOOLCHAIN_PREFIX aarch64-linux-gnu) +set(ANDROID_NDK /opt/android-ndk-r25b-linux/android-ndk-r25b) +set (CMAKE_SYSTEM_NAME Android) +set (CMAKE_ANDROID_ARCH_ABI x86_64) +#set (CMAKE_ANDROID_STANDALONE_TOOLCHAIN ${ANDROID_NDK}/build/cmake/andriod.toolchain.cmake) +set (CMAKE_C_COMPILER ${TOOLCHAIN_PREFIX}-gcc) +set (CMAKE_CXX_COMPILER ${TOOLCHAIN_PREFIX}-g++) +#set (CMAKE_RC_COMPILER ${TOOLCHAIN_PREFIX}-windres) +set (CMAKE_Fortran_COMPILER ${TOOLCHAIN_PREFIX}-gfortran) +set (CMAKE_FIND_ROOT_PATH /usr/${TOOLCHAIN_PREFIX}) +set (CMAKE_FIND_ROOT_PATH_MODE_PROGRAM NEVER) +set (CMAKE_FIND_ROOT_PATH_MODE_LIBRARY ONLY) +set (CMAKE_FIND_ROOT_PATH_MODE_INCLUDE ONLY) +set (CMAKE_CROSSCOMPILING_EMULATOR qemu-aarch64) + +include_directories(/usr/${TOOLCHAIN_PREFIX}/include) + diff --git a/config/toolchain/GCC.cmake b/config/toolchain/gcc.cmake index c41d0ca..c41d0ca 100644 --- a/config/toolchain/GCC.cmake +++ b/config/toolchain/gcc.cmake diff --git a/config/toolchain/intel.cmake b/config/toolchain/intel.cmake index ae1d2f8..f8f60b2 100644 --- a/config/toolchain/intel.cmake +++ b/config/toolchain/intel.cmake @@ -3,16 +3,9 @@ set(CMAKE_COMPILER_VENDOR "intel") -if(USE_SANITIZER) - set(CMAKE_C_COMPILER icl) - set(CMAKE_CXX_COMPILER icl++) - set(CMAKE_Fortran_COMPILER ifort) - set(INTEL_CLANG ON) -else () - set(CMAKE_C_COMPILER icc) - set(CMAKE_CXX_COMPILER icpc) - set(CMAKE_Fortran_COMPILER ifort) -endif () +set(CMAKE_C_COMPILER icx) +set(CMAKE_CXX_COMPILER icpx) +set(CMAKE_Fortran_COMPILER ifx) # the following is used if cross-compiling set(CMAKE_CROSSCOMPILING_EMULATOR "") diff --git a/config/toolchain/intel_classic.cmake b/config/toolchain/intel_classic.cmake new file mode 100644 index 0000000..97f6a64 --- /dev/null +++ b/config/toolchain/intel_classic.cmake @@ -0,0 +1,11 @@ +# Uncomment the following to use cross-compiling +#set(CMAKE_SYSTEM_NAME Linux) + +set(CMAKE_COMPILER_VENDOR "intel") + +set(CMAKE_C_COMPILER icc) +set(CMAKE_CXX_COMPILER icpc) +set(CMAKE_Fortran_COMPILER ifort) + +# the following is used if cross-compiling +set(CMAKE_CROSSCOMPILING_EMULATOR "") diff --git a/config/toolchain/PGI.cmake b/config/toolchain/pgi.cmake index ec58cbb..ec58cbb 100644 --- a/config/toolchain/PGI.cmake +++ b/config/toolchain/pgi.cmake diff --git a/examples/h5_debug_trace.c b/examples/h5_debug_trace.c index 0f0ec97..b7f6e5b 100644 --- a/examples/h5_debug_trace.c +++ b/examples/h5_debug_trace.c @@ -5,12 +5,10 @@ * * * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * - * the files COPYING and Copyright.html. COPYING can be found at the root * - * of the source code distribution tree; Copyright.html can be found at the * - * root level of an installed copy of the electronic HDF5 document set and * - * is linked from the top-level documents page. It can also be found at * - * http://hdfgroup.org/HDF5/doc/Copyright.html. If you do not have * - * access to either file, you may request a copy from help@hdfgroup.org. * + * the COPYING file, which can be found at the root of the source code * + * distribution tree, or in https://www.hdfgroup.org/licenses. * + * If you do not have access to either file, you may request a copy from * + * help@hdfgroup.org. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ /* This example demonstrates debug trace output. diff --git a/examples/h5_vds-eiger.c b/examples/h5_vds-eiger.c index 9608319..fcde490 100644 --- a/examples/h5_vds-eiger.c +++ b/examples/h5_vds-eiger.c @@ -128,7 +128,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset \n"); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); if (H5Sget_select_type(vspace) == H5S_SEL_HYPERSLABS) { if (H5Sis_regular_hyperslab(vspace)) { diff --git a/examples/h5_vds-exc.c b/examples/h5_vds-exc.c index e03ffd8..fa6b720 100644 --- a/examples/h5_vds-exc.c +++ b/examples/h5_vds-exc.c @@ -156,7 +156,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset \n"); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); if (H5Sget_select_type(vspace) == H5S_SEL_HYPERSLABS) { if (H5Sis_regular_hyperslab(vspace)) { diff --git a/examples/h5_vds-exclim.c b/examples/h5_vds-exclim.c index f7b3fdd..58ed1f8 100644 --- a/examples/h5_vds-exclim.c +++ b/examples/h5_vds-exclim.c @@ -153,7 +153,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset \n"); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); if (H5Sget_select_type(vspace) == H5S_SEL_HYPERSLABS) { if (H5Sis_regular_hyperslab(vspace)) { diff --git a/examples/h5_vds-percival-unlim-maxmin.c b/examples/h5_vds-percival-unlim-maxmin.c index 507f088..9ef514d 100644 --- a/examples/h5_vds-percival-unlim-maxmin.c +++ b/examples/h5_vds-percival-unlim-maxmin.c @@ -246,7 +246,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset \n"); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); if (H5Sget_select_type(vspace) == H5S_SEL_HYPERSLABS) { if (H5Sis_regular_hyperslab(vspace)) { diff --git a/examples/h5_vds-percival-unlim.c b/examples/h5_vds-percival-unlim.c index badead3..ddbcdec 100644 --- a/examples/h5_vds-percival-unlim.c +++ b/examples/h5_vds-percival-unlim.c @@ -223,7 +223,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset \n"); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); if (H5Sget_select_type(vspace) == H5S_SEL_HYPERSLABS) { if (H5Sis_regular_hyperslab(vspace)) { diff --git a/examples/h5_vds-percival.c b/examples/h5_vds-percival.c index 9507931..82c8ef4 100644 --- a/examples/h5_vds-percival.c +++ b/examples/h5_vds-percival.c @@ -183,7 +183,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset \n"); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); if (H5Sget_select_type(vspace) == H5S_SEL_HYPERSLABS) { if (H5Sis_regular_hyperslab(vspace)) { diff --git a/examples/h5_vds-simpleIO.c b/examples/h5_vds-simpleIO.c index cb69645..f516af9 100644 --- a/examples/h5_vds-simpleIO.c +++ b/examples/h5_vds-simpleIO.c @@ -132,7 +132,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset "); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); /* Make sure it is ALL selection and then print selection. */ diff --git a/examples/h5_vds.c b/examples/h5_vds.c index 48c6d7d..e02d246 100644 --- a/examples/h5_vds.c +++ b/examples/h5_vds.c @@ -169,7 +169,7 @@ main(void) for (i = 0; i < (int)num_map; i++) { printf(" Mapping %d \n", i); printf(" Selection in the virtual dataset "); - /* Get selection in the virttual dataset */ + /* Get selection in the virtual dataset */ vspace = H5Pget_virtual_vspace(dcpl, (size_t)i); /* Make sure that this is a hyperslab selection and then print information. */ diff --git a/fortran/examples/compound_complex_fortran2003.f90 b/fortran/examples/compound_complex_fortran2003.f90 index 2a86631..cf1665fb 100644 --- a/fortran/examples/compound_complex_fortran2003.f90 +++ b/fortran/examples/compound_complex_fortran2003.f90 @@ -1,14 +1,15 @@ ! This is the F2003 version of the h5_compound.c example source code. ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * +! Copyright by The HDF Group. * ! Copyright by the Board of Trustees of the University of Illinois. * ! All rights reserved. * ! * ! This file is part of HDF5. The full HDF5 copyright notice, including * ! terms governing use, modification, and redistribution, is contained in * -! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * -! If you do not have access to either file, you may request a copy from * -! help@hdfgroup.org. * +! the COPYING file, which can be found at the root of the source code * +! distribution tree, or in https://www.hdfgroup.org/licenses. * +! If you do not have access to either file, you may request a copy from * +! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * ! ! This example shows how to create an array of a compound datatype which diff --git a/fortran/examples/compound_fortran2003.f90 b/fortran/examples/compound_fortran2003.f90 index 1a05e1d..d6202be 100644 --- a/fortran/examples/compound_fortran2003.f90 +++ b/fortran/examples/compound_fortran2003.f90 @@ -1,14 +1,15 @@ ! This is the F2003 version of the h5_compound.c example source code. ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * +! Copyright by The HDF Group. * ! Copyright by the Board of Trustees of the University of Illinois. * ! All rights reserved. * ! * ! This file is part of HDF5. The full HDF5 copyright notice, including * ! terms governing use, modification, and redistribution, is contained in * -! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * -! If you do not have access to either file, you may request a copy from * -! help@hdfgroup.org. * +! the COPYING file, which can be found at the root of the source code * +! distribution tree, or in https://www.hdfgroup.org/licenses. * +! If you do not have access to either file, you may request a copy from * +! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * ! ! This example shows how to create a compound data type, diff --git a/fortran/examples/h5_subset.f90 b/fortran/examples/h5_subset.f90 index fa9e3a8..f08b97b 100644 --- a/fortran/examples/h5_subset.f90 +++ b/fortran/examples/h5_subset.f90 @@ -1,14 +1,14 @@ ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * -! Copyright by The HDF Group. * -! Copyright by the Board of Trustees of the University of Illinois. * -! All rights reserved. * +! Copyright by The HDF Group. * +! Copyright by the Board of Trustees of the University of Illinois. * +! All rights reserved. * ! * -! This file is part of HDF5. The full HDF5 copyright notice, including * -! terms governing use, modification, and redistribution, is contained in * -! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * -! If you do not have access to either file, you may request a copy from * -! help@hdfgroup.org. * +! This file is part of HDF5. The full HDF5 copyright notice, including * +! terms governing use, modification, and redistribution, is contained in * +! the COPYING file, which can be found at the root of the source code * +! distribution tree, or in https://www.hdfgroup.org/licenses. * +! If you do not have access to either file, you may request a copy from * +! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * ! ! This example shows how to write and read a hyperslab. diff --git a/fortran/examples/nested_derived_type.f90 b/fortran/examples/nested_derived_type.f90 index ecf1481..38f3854 100644 --- a/fortran/examples/nested_derived_type.f90 +++ b/fortran/examples/nested_derived_type.f90 @@ -1,14 +1,15 @@ ! ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * +! Copyright by The HDF Group. * ! Copyright by the Board of Trustees of the University of Illinois. * ! All rights reserved. * ! * ! This file is part of HDF5. The full HDF5 copyright notice, including * ! terms governing use, modification, and redistribution, is contained in * -! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * -! If you do not have access to either file, you may request a copy from * -! help@hdfgroup.org. * +! the COPYING file, which can be found at the root of the source code * +! distribution tree, or in https://www.hdfgroup.org/licenses. * +! If you do not have access to either file, you may request a copy from * +! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * ! ! This example shows how to create a nested compound data type, diff --git a/fortran/examples/refobjexample.f90 b/fortran/examples/refobjexample.f90 index 873b420..0d6595c 100644 --- a/fortran/examples/refobjexample.f90 +++ b/fortran/examples/refobjexample.f90 @@ -65,7 +65,7 @@ ! CALL h5gcreate_f(file_id, groupname1, grp1_id, error) ! - ! Create a group inside the created gorup + ! Create a group inside the created group ! CALL h5gcreate_f(grp1_id, groupname2, grp2_id, error) ! diff --git a/fortran/src/H5Iff.F90 b/fortran/src/H5Iff.F90 index b67d807..aa441d3 100644 --- a/fortran/src/H5Iff.F90 +++ b/fortran/src/H5Iff.F90 @@ -11,12 +11,12 @@ ! ! COPYRIGHT ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * -! Copyright by The HDF Group. * -! Copyright by the Board of Trustees of the University of Illinois. * -! All rights reserved. * -! * -! This file is part of HDF5. The full HDF5 copyright notice, including * -! terms governing use, modification, and redistribution, is contained in * +! Copyright by The HDF Group. * +! Copyright by the Board of Trustees of the University of Illinois. * +! All rights reserved. * +! * +! This file is part of HDF5. The full HDF5 copyright notice, including * +! terms governing use, modification, and redistribution, is contained in * ! the COPYING file, which can be found at the root of the source code * ! distribution tree, or in https://www.hdfgroup.org/licenses. * ! If you do not have access to either file, you may request a copy from * diff --git a/fortran/src/H5Sff.F90 b/fortran/src/H5Sff.F90 index 8532694..0e606e6 100644 --- a/fortran/src/H5Sff.F90 +++ b/fortran/src/H5Sff.F90 @@ -443,7 +443,7 @@ CONTAINS !! !! \brief Verifies that the selection is within the extent of the dataspace. !! -!! \param space_id Identifier for the dataspace for whichselection is verified +!! \param space_id Identifier for the dataspace for which selection is verified !! \param status TRUE if the selection is contained within the extent, FALSE otherwise. !! \param hdferr \fortran_error !! diff --git a/fortran/src/H5Tf.c b/fortran/src/H5Tf.c index d9f75d9..eee0020 100644 --- a/fortran/src/H5Tf.c +++ b/fortran/src/H5Tf.c @@ -887,7 +887,7 @@ h5tget_ebias_c(hid_t_f *type_id, size_t_f *ebias) * floating-point type of the datatype * INPUTS * type_id - identifier of the dataspace - * ebias - exponent bias of a floating-point type of the datatyp + * ebias - exponent bias of a floating-point type of the datatype * RETURNS * 0 on success, -1 on failure * AUTHOR diff --git a/fortran/src/H5_buildiface.F90 b/fortran/src/H5_buildiface.F90 index 30d29ba..2af242d 100644 --- a/fortran/src/H5_buildiface.F90 +++ b/fortran/src/H5_buildiface.F90 @@ -34,7 +34,7 @@ ! This file is part of HDF5. The full HDF5 copyright notice, including * ! terms governing use, modification, and redistribution, is contained in * ! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * +! distribution tree, or in https://www.hdfgroup.org/licenses. * ! If you do not have access to either file, you may request a copy from * ! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * diff --git a/fortran/src/H5f90global.F90 b/fortran/src/H5f90global.F90 index 840b8a7..d71a50c 100644 --- a/fortran/src/H5f90global.F90 +++ b/fortran/src/H5f90global.F90 @@ -18,7 +18,7 @@ ! then the PARAMETER of that corresponding array in Fortran must also be changed. ! ! USES -! H5FORTRAN_TYPES - This module is generated at run time. See +! H5FORTRAN_TYPES - H5FORTRAN_TYPES is generated at run time by H5match_types.c ! ! COPYRIGHT ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * diff --git a/fortran/test/H5_test_buildiface.F90 b/fortran/test/H5_test_buildiface.F90 index 94beab1..a53cbe4 100644 --- a/fortran/test/H5_test_buildiface.F90 +++ b/fortran/test/H5_test_buildiface.F90 @@ -14,7 +14,7 @@ ! ! NOTES ! This program uses the Fortran 2008 intrinsic function STORAGE_SIZE or SIZEOF -! depending on availablity.It generates code that makes use of +! depending on availability.It generates code that makes use of ! STORAGE_SIZE/SIZEOF in H5fortran_detect.f90. STORAGE_SIZE is standard ! compliant and should always be chosen over SIZEOF. ! @@ -34,7 +34,7 @@ ! This file is part of HDF5. The full HDF5 copyright notice, including * ! terms governing use, modification, and redistribution, is contained in * ! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * +! distribution tree, or in https://www.hdfgroup.org/licenses. * ! If you do not have access to either file, you may request a copy from * ! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * diff --git a/fortran/test/fflush1.F90 b/fortran/test/fflush1.F90 index c192d93..5684f81 100644 --- a/fortran/test/fflush1.F90 +++ b/fortran/test/fflush1.F90 @@ -88,7 +88,7 @@ data_dims(2) = NY ! - !Initialize FORTRAN predifined datatypes + !Initialize FORTRAN predefined datatypes ! CALL h5open_f(error) CALL check("h5open_f",error,total_error) diff --git a/fortran/test/fflush2.F90 b/fortran/test/fflush2.F90 index dc3bc85..a6e16c1 100644 --- a/fortran/test/fflush2.F90 +++ b/fortran/test/fflush2.F90 @@ -81,7 +81,7 @@ PROGRAM FFLUSH2EXAMPLE data_dims(2) = NY ! - !Initialize FORTRAN predifined datatypes + !Initialize FORTRAN predefined datatypes ! CALL h5open_f(error) CALL check("h5open_f",error,total_error) @@ -163,7 +163,7 @@ PROGRAM FFLUSH2EXAMPLE CALL check("h5fclose_f",error,total_error) ! - !Close FORTRAN predifined datatypes + !Close FORTRAN predefined datatypes ! CALL h5_cleanup_f(filename, H5P_DEFAULT_F, error) CALL h5close_f(error) diff --git a/fortran/test/tH5A.F90 b/fortran/test/tH5A.F90 index 4d56bed..545de2d 100644 --- a/fortran/test/tH5A.F90 +++ b/fortran/test/tH5A.F90 @@ -81,7 +81,7 @@ CONTAINS INTEGER(HID_T) :: atype5_id !Integer Attribute Datatype identifier INTEGER(HSIZE_T), DIMENSION(1) :: adims = (/2/) ! Attribute dimension INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension - INTEGER :: arank = 1 ! Attribure rank + INTEGER :: arank = 1 ! Attribute rank INTEGER(SIZE_T) :: attrlen ! Length of the attribute string INTEGER(HID_T) :: attr_space !Returned String Attribute Space identifier @@ -383,41 +383,41 @@ CONTAINS CALL h5dopen_f(file_id, dsetname, dset_id, error) CALL check("h5dopen_f",error,total_error) ! - !open the String attrbute by name + !open the String attribute by name ! CALL h5aopen_name_f(dset_id, aname, attr_id, error) CALL check("h5aopen_name_f",error,total_error) ! - !open the CHARACTER attrbute by name + !open the CHARACTER attribute by name ! CALL h5aopen_name_f(dset_id, aname2, attr2_id, error) CALL check("h5aopen_name_f",error,total_error) ! - !open the DOUBLE attrbute by name + !open the DOUBLE attribute by name ! CALL h5aopen_name_f(dset_id, aname3, attr3_id, error) CALL check("h5aopen_name_f",error,total_error) ! - !open the REAL attrbute by name + !open the REAL attribute by name ! CALL h5aopen_name_f(dset_id, aname4, attr4_id, error) CALL check("h5aopen_name_f",error,total_error) ! - !open the INTEGER attrbute by name + !open the INTEGER attribute by name ! CALL h5aopen_name_f(dset_id, aname5, attr5_id, error) CALL check("h5aopen_idx_f",error,total_error) ! - !open the NULL attrbute by name + !open the NULL attribute by name ! CALL h5aopen_name_f(dset_id, aname6, attr6_id, error) CALL check("h5aopen_idx_f",error,total_error) ! - !get the attrbute name + !get the attribute name ! CALL h5aget_name_f(attr5_id, name_size, attr_name, error) CALL check("h5aget_name_f",error,total_error) @@ -429,44 +429,44 @@ CONTAINS END IF ! - !get the STRING attrbute space + !get the STRING attribute space ! CALL h5aget_space_f(attr_id, attr_space, error) CALL check("h5aget_space_f",error,total_error) ! - !get other attrbute space + !get other attribute space ! CALL h5aget_space_f(attr2_id, attr2_space, error) CALL check("h5aget_space_f",error,total_error) ! - !get the string attrbute datatype + !get the string attribute datatype ! CALL h5aget_type_f(attr_id, attr_type, error) CALL check("h5aget_type_f",error,total_error) ! - !get the character attrbute datatype + !get the character attribute datatype ! CALL h5aget_type_f(attr2_id, attr2_type, error) CALL check("h5aget_type_f",error,total_error) ! - !get the double attrbute datatype + !get the double attribute datatype ! CALL h5aget_type_f(attr3_id, attr3_type, error) CALL check("h5aget_type_f",error,total_error) ! - !get the real attrbute datatype + !get the real attribute datatype ! CALL h5aget_type_f(attr4_id, attr4_type, error) CALL check("h5aget_type_f",error,total_error) ! - !get the integer attrbute datatype + !get the integer attribute datatype ! CALL h5aget_type_f(attr5_id, attr5_type, error) CALL check("h5aget_type_f",error,total_error) ! - !get the null attrbute datatype + !get the null attribute datatype ! CALL h5aget_type_f(attr6_id, attr6_type, error) CALL check("h5aget_type_f",error,total_error) @@ -497,7 +497,7 @@ CONTAINS CALL check("h5aread_f",error,total_error) IF ( (aread_data(1) .NE. attr_data(1)) .OR. (aread_data(2) .NE. attr_data(2)) ) THEN - WRITE(*,*) "Read back string attrbute is wrong", aread_data(1), aread_data(2) + WRITE(*,*) "Read back string attribute is wrong", aread_data(1), aread_data(2) total_error = total_error + 1 END IF @@ -507,7 +507,7 @@ CONTAINS CALL h5aread_f(attr2_id, H5T_NATIVE_CHARACTER, aread_character_data, data_dims, error) CALL check("h5aread_f",error,total_error) IF (aread_character_data .NE. 'A' ) THEN - WRITE(*,*) "Read back character attrbute is wrong ",aread_character_data + WRITE(*,*) "Read back character attribute is wrong ",aread_character_data total_error = total_error + 1 END IF ! @@ -516,7 +516,7 @@ CONTAINS data_dims(1) = 1 CALL h5aread_f(attr3_id, H5T_NATIVE_DOUBLE, aread_double_data, data_dims, error) CALL check("h5aread_f",error,total_error) - CALL VERIFY("Read back double attrbute is wrong", aread_double_data(1),3.459_Fortran_DOUBLE,total_error) + CALL VERIFY("Read back double attribute is wrong", aread_double_data(1),3.459_Fortran_DOUBLE,total_error) ! !read the real attribute data back to memory @@ -524,7 +524,7 @@ CONTAINS data_dims(1) = 1 CALL h5aread_f(attr4_id, H5T_NATIVE_REAL, aread_real_data, data_dims, error) CALL check("h5aread_f",error,total_error) - CALL VERIFY("Read back real attrbute is wrong", aread_real_data(1),4.0,total_error) + CALL VERIFY("Read back real attribute is wrong", aread_real_data(1),4.0,total_error) ! !read the Integer attribute data back to memory ! @@ -532,7 +532,7 @@ CONTAINS CALL h5aread_f(attr5_id, H5T_NATIVE_INTEGER, aread_integer_data, data_dims, error) CALL check("h5aread_f",error,total_error) IF (aread_integer_data(1) .NE. 5 ) THEN - WRITE(*,*) "Read back integer attrbute is wrong ", aread_integer_data + WRITE(*,*) "Read back integer attribute is wrong ", aread_integer_data total_error = total_error + 1 END IF ! @@ -542,7 +542,7 @@ CONTAINS CALL h5aread_f(attr6_id, H5T_NATIVE_INTEGER, aread_null_data, data_dims, error) CALL check("h5aread_f",error,total_error) IF (aread_null_data(1) .NE. 7 ) THEN - WRITE(*,*) "Read back null attrbute is wrong ", aread_null_data + WRITE(*,*) "Read back null attribute is wrong ", aread_null_data total_error = total_error + 1 END IF diff --git a/fortran/test/tH5A_1_8.F90 b/fortran/test/tH5A_1_8.F90 index 4e02c58..eb781a5 100644 --- a/fortran/test/tH5A_1_8.F90 +++ b/fortran/test/tH5A_1_8.F90 @@ -64,7 +64,7 @@ SUBROUTINE attribute_test_1_8(cleanup, total_error) INTEGER :: ret_total_error ! ******************** -! test_attr equivelent +! test_attr equivalent ! ******************** ! WRITE(*,*) "TESTING ATTRIBUTES" @@ -840,7 +840,7 @@ SUBROUTINE test_attr_info_by_idx(new_format, fcpl, fapl, total_error) my_dataset = dset3 END SELECT - ! Check for query on non-existant attribute + ! Check for query on non-existent attribute n = 0 @@ -1111,7 +1111,7 @@ SUBROUTINE test_attr_shared_rename( fcpl, fapl, total_error) INTEGER :: test_shared INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension - INTEGER :: arank = 1 ! Attribure rank + INTEGER :: arank = 1 ! Attribute rank ! Initialize "big" attribute data @@ -1497,7 +1497,7 @@ SUBROUTINE test_attr_delete_by_idx(new_format, fcpl, fapl, total_error) END SELECT - ! Check for deleting non-existant attribute + ! Check for deleting non-existent attribute !EP CALL H5Adelete_by_idx_f(my_dataset, '.', idx_type, order, 0_HSIZE_T,error, lapl_id=H5P_DEFAULT_F) CALL H5Adelete_by_idx_f(my_dataset, '.', idx_type, order, hzero,error, lapl_id=H5P_DEFAULT_F) CALL verify("H5Adelete_by_idx_f",error,minusone,total_error) @@ -1773,7 +1773,7 @@ SUBROUTINE test_attr_shared_delete(fcpl, fapl, total_error) INTEGER :: test_shared INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension - INTEGER :: arank = 1 ! Attribure rank + INTEGER :: arank = 1 ! Attribute rank ! Output message about test being performed diff --git a/fortran/test/tH5D.F90 b/fortran/test/tH5D.F90 index 006d62a..5f8484e 100644 --- a/fortran/test/tH5D.F90 +++ b/fortran/test/tH5D.F90 @@ -187,7 +187,7 @@ CONTAINS DO i = 1, 4 DO j = 1, 6 IF (data_out(i,j) .NE. dset_data(i, j)) THEN - WRITE(*, *) "dataset test error occured" + WRITE(*, *) "dataset test error occurred" WRITE(*,*) "data read is not the same as the data written" END IF END DO @@ -196,7 +196,7 @@ CONTAINS ! Check if no change to null_dset_data ! IF (null_dset_data .NE. 1) THEN - WRITE(*, *) "null dataset test error occured" + WRITE(*, *) "null dataset test error occurred" END IF ! ! End access to the dataset and release resources used by it. @@ -258,7 +258,7 @@ CONTAINS INTEGER(HID_T) :: dset_id ! Dataset identifier INTEGER(HID_T) :: dataspace ! Dataspace identifier INTEGER(HID_T) :: memspace ! memory Dataspace identifier - INTEGER(HID_T) :: crp_list ! dataset creatation property identifier + INTEGER(HID_T) :: crp_list ! dataset creation property identifier ! !dataset dimensions at creation time @@ -313,7 +313,7 @@ CONTAINS END DO ! - !Initialize FORTRAN predifined datatypes + !Initialize FORTRAN predefined datatypes ! ! CALL h5init_types_f(error) ! CALL check("h5init_types_f",error,total_error) @@ -347,7 +347,7 @@ CONTAINS CALL check("h5pset_chunk_f",error,total_error) ! - !Create a dataset with 3X3 dimensions using cparms creation propertie . + !Create a dataset with 3X3 dimensions using cparms creation properties. ! CALL h5dcreate_f(file_id, dsetname, H5T_NATIVE_INTEGER, dataspace, dset_id, error, crp_list ) CALL check("h5dcreate_f",error,total_error) @@ -426,7 +426,7 @@ CONTAINS CALL h5sget_simple_extent_ndims_f(dataspace, rankr, error) CALL check("h5sget_simple_extent_ndims_f",error,total_error) IF (rankr .NE. RANK) THEN - WRITE(*,*) "dataset rank error occured" + WRITE(*,*) "dataset rank error occurred" STOP END IF @@ -436,7 +436,7 @@ CONTAINS CALL h5sget_simple_extent_dims_f(dataspace, dimsr, maxdimsr, error) CALL check("h5sget_simple_extent_dims_f",error,total_error) IF ((dimsr(1) .NE. dims1(1)) .OR. (dimsr(2) .NE. dims1(2))) THEN - WRITE(*,*) "dataset dimensions error occured" + WRITE(*,*) "dataset dimensions error occurred" STOP END IF @@ -466,7 +466,7 @@ CONTAINS DO ih = 1, dims1(1) DO jh = 1, dims1(2) IF (data_out(ih,jh) .NE. data_in(ih, jh)) THEN - WRITE(*, *) "extend dataset test error occured" + WRITE(*, *) "extend dataset test error occurred" WRITE(*, *) "read value is not the same as the written values" END IF END DO diff --git a/fortran/test/tH5F.F90 b/fortran/test/tH5F.F90 index 2f1fcd1..936cf84 100644 --- a/fortran/test/tH5F.F90 +++ b/fortran/test/tH5F.F90 @@ -508,7 +508,7 @@ CONTAINS do i = 1, NX do j = 1, NY IF (data_out(i,j) .NE. dset_data(i, j)) THEN - write(*, *) "reopen test error occured" + write(*, *) "reopen test error occurred" END IF end do end do diff --git a/fortran/test/tH5G_1_8.F90 b/fortran/test/tH5G_1_8.F90 index ec0424f..4cd1eef 100644 --- a/fortran/test/tH5G_1_8.F90 +++ b/fortran/test/tH5G_1_8.F90 @@ -654,7 +654,7 @@ SUBROUTINE group_info(cleanup, fapl, total_error) INTEGER(HID_T) :: file, scalar, grp, d1 CHARACTER(LEN=12), PARAMETER :: filename ='TestLinks.h5' INTEGER(HSIZE_T), DIMENSION(1) :: adims2 = (/1/) ! Attribute dimension - INTEGER :: arank = 1 ! Attribure rank + INTEGER :: arank = 1 ! Attribute rank INTEGER :: error INTEGER :: cset ! Indicates the character set used for the link’s name. diff --git a/fortran/test/tH5MISC_1_8.F90 b/fortran/test/tH5MISC_1_8.F90 index 64cbad3..474e34e 100644 --- a/fortran/test/tH5MISC_1_8.F90 +++ b/fortran/test/tH5MISC_1_8.F90 @@ -160,7 +160,7 @@ SUBROUTINE test_genprop_basic_class(total_error) CALL verify("H5Pequal_f", flag, .TRUE., total_error) - ! Make certain false postives aren't being returned + ! Make certain false positives aren't being returned CALL H5Pequal_f(cid2, H5P_FILE_CREATE_F, flag, error) CALL check("H5Pequal_f", error, total_error) CALL verify("H5Pequal_f", flag, .FALSE., total_error) diff --git a/fortran/test/tH5P.F90 b/fortran/test/tH5P.F90 index 1d1208c..d664dd7 100644 --- a/fortran/test/tH5P.F90 +++ b/fortran/test/tH5P.F90 @@ -349,8 +349,8 @@ SUBROUTINE multi_file_test(cleanup, total_error) DO i = 1, 4 DO j = 1, 6 IF (data_out(i,j) .NE. dset_data(i, j)) THEN - WRITE(*, *) "dataset test error occured" - WRITE(*,*) "data read is not the same as the data writen" + WRITE(*, *) "dataset test error occurred" + WRITE(*,*) "data read is not the same as the data written" END IF END DO END DO @@ -541,7 +541,7 @@ SUBROUTINE test_chunk_cache(cleanup, total_error) CALL H5Dclose_f(dsid, error) CALL H5Oopen_f(fid, "dset", dsid, error, dapl1) - ! Retrieve dapl from dataset, verfiy cache values are the same as on dapl1 + ! Retrieve dapl from dataset, verify cache values are the same as on dapl1 ! ! Note we rely on the knowledge that H5Pget_chunk_cache retrieves these ! values directly from the dataset structure, and not from a copy of the @@ -563,7 +563,7 @@ SUBROUTINE test_chunk_cache(cleanup, total_error) CALL H5Oopen_f(fid, "dset", dsid, error) CALL check("H5Oopen_f", error, total_error) - ! Retrieve dapl from dataset, verfiy cache values are the same as on fapl_local + ! Retrieve dapl from dataset, verify cache values are the same as on fapl_local CALL H5Dget_access_plist_f(dsid, dapl2, error) CALL check("H5Dget_access_plist_f", error, total_error) @@ -575,7 +575,7 @@ SUBROUTINE test_chunk_cache(cleanup, total_error) CALL H5Pclose_f(dapl2,error) CALL check("H5Pclose_f", error, total_error) - ! Similary, test use of H5Dcreate2 with H5P_DEFAULT + ! Similarly, test use of H5Dcreate2 with H5P_DEFAULT CALL H5Dclose_f(dsid, error) CALL check("H5Dclose_f", error, total_error) diff --git a/fortran/test/tH5P_F03.F90 b/fortran/test/tH5P_F03.F90 index ebc11c8..0b0a320 100644 --- a/fortran/test/tH5P_F03.F90 +++ b/fortran/test/tH5P_F03.F90 @@ -205,7 +205,7 @@ SUBROUTINE test_create(total_error) CALL h5fclose_f(file,error) CALL check("h5fclose_f", error, total_error) - ! Open the file and get the dataset fill value from each dataset + ! Open the file and get the dataset fill value from each dataset CALL h5pcreate_f(H5P_FILE_ACCESS_F, fapl, error) CALL check("h5pcreate_f",error, total_error) @@ -553,7 +553,7 @@ SUBROUTINE external_test_offset(cleanup,total_error) ! ! Create the file and an initial group. CALL h5pcreate_f(H5P_FILE_ACCESS_F, fapl, error) - CALL h5fcreate_f('extren_raw.h5', H5F_ACC_TRUNC_F, file, error, access_prp=fapl) + CALL h5fcreate_f('extern_raw.h5', H5F_ACC_TRUNC_F, file, error, access_prp=fapl) CALL check("h5fcreate_f",error,total_error) CALL h5gcreate_f(file, "emit-diagnostics", grp, error) @@ -632,7 +632,7 @@ SUBROUTINE external_test_offset(cleanup,total_error) CALL h5_cleanup_f(filename, H5P_DEFAULT_F, error) CALL check("h5_cleanup_f", error, total_error) ENDDO - IF(cleanup) CALL h5_cleanup_f("extren_raw.h5", H5P_DEFAULT_F, error) + IF(cleanup) CALL h5_cleanup_f("extern_raw.h5", H5P_DEFAULT_F, error) CALL check("h5_cleanup_f", error, total_error) END SUBROUTINE external_test_offset @@ -999,28 +999,28 @@ SUBROUTINE test_vds(total_error) CALL check("H5Pget_virtual_filename_f", error, total_error) IF(nsize.NE.LEN(SRC_FILE_LEN_EXACT))THEN - PRINT*,"virtual filenname size is incorrect" + PRINT*,"virtual filename size is incorrect" total_error = total_error + 1 ENDIF ! check passing a buffer that is very small CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_TINY, error) CALL check("H5Pget_virtual_filename_f", error, total_error) IF(SRC_FILE_LEN_TINY.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_TINY)))THEN - PRINT*,"virtual filenname returned is incorrect" + PRINT*,"virtual filename returned is incorrect" total_error = total_error + 1 ENDIF ! check passing a buffer that small by one CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_SMALL, error) CALL check("H5Pget_virtual_filename_f", error, total_error) IF(SRC_FILE_LEN_SMALL.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_SMALL)))THEN - PRINT*,"virtual filenname returned is incorrect" + PRINT*,"virtual filename returned is incorrect" total_error = total_error + 1 ENDIF ! check passing a buffer that is exact CALL H5Pget_virtual_filename_f(dcpl, INT(i-1, size_t), SRC_FILE_LEN_EXACT, error) CALL check("H5Pget_virtual_filename_f", error, total_error) IF(SRC_FILE_LEN_EXACT.NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)))THEN - PRINT*,"virtual filenname returned is incorrect" + PRINT*,"virtual filename returned is incorrect" total_error = total_error + 1 ENDIF ! check passing a buffer that bigger by one @@ -1028,7 +1028,7 @@ SUBROUTINE test_vds(total_error) CALL check("H5Pget_virtual_filename_f", error, total_error) IF(SRC_FILE_LEN_LARGE(1:LEN(SRC_FILE_LEN_EXACT)).NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)).AND. & SRC_FILE_LEN_LARGE(LEN(SRC_FILE_LEN_EXACT):).NE.'')THEN - PRINT*,"virtual filenname returned is incorrect" + PRINT*,"virtual filename returned is incorrect" total_error = total_error + 1 ENDIF ! check passing a buffer that is very big @@ -1036,7 +1036,7 @@ SUBROUTINE test_vds(total_error) CALL check("H5Pget_virtual_filename_f", error, total_error) IF(SRC_FILE_LEN_HUGE(1:LEN(SRC_FILE_LEN_EXACT)).NE.SRC_FILE(i)(1:LEN(SRC_FILE_LEN_EXACT)).AND. & SRC_FILE_LEN_HUGE(LEN(SRC_FILE_LEN_EXACT):).NE.'')THEN - PRINT*,"virtual filenname returned is incorrect" + PRINT*,"virtual filename returned is incorrect" total_error = total_error + 1 ENDIF ! Get source dataset name diff --git a/fortran/test/tH5R.F90 b/fortran/test/tH5R.F90 index 5144d83..4e62b20 100644 --- a/fortran/test/tH5R.F90 +++ b/fortran/test/tH5R.F90 @@ -142,7 +142,7 @@ SUBROUTINE refobjtest(cleanup, total_error) CALL check("h5gclose_f",error,total_error) ! - ! Craete references to two groups, integer dataset and shared datatype + ! Create references to two groups, integer dataset and shared datatype ! and write it to the dataset in the file ! CALL h5rcreate_f(file_id, groupname1, ref(1), error) diff --git a/fortran/test/tH5S.F90 b/fortran/test/tH5S.F90 index 08f0b59..302db51 100644 --- a/fortran/test/tH5S.F90 +++ b/fortran/test/tH5S.F90 @@ -151,7 +151,7 @@ CONTAINS CALL h5sget_simple_extent_dims_f(space2_id, dimsout, maxdimsout, error) CALL check("h5sget_simple_extent_dims_f", error, total_error) IF ((dimsout(1) .NE. dims1(1)) .OR. (dimsout(2) .NE. dims1(2)) ) THEN - write(*,*)"error occured, copied dims not same" + write(*,*)"error occurred, copied dims not same" END IF ! @@ -159,14 +159,14 @@ CONTAINS ! CALL h5sget_simple_extent_ndims_f(space2_id, rank2, error) CALL check("h5sget_simple_extent_ndims_f", error, total_error) - IF (rank2 .NE. rank1) write(*,*)"error occured, copied ranks not same" + IF (rank2 .NE. rank1) write(*,*)"error occurred, copied ranks not same" ! !get the copied space's number of elements. ! CALL h5sget_simple_extent_npoints_f(space2_id, npoints, error) CALL check("h5sget_simple_extent_npoints_f", error, total_error) - IF (npoints .NE. 24) write(*,*)"error occured, number of elements not correct" + IF (npoints .NE. 24) write(*,*)"error occurred, number of elements not correct" ! @@ -194,7 +194,7 @@ CONTAINS CALL h5sget_simple_extent_dims_f(space2_id, dimsout, maxdimsout, error) CALL check("h5sget_simple_extent_dims_f", error, total_error) IF ((dimsout(1) .NE. dims2(1)) .OR. (dimsout(2) .NE. dims2(2)) ) THEN - write(*,*)"error occured, copied dims not same" + write(*,*)"error occurred, copied dims not same" END IF ! @@ -235,8 +235,8 @@ CONTAINS do i = 1, 4 do j = 1, 6 IF (data1_out(i,j) .NE. data1_in(i, j)) THEN - write(*, *) "dataset test error occured" - write(*,*) "data read is not the same as the data writen" + write(*, *) "dataset test error occurred" + write(*,*) "data read is not the same as the data written" END IF end do end do @@ -256,8 +256,8 @@ CONTAINS do i = 1, 6 do j = 1, 6 IF (data2_out(i,j) .NE. data2_in(i, j)) THEN - write(*, *) "dataset test error occured" - write(*,*) "data read is not the same as the data writen" + write(*, *) "dataset test error occurred" + write(*,*) "data read is not the same as the data written" END IF end do end do diff --git a/fortran/test/tH5Sselect.F90 b/fortran/test/tH5Sselect.F90 index 178ac32..41cd16f 100644 --- a/fortran/test/tH5Sselect.F90 +++ b/fortran/test/tH5Sselect.F90 @@ -149,7 +149,7 @@ CONTAINS ! ! - !Initialize FORTRAN predifined datatypes + !Initialize FORTRAN predefined datatypes ! ! CALL h5init_types_f(error) ! CALL check("h5init_types_f", error, total_error) @@ -430,7 +430,7 @@ CONTAINS end do ! - !Initialize FORTRAN predifined datatypes + !Initialize FORTRAN predefined datatypes ! ! CALL h5init_types_f(error) ! CALL check("h5init_types_f", error, total_error) @@ -912,7 +912,7 @@ CONTAINS ! CALL h5sget_select_hyper_nblocks_f(dataspace, num_blocks, error) CALL check("h5sget_select_hyper_nblocks_f", error, total_error) - IF (num_blocks .NE. 4) write (*,*) "error occured with num_blocks" + IF (num_blocks .NE. 4) write (*,*) "error occurred with num_blocks" !write(*,*) num_blocks !result of num_blocks is 4 @@ -945,11 +945,11 @@ CONTAINS CALL h5sget_select_bounds_f(dataspace, startout, endout, error) CALL check("h5sget_select_bounds_f", error, total_error) IF ( (startout(1) .ne. 1) .or. (startout(2) .ne. 1) ) THEN - write(*,*) "error occured to select_bounds's start position" + write(*,*) "error occurred to select_bounds's start position" END IF IF ( (endout(1) .ne. 5) .or. (endout(2) .ne. 5) ) THEN - write(*,*) "error occured to select_bounds's end position" + write(*,*) "error occurred to select_bounds's end position" END IF !write(*,*) (startout(i), i = 1, RANK) !result of startout is 0, 0 @@ -978,7 +978,7 @@ CONTAINS ! CALL h5sget_select_elem_npoints_f(dataspace, num_points, error) CALL check("h5sget_select_elem_npoints_f", error, total_error) - IF (num_points .NE. 10) write(*,*) "error occured with num_points" + IF (num_points .NE. 10) write(*,*) "error occurred with num_points" !write(*,*) num_points ! result of num_points is 10 diff --git a/fortran/test/tH5T.F90 b/fortran/test/tH5T.F90 index 8b10136..7673993 100644 --- a/fortran/test/tH5T.F90 +++ b/fortran/test/tH5T.F90 @@ -72,7 +72,7 @@ CONTAINS INTEGER(HID_T) :: dt4_id ! Memory datatype identifier (for real field) INTEGER(HID_T) :: dt5_id ! Memory datatype identifier INTEGER(HID_T) :: membtype_id ! Datatype identifier - INTEGER(HID_T) :: plist_id ! Dataset trasfer property + INTEGER(HID_T) :: plist_id ! Dataset transfer property INTEGER(HSIZE_T), DIMENSION(1) :: dims = (/dimsize/) ! Dataset dimensions @@ -98,7 +98,7 @@ CONTAINS INTEGER :: class ! Datatype class INTEGER :: num_members ! Number of members in the compound datatype CHARACTER(LEN=256) :: member_name - INTEGER :: len ! Lenght of the name of the compound datatype member + INTEGER :: len ! Length of the name of the compound datatype member INTEGER :: member_index ! index of the field INTEGER(HSIZE_T), DIMENSION(3) :: array_dims=(/2,3,4/) INTEGER :: array_dims_range = 3 @@ -219,9 +219,9 @@ CONTAINS CALL h5tcreate_f(H5T_COMPOUND_F, type_size, dtype_id, error) CALL check("h5tcreate_f", error, total_error) ! - ! Insert memebers + ! Insert members ! - ! CHARACTER*2 memeber + ! CHARACTER*2 member ! offset = 0 CALL h5tinsert_f(dtype_id, "char_field", offset, dt5_id, error) @@ -229,13 +229,13 @@ CONTAINS ! ! INTEGER member ! - offset = offset + type_sizec ! Offset of the second memeber is 2 + offset = offset + type_sizec ! Offset of the second member is 2 CALL h5tinsert_f(dtype_id, "integer_field", offset, H5T_NATIVE_INTEGER, error) CALL check("h5tinsert_f", error, total_error) ! ! DOUBLE PRECISION member ! - offset = offset + type_sizei ! Offset of the third memeber is 6 + offset = offset + type_sizei ! Offset of the third member is 6 CALL h5tinsert_f(dtype_id, "double_field", offset, H5T_NATIVE_DOUBLE, error) CALL check("h5tinsert_f", error, total_error) ! @@ -416,7 +416,7 @@ CONTAINS CALL h5tequal_f(membtype_id, H5T_NATIVE_INTEGER, flag, error) CALL check("h5tequal_f", error, total_error) if(.not. flag) then - write(*,*) "Wrong member type returned for integer memebr" + write(*,*) "Wrong member type returned for integer member" total_error = total_error + 1 endif CALL h5tget_member_class_f(dtype_id, i-1, class, error) @@ -435,7 +435,7 @@ CONTAINS CALL h5tequal_f(membtype_id, H5T_NATIVE_DOUBLE, flag, error) CALL check("h5tequal_f", error, total_error) if(.not. flag) then - write(*,*) "Wrong member type returned for double precision memebr" + write(*,*) "Wrong member type returned for double precision member" total_error = total_error + 1 endif CALL h5tget_member_class_f(dtype_id, i-1, class, error) @@ -454,7 +454,7 @@ CONTAINS CALL h5tequal_f(membtype_id, H5T_NATIVE_REAL, flag, error) CALL check("h5tequal_f", error, total_error) if(.not. flag) then - write(*,*) "Wrong member type returned for real memebr" + write(*,*) "Wrong member type returned for real member" total_error = total_error + 1 endif CALL h5tget_member_class_f(dtype_id, i-1, class, error) @@ -862,7 +862,7 @@ CONTAINS CALL h5fcreate_f(fix_filename,H5F_ACC_TRUNC_F,file_id,error) CALL check("h5fcreate_f", error, total_error) ! - ! Create enumeration datatype with tow values + ! Create enumeration datatype with two values ! CALL h5tenum_create_f(H5T_NATIVE_INTEGER,dtype_id,error) CALL check("h5tenum_create_f", error, total_error) @@ -871,7 +871,7 @@ CONTAINS CALL h5tenum_insert_f(dtype_id,false,DATA(2),error) CALL check("h5tenum_insert_f", error, total_error) ! - ! Create write and close a dataset with enum datatype + ! Create write and close a dataset with enum datatype ! CALL h5screate_simple_f(1,dsize,dspace_id,error) CALL check("h5screate_simple_f", error, total_error) diff --git a/fortran/test/tH5VL.F90 b/fortran/test/tH5VL.F90 index 2269fec..eb0470c 100644 --- a/fortran/test/tH5VL.F90 +++ b/fortran/test/tH5VL.F90 @@ -147,7 +147,7 @@ CONTAINS CALL check("h5dvlen_get_max_len_f", error, total_error) if(max_len .ne. data_dims(1)) then total_error = total_error + 1 - write(*,*) "Wrong number of elemets returned by h5dvlen_get_max_len_f" + write(*,*) "Wrong number of elements returned by h5dvlen_get_max_len_f" endif ! ! Read the dataset. @@ -317,7 +317,7 @@ CONTAINS CALL check("h5dvlen_get_max_len_f", error, total_error) if(max_len .ne. data_dims(1)) then total_error = total_error + 1 - write(*,*) "Wrong number of elemets returned by h5dvlen_get_max_len_f" + write(*,*) "Wrong number of elements returned by h5dvlen_get_max_len_f" endif ! ! Read the dataset. diff --git a/fortran/test/tH5Z.F90 b/fortran/test/tH5Z.F90 index 8c39fea..799067a 100644 --- a/fortran/test/tH5Z.F90 +++ b/fortran/test/tH5Z.F90 @@ -374,7 +374,7 @@ CONTAINS do i = 1, N do j = 1, M IF (data_out(i,j) .NE. dset_data(i, j)) THEN - write(*, *) "dataset test error occured" + write(*, *) "dataset test error occurred" write(*,*) "data read is not the same as the data written" num_errors = num_errors + 1 IF (num_errors .GE. 512) THEN diff --git a/fortran/test/tHDF5.F90 b/fortran/test/tHDF5.F90 index db829af..2bd373d 100644 --- a/fortran/test/tHDF5.F90 +++ b/fortran/test/tHDF5.F90 @@ -7,7 +7,7 @@ ! src/fortran/test/tHDF5.f90 ! ! PURPOSE -! This is the test module used for testing the Fortran90 HDF library APIs. +! This is the test module used for testing the Fortran90 HDF library APIs. ! ! COPYRIGHT ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * diff --git a/fortran/test/tHDF5_F03.F90 b/fortran/test/tHDF5_F03.F90 index 6d318ea..1194b1e 100644 --- a/fortran/test/tHDF5_F03.F90 +++ b/fortran/test/tHDF5_F03.F90 @@ -7,7 +7,7 @@ ! src/fortran/test/tHDF5_F03.f90 ! ! PURPOSE -! This is the test module used for testing the Fortran2003 HDF +! This is the test module used for testing the Fortran2003 HDF ! library APIS. ! ! COPYRIGHT diff --git a/hl/fortran/src/H5HL_buildiface.F90 b/hl/fortran/src/H5HL_buildiface.F90 index ca66ebf..9c783d3 100644 --- a/hl/fortran/src/H5HL_buildiface.F90 +++ b/hl/fortran/src/H5HL_buildiface.F90 @@ -34,7 +34,7 @@ ! This file is part of HDF5. The full HDF5 copyright notice, including * ! terms governing use, modification, and redistribution, is contained in * ! the COPYING file, which can be found at the root of the source code * -! distribution tree, or in https://www.hdfgroup.org/licenses. * +! distribution tree, or in https://www.hdfgroup.org/licenses. * ! If you do not have access to either file, you may request a copy from * ! help@hdfgroup.org. * ! * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * diff --git a/hl/fortran/src/H5TBfc.c b/hl/fortran/src/H5TBfc.c index 03bbd00..0997c68 100644 --- a/hl/fortran/src/H5TBfc.c +++ b/hl/fortran/src/H5TBfc.c @@ -83,7 +83,7 @@ h5tbmake_table_c(size_t_f *namelen1, _fcd name1, hid_t_f *loc_id, size_t_f *name (tmp = (char *)HD5f2cstring(field_names, (size_t) * (max_char_size_field_names) * (size_t)num_elem))) HGOTO_DONE(FAIL) /* - * move data from temorary buffer + * move data from temporary buffer */ tmp_p = tmp; for (i = 0; i < num_elem; i++) { @@ -188,7 +188,7 @@ h5tbmake_table_ptr_c(size_t_f *namelen1, _fcd name1, hid_t_f *loc_id, size_t_f * (tmp = (char *)HD5f2cstring(field_names, (size_t) * (max_char_size_field_names) * (size_t)num_elem))) HGOTO_DONE(FAIL) /* - * move data from temorary buffer + * move data from temporary buffer */ tmp_p = tmp; for (i = 0; i < num_elem; i++) { diff --git a/hl/fortran/test/tsttable.F90 b/hl/fortran/test/tsttable.F90 index a77068c..e7bd88e 100644 --- a/hl/fortran/test/tsttable.F90 +++ b/hl/fortran/test/tsttable.F90 @@ -172,11 +172,11 @@ SUBROUTINE test_table1() ! offset = 0 field_offset(1) = offset - offset = offset + type_sizec ! Offset of the second memeber is 2 + offset = offset + type_sizec ! Offset of the second member is 2 field_offset(2) = offset - offset = offset + type_sizei ! Offset of the second memeber is 6 + offset = offset + type_sizei ! Offset of the second member is 6 field_offset(3) = offset - offset = offset + type_sized ! Offset of the second memeber is 14 + offset = offset + type_sized ! Offset of the second member is 14 field_offset(4) = offset !------------------------------------------------------------------------- @@ -571,7 +571,7 @@ SUBROUTINE test_table2() INTEGER(HSIZE_T), PARAMETER :: nrecords = 8 ! nrecords CHARACTER(len=8), PARAMETER :: filename = "f2tab.h5" ! File name - CHARACTER(LEN=5), PARAMETER :: table_name = "tabel" ! table name + CHARACTER(LEN=5), PARAMETER :: table_name = "table" ! table name CHARACTER(LEN=10), PARAMETER :: table_name_fill = "tabel_fill" ! table name ! Define field information diff --git a/hl/src/H5DS.c b/hl/src/H5DS.c index 59fc197..164bc35 100644 --- a/hl/src/H5DS.c +++ b/hl/src/H5DS.c @@ -55,7 +55,7 @@ H5DSset_scale(hid_t dsid, const char *dimname) return FAIL; /*------------------------------------------------------------------------- - * check if the dataset is a dataset wich has references to dimension scales + * check if the dataset is a dataset which has references to dimension scales *------------------------------------------------------------------------- */ @@ -864,7 +864,7 @@ out: dsbuf = NULL; } if (buf) { - /* Failure occured before H5Dvlen_reclaim was called; + /* Failure occurred before H5Dvlen_reclaim was called; free the pointers allocated when we read data in */ for (i = 0; i < rank; i++) { if (buf[i].p) @@ -1224,7 +1224,7 @@ H5DSiterate_scales(hid_t did, unsigned int dim, int *ds_idx, H5DS_iterate_t visi if (H5I_DATASET != it) return FAIL; - /* get the number of scales assotiated with this DIM */ + /* get the number of scales associated with this DIM */ if ((nscales = H5DSget_num_scales(did, dim)) < 0) return FAIL; diff --git a/hl/test/test_ds.c b/hl/test/test_ds.c index 52746ca..ddc4b80 100644 --- a/hl/test/test_ds.c +++ b/hl/test/test_ds.c @@ -3020,7 +3020,7 @@ test_simple(void) } /*------------------------------------------------------------------------- - * dettach for DIM0 + * detach for DIM0 *------------------------------------------------------------------------- */ @@ -3318,11 +3318,11 @@ test_simple(void) goto out; dim = 0; - /* iterate trough the 1st dimension of "dset_a" and verify that its DS is valid */ + /* iterate through the 1st dimension of "dset_a" and verify that its DS is valid */ if (H5DSiterate_scales(did, dim, NULL, verify_scale, NULL) < 0) goto out; - /* iterate trough the 2nd dimension of "dset_a" and verify that its DS is valid + /* iterate through the 2nd dimension of "dset_a" and verify that its DS is valid start at DS index 2 */ dim = 1; scale_idx = 2; @@ -3343,11 +3343,11 @@ test_simple(void) goto out; dim = 0; - /* iterate trough the 1st dimension of "dset_a" and read the DS */ + /* iterate through the 1st dimension of "dset_a" and read the DS */ if (H5DSiterate_scales(did, dim, NULL, read_scale, s1_wbuf) < 0) goto out; - /* iterate trough the 2nd dimension of "dset_a" and read the DS + /* iterate through the 2nd dimension of "dset_a" and read the DS start at DS index 2 */ dim = 1; scale_idx = 2; @@ -3380,9 +3380,9 @@ test_simple(void) goto out; { int match_size; /* does this scale size matches the dataset DIM size */ - int idx = 0; /* scale index to start iterating, on return, index where iterator stoped */ + int idx = 0; /* scale index to start iterating, on return, index where iterator stopped */ - /* iterate trough all the dimensions */ + /* iterate through all the dimensions */ for (dim = 0; dim < (unsigned)rank; dim++) { if ((match_size = H5DSiterate_scales(did, dim, &idx, match_dim_scale, NULL)) < 0) goto out; @@ -3474,9 +3474,9 @@ test_simple(void) goto out; { int match_size; /* does this scale size matches the dataset DIM size */ - int idx; /* scale index to start iterating, on return, index where iterator stoped */ + int idx; /* scale index to start iterating, on return, index where iterator stopped */ - /* iterate trough all the dimensions */ + /* iterate through all the dimensions */ for (dim = 0; dim < (unsigned)rank; dim++) { /* always start at 1st scale */ idx = 0; @@ -4024,7 +4024,7 @@ test_errors(void) PASSED(); /*------------------------------------------------------------------------- - * try to attach a scale to an image, pallete or table + * try to attach a scale to an image, palette or table *------------------------------------------------------------------------- */ @@ -4035,7 +4035,7 @@ test_errors(void) goto out; /* make a palette */ - if (H5IMmake_palette(fid, "pallete", pal_dims, NULL) < 0) + if (H5IMmake_palette(fid, "palette", pal_dims, NULL) < 0) goto out; /* open the previous written "ds_b" */ @@ -4247,7 +4247,7 @@ test_iterators(void) if ((did = H5Dopen2(fid, "dset_a", H5P_DEFAULT)) < 0) goto out; - /* try to iterate trough the 1st dimension of "dset_a", return error */ + /* try to iterate through the 1st dimension of "dset_a", return error */ if (H5DSiterate_scales(did, 0, NULL, verify_scale, NULL) < 0) goto out; @@ -4268,7 +4268,7 @@ test_iterators(void) if ((did = H5Dopen2(fid, "dset_a", H5P_DEFAULT)) < 0) goto out; - /* try to iterate trough the 3rd dimension of "dset_a", return error */ + /* try to iterate through the 3rd dimension of "dset_a", return error */ if (H5DSiterate_scales(did, 3, NULL, verify_scale, NULL) == SUCCEED) goto out; @@ -4305,7 +4305,7 @@ test_iterators(void) goto out; } - /* iterate trough the 1st dimension of "dset_a" */ + /* iterate through the 1st dimension of "dset_a" */ if (H5DSiterate_scales(did, 0, NULL, op_continue, NULL) < 0) goto out; diff --git a/hl/test/test_file_image.c b/hl/test/test_file_image.c index 7bcf491..cc8fc33 100644 --- a/hl/test/test_file_image.c +++ b/hl/test/test_file_image.c @@ -243,7 +243,7 @@ test_file_image(size_t open_images, size_t nflags, const unsigned *flags) */ HDmemset((uint8_t *)tmp_ptr + SUPER_STATUS_FLAGS_OFF_V0_V1, (int)0, (size_t)SUPER_STATUS_FLAGS_SIZE_V0_V1); - /* Does the comparision */ + /* Does the comparison */ if (HDmemcmp(tmp_ptr, buf_ptr[i], (size_t)buf_size[i]) != 0) FAIL_PUTS_ERROR("comparison of TMP vfd and user buffer failed"); /* Free the temporary buffer */ diff --git a/hl/test/test_h5do_compat.c b/hl/test/test_h5do_compat.c index c554ae7..6f19d4b 100644 --- a/hl/test/test_h5do_compat.c +++ b/hl/test/test_h5do_compat.c @@ -31,7 +31,7 @@ * * Purpose: Test the basic functionality of H5DOwrite_chunk * - * Return: Success: An identifer for the dataset used in the tests + * Return: Success: An identifier for the dataset used in the tests * Failure: H5I_INVALID_HID * *------------------------------------------------------------------------- diff --git a/hl/test/test_image.c b/hl/test/test_image.c index 40db7c8..8c33f16 100644 --- a/hl/test/test_image.c +++ b/hl/test/test_image.c @@ -219,7 +219,7 @@ test_simple(void) *------------------------------------------------------------------------- */ - HL_TESTING2("pallete functions"); + HL_TESTING2("palette functions"); if (H5IMget_npalettes(fid, IMAGE1_NAME, &npals) < 0) goto out; @@ -422,7 +422,7 @@ test_data(void) /*------------------------------------------------------------------------- * palette #4. blue-red - * make a palette whith blue to red colors + * make a palette with blue to red colors *------------------------------------------------------------------------- */ for (i = 0, n = 0; i < 256 * 3; i += 3, n++) { diff --git a/hl/test/test_ld.c b/hl/test/test_ld.c index 5f6b8e5..2a5dbb5 100644 --- a/hl/test/test_ld.c +++ b/hl/test/test_ld.c @@ -189,7 +189,7 @@ typedef struct set_t { * This will fail on heiwa and amani when VALID_FIELDS1 is "field1,field3,field4" * because of alignment problems: * amani and heiwa - 8 byte alignment - * jam - 4 byte alignemnt + * jam - 4 byte alignment * This will need to be fixed in the library for H5Tget_native_type(). */ /* VALID_FIELDS1 "field1,field2.a,field3,field4" */ @@ -213,8 +213,8 @@ typedef struct test_valid_fields2 { #define TEST_BUF_SIZE 100 int *iibuf; /* buffer for storing retrieved elements */ int *ibuf; /* buffer for storing retrieved elements (integer) */ -set_t *cbuf; /* buffer for storing retrieved elemnets (compound) */ -set_t *ccbuf; /* buffer for storing retrieved elemnets (compound) */ +set_t *cbuf; /* buffer for storing retrieved elements (compound) */ +set_t *ccbuf; /* buffer for storing retrieved elements (compound) */ test_valid_fields1 *vbuf1; /* buffer for storing retrieved elements (FIELDS1) */ test_valid_fields2 *vbuf2; /* buffer for storing retrieved elements (FIELDS2) */ diff --git a/hl/test/test_packet_vlen.c b/hl/test/test_packet_vlen.c index d812247..9af812f 100644 --- a/hl/test/test_packet_vlen.c +++ b/hl/test/test_packet_vlen.c @@ -677,7 +677,7 @@ error: /* An error has occurred. Clean up and exit. */ * * Description: * - Added a fixed-length packet table to the file for variety - * - Use the helper funtion verify_ptlengthtype to test H5PTis_varlen + * - Use the helper function verify_ptlengthtype to test H5PTis_varlen * on each packet table. * * 2016/01/27 -BMR diff --git a/hl/test/test_table.c b/hl/test/test_table.c index 9b5d8ac..24b7336 100644 --- a/hl/test/test_table.c +++ b/hl/test/test_table.c @@ -1464,7 +1464,7 @@ test_table(hid_t fid, int do_write) HL_TESTING2("getting field info"); - /* alocate */ + /* allocate */ names_out = (char **)HDmalloc(sizeof(char *) * (size_t)NFIELDS); for (i = 0; i < NFIELDS; i++) { names_out[i] = (char *)HDmalloc(sizeof(char) * 255); diff --git a/java/src/hdf/hdf5lib/H5.java b/java/src/hdf/hdf5lib/H5.java index aafffaa..768110d 100644 --- a/java/src/hdf/hdf5lib/H5.java +++ b/java/src/hdf/hdf5lib/H5.java @@ -11171,7 +11171,7 @@ public class H5 implements java.io.Serializable { * <pre> * version_info[0] = boot // boot block version number * version_info[1] = freelist // global freelist version - * version_info[2] = stab // symbol tabl version number + * version_info[2] = stab // symbol table version number * version_info[3] = shhdr // shared object header version * </pre> * @return a non-negative value, with the values of version_info initialized, if successful @@ -14385,7 +14385,7 @@ public class H5 implements java.io.Serializable { /** * @ingroup JH5T * - * H5Tset_sign sets the sign proprety for an integer type. + * H5Tset_sign sets the sign property for an integer type. * * @param type_id * IN: Identifier of datatype to set. diff --git a/java/src/hdf/hdf5lib/callbacks/package-info.java b/java/src/hdf/hdf5lib/callbacks/package-info.java index 5ef3fab..bf7bf3c 100644 --- a/java/src/hdf/hdf5lib/callbacks/package-info.java +++ b/java/src/hdf/hdf5lib/callbacks/package-info.java @@ -1,6 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * Copyright by The HDF Group. * - * * * All rights reserved. * * * * This file is part of HDF5. The full HDF5 copyright notice, including * diff --git a/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java b/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java index 677e57f..660b919 100644 --- a/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java +++ b/java/src/hdf/hdf5lib/exceptions/HDF5LibraryException.java @@ -108,7 +108,7 @@ public class HDF5LibraryException extends HDF5Exception { /** * @ingroup JERRLIB * - * Return a error message for the minor error number. + * Return an error message for the minor error number. * * These messages come from <b>@ref H5E</b>. * diff --git a/java/src/hdf/hdf5lib/exceptions/package-info.java b/java/src/hdf/hdf5lib/exceptions/package-info.java index 2ac7806..fa261ca 100644 --- a/java/src/hdf/hdf5lib/exceptions/package-info.java +++ b/java/src/hdf/hdf5lib/exceptions/package-info.java @@ -1,6 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * Copyright by The HDF Group. * - * * * All rights reserved. * * * * This file is part of HDF5. The full HDF5 copyright notice, including * diff --git a/java/src/hdf/hdf5lib/package-info.java b/java/src/hdf/hdf5lib/package-info.java index 7ae4df9..e742197 100644 --- a/java/src/hdf/hdf5lib/package-info.java +++ b/java/src/hdf/hdf5lib/package-info.java @@ -1,6 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * Copyright by The HDF Group. * - * * * All rights reserved. * * * * This file is part of HDF5. The full HDF5 copyright notice, including * diff --git a/java/src/hdf/hdf5lib/structs/package-info.java b/java/src/hdf/hdf5lib/structs/package-info.java index a3ec6cd..7bd3dee 100644 --- a/java/src/hdf/hdf5lib/structs/package-info.java +++ b/java/src/hdf/hdf5lib/structs/package-info.java @@ -1,6 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * Copyright by The HDF Group. * - * * * All rights reserved. * * * * This file is part of HDF5. The full HDF5 copyright notice, including * diff --git a/m4/aclocal_cxx.m4 b/m4/aclocal_cxx.m4 index 2779066..188f7ca 100644 --- a/m4/aclocal_cxx.m4 +++ b/m4/aclocal_cxx.m4 @@ -20,7 +20,7 @@ dnl Contains Macros for HDF5 C++ dnl ********************************* dnl dnl Special characteristics that have no autoconf counterpart but that -dnl we need as part of the C++ support. To distinquish these, they +dnl we need as part of the C++ support. To distinguish these, they dnl have a [PAC] prefix. dnl Checking if C++ needs old style header files in includes diff --git a/release_docs/HISTORY-1_10.txt b/release_docs/HISTORY-1_10.txt index e21bfd8..9634832 100644 --- a/release_docs/HISTORY-1_10.txt +++ b/release_docs/HISTORY-1_10.txt @@ -514,7 +514,7 @@ Bug Fixes since HDF5-1.10.7 release function. This improves the I/O performance. In the case of the parallel library, it also avoids breaking to independent I/O, which makes it possible to apply a filter when writing or reading data to or from - teh HDF5 file. + te HDF5 file. (JWSB - 2021/09/13) @@ -542,7 +542,7 @@ Bug Fixes since HDF5-1.10.7 release dimension sizes. The problem was fixed by ensuring that decoding the dimension sizes - and max values will not go beyong the end of the buffer. + and max values will not go beyond the end of the buffer. (BMR - 2021/05/12, HDFFV-11223) @@ -610,7 +610,7 @@ Bug Fixes since HDF5-1.10.7 release - Fixed CMake C++ compiler flags - A recent refactoring of the C++ configure files accidently removed the + A recent refactoring of the C++ configure files accidentally removed the file that executed the enable_language command for C++ needed by the HDFCXXCompilerFlags.cmake file. Also updated the intel warnings files, including adding support for windows platforms. @@ -2382,7 +2382,7 @@ Known Problems e.g. within a single parallel test (testphdf5) the subtests (cdsetr, eidsetr) report data verification errors before eventually aborting. The most recent versions of OpenMPI (v3.1.3 or newer) have evidently - resolved these isses and parallel HDF5 testing does not currently report + resolved these issues and parallel HDF5 testing does not currently report errors though occasional hangs have been observed. Known problems in previous releases can be found in the HISTORY*.txt files @@ -3085,7 +3085,7 @@ Known Problems e.g. within a single parallel test (testphdf5) the subtests (cdsetr, eidsetr) report data verification errors before eventually aborting. The most recent versions of OpenMPI (v3.1.3 or newer) have evidently - resolved these isses and parallel HDF5 testing does not currently report + resolved these issues and parallel HDF5 testing does not currently report errors though occasional hangs have been observed. Known problems in previous releases can be found in the HISTORY*.txt files diff --git a/release_docs/HISTORY-1_8.txt b/release_docs/HISTORY-1_8.txt index 1e31010..461e0be 100644 --- a/release_docs/HISTORY-1_8.txt +++ b/release_docs/HISTORY-1_8.txt @@ -1266,7 +1266,7 @@ New Features (BMR, 2017/05/13, HDFFV-10004, HDFFV-10156) - - New class LinkAccPropList is added for link access proprety list + - New class LinkAccPropList is added for link access property list (BMR, 2017/05/13, HDFFV-10156) @@ -176,7 +176,7 @@ H5_init_library(void) * Install atexit() library cleanup routines unless the H5dont_atexit() * has been called. Once we add something to the atexit() list it stays * there permanently, so we set H5_dont_atexit_g after we add it to prevent - * adding it again later if the library is cosed and reopened. + * adding it again later if the library is closed and reopened. */ if (!H5_dont_atexit_g) { @@ -1193,7 +1193,7 @@ done: if all attributes were processed. DESCRIPTION - This function interates over the attributes of dataset or group + This function iterates over the attributes of dataset or group specified with 'loc_id' & 'obj_name'. For each attribute of the object, the 'op_data' and some additional information (specified below) are passed to the 'op' function. The iteration begins with the '*idx' @@ -1222,7 +1222,7 @@ H5Aiterate2(hid_t loc_id, H5_index_t idx_type, H5_iter_order_t order, hsize_t *i FUNC_ENTER_API(FAIL) H5TRACE6("e", "iIiIo*hx*x", loc_id, idx_type, order, idx, op, op_data); - /* check arguments */ + /* Check arguments */ if (H5I_ATTR == H5I_get_type(loc_id)) HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "location is not valid for an attribute") if (idx_type <= H5_INDEX_UNKNOWN || idx_type >= H5_INDEX_N) @@ -1261,7 +1261,7 @@ done: if all attributes were processed. DESCRIPTION - This function interates over the attributes of dataset or group + This function iterates over the attributes of dataset or group specified with 'loc_id' & 'obj_name'. For each attribute of the object, the 'op_data' and some additional information (specified below) are passed to the 'op' function. The iteration begins with the '*idx' @@ -1273,7 +1273,7 @@ done: * metadata cache flush. * * Initially, this means setting up the slist prior to the - * flush. We do this in a seperate call because + * flush. We do this in a separate call because * H5F__flush_phase2() make repeated calls to H5AC_flush(). * Handling this detail in separate calls allows us to avoid * the overhead of setting up and taking down the skip list @@ -1325,7 +1325,7 @@ done: * flush. * * Initially, this means taking down the slist after the - * flush. We do this in a seperate call because + * flush. We do this in a separate call because * H5F__flush_phase2() make repeated calls to H5AC_flush(). * Handling this detail in separate calls allows us to avoid * the overhead of setting up and taking down the skip list @@ -2480,7 +2480,7 @@ done: /*------------------------------------------------------------------------------ * Function: H5AC_expunge_tag_type_metadata() * - * Purpose: Wrapper for cache level function which expunge entries with + * Purpose: Wrapper for cache level function which expunges entries with * a specific tag and type id. * * Return: SUCCEED on success, FAIL otherwise. diff --git a/src/H5ACdbg.c b/src/H5ACdbg.c index b0d01a8..97b0ba3 100644 --- a/src/H5ACdbg.c +++ b/src/H5ACdbg.c @@ -197,9 +197,9 @@ done: * If either the parent or the child is not in the metadata * cache, the function sets *fd_exists_ptr to FALSE. * - * If both are in the cache, the childs list of parents is + * If both are in the cache, the child's list of parents is * searched for the proposed parent. If the proposed parent - * is found in the childs parent list, the function sets + * is found in the child's parent list, the function sets * *fd_exists_ptr to TRUE. In all other non-error cases, * the function sets *fd_exists_ptr FALSE. * diff --git a/src/H5ACmpio.c b/src/H5ACmpio.c index 0db5cd9..95aba61 100644 --- a/src/H5ACmpio.c +++ b/src/H5ACmpio.c @@ -1140,7 +1140,7 @@ done: * * While the list of candidate cache entries is prepared * elsewhere, this function is the main routine for distributing - * and applying the list. It must be run simultaniously on + * and applying the list. It must be run simultaneously on * all processes that have the relevant file open. To ensure * proper synchronization, there is a barrier at the beginning * of this function. @@ -1319,7 +1319,7 @@ done: * * This function is the main routine for handling this * notification procedure. It must be called - * simultaniously on all processes that have the relevant + * simultaneously on all processes that have the relevant * file open. To this end, it is called only during a * sync point, with a barrier prior to the call. * @@ -1524,7 +1524,7 @@ H5AC__receive_and_apply_clean_list(H5F_t *f) HGOTO_ERROR(H5E_CACHE, H5E_CANTGET, FAIL, "can't receive clean list") if (num_entries > 0) - /* mark the indicated entries as clean */ + /* Mark the indicated entries as clean */ if (H5C_mark_entries_as_clean(f, num_entries, haddr_buf_ptr) < 0) HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "Can't mark entries clean.") diff --git a/src/H5ACpkg.h b/src/H5ACpkg.h index 5289e3a..0ffd5e9 100644 --- a/src/H5ACpkg.h +++ b/src/H5ACpkg.h @@ -124,7 +124,7 @@ H5FL_EXTERN(H5AC_aux_t); * Maintaining this count is easy for all processes not on process 0 -- * all that is necessary is to add the size of the entry to the total * whenever there is an insertion, a move of a previously clean entry, - * or whever a previously clean entry is marked dirty in an unprotect. + * or wherever a previously clean entry is marked dirty in an unprotect. * * On process 0, we have to be careful not to count dirty bytes twice. * If an entry is marked dirty, flushed, and marked dirty again, all @@ -293,12 +293,12 @@ H5FL_EXTERN(H5AC_aux_t); * dirtied flag set and the entry does not already appear * in the dirty entry list. * - * Entries are added to the dirty entry list whever they cause + * Entries are added to the dirty entry list wherever they cause * the dirty bytes count to be increased. They are removed * when they appear in a clean entries broadcast. Note that * moves must be reflected in the dirty entry list. * - * To reitterate, this field is only used on process 0 -- it + * To reiterate, this field is only used on process 0 -- it * should be NULL on all other processes. * * c_slist_ptr: Pointer to an instance of H5SL_t used to maintain a list diff --git a/src/H5Abtree2.c b/src/H5Abtree2.c index 54deb8a..ff9ec73 100644 --- a/src/H5Abtree2.c +++ b/src/H5Abtree2.c @@ -177,7 +177,7 @@ H5A__dense_fh_name_cmp(const void *obj, size_t obj_len, void *_udata) } /* end if */ done: - /* Release the space allocated for the attrbute */ + /* Release the space allocated for the attribute */ if (attr && !took_ownership) H5O_msg_free(H5O_ATTR_ID, attr); diff --git a/src/H5Adense.c b/src/H5Adense.c index 5ec61a2..8f652f0 100644 --- a/src/H5Adense.c +++ b/src/H5Adense.c @@ -341,7 +341,7 @@ H5A__dense_open(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name) H5HF_t *fheap = NULL; /* Fractal heap handle */ H5HF_t *shared_fheap = NULL; /* Fractal heap handle for shared header messages */ H5B2_t *bt2_name = NULL; /* v2 B-tree handle for name index */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ htri_t attr_exists; /* Attribute exists in v2 B-tree */ H5A_t *ret_value = NULL; /* Return value */ @@ -360,7 +360,7 @@ H5A__dense_open(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name) if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, NULL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ @@ -432,7 +432,7 @@ H5A__dense_insert(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr) H5WB_t *wb = NULL; /* Wrapped buffer for attribute data */ uint8_t attr_buf[H5A_ATTR_BUF_SIZE]; /* Buffer for serializing message */ unsigned mesg_flags = 0; /* Flags for storing message */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ herr_t ret_value = SUCCEED; /* Return value */ FUNC_ENTER_PACKAGE @@ -446,7 +446,7 @@ H5A__dense_insert(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr) if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ htri_t shared_mesg; /* Should this message be stored in the Shared Message table? */ @@ -730,7 +730,7 @@ H5A__dense_write(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr) H5HF_t *fheap = NULL; /* Fractal heap handle */ H5HF_t *shared_fheap = NULL; /* Fractal heap handle for shared header messages */ H5B2_t *bt2_name = NULL; /* v2 B-tree handle for name index */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ herr_t ret_value = SUCCEED; /* Return value */ FUNC_ENTER_PACKAGE @@ -746,7 +746,7 @@ H5A__dense_write(H5F_t *f, const H5O_ainfo_t *ainfo, H5A_t *attr) if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ @@ -868,7 +868,7 @@ H5A__dense_rename(H5F_t *f, const H5O_ainfo_t *ainfo, const char *old_name, cons H5B2_t *bt2_name = NULL; /* v2 B-tree handle for name index */ H5B2_t *bt2_corder = NULL; /* v2 B-tree handle for creation order ndex */ H5A_t *attr_copy = NULL; /* Copy of attribute to rename */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ htri_t shared_mesg; /* Should this message be stored in the Shared Message table? */ htri_t attr_exists; /* Attribute exists in v2 B-tree */ herr_t ret_value = SUCCEED; /* Return value */ @@ -885,7 +885,7 @@ H5A__dense_rename(H5F_t *f, const H5O_ainfo_t *ainfo, const char *old_name, cons if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ @@ -1182,7 +1182,7 @@ H5A__dense_iterate(H5F_t *f, hid_t loc_id, const H5O_ainfo_t *ainfo, H5_index_t /* Check on iteration order */ if (order == H5_ITER_NATIVE && H5F_addr_defined(bt2_addr)) { H5A_bt2_ud_it_t udata; /* User data for iterator callback */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ /* Open the fractal heap */ if (NULL == (fheap = H5HF_open(f, ainfo->fheap_addr))) @@ -1192,7 +1192,7 @@ H5A__dense_iterate(H5F_t *f, hid_t loc_id, const H5O_ainfo_t *ainfo, H5_index_t if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ @@ -1339,7 +1339,7 @@ H5A__dense_remove(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name) H5HF_t *shared_fheap = NULL; /* Fractal heap handle for shared header messages */ H5B2_t *bt2_name = NULL; /* v2 B-tree handle for name index */ H5A_t *attr_copy = NULL; /* Copy of attribute to remove */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ herr_t ret_value = SUCCEED; /* Return value */ FUNC_ENTER_PACKAGE @@ -1357,7 +1357,7 @@ H5A__dense_remove(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name) if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ @@ -1586,7 +1586,7 @@ H5A__dense_remove_by_idx(H5F_t *f, const H5O_ainfo_t *ainfo, H5_index_t idx_type /* If there is an index defined for the field, use it */ if (H5F_addr_defined(bt2_addr)) { H5A_bt2_ud_rmbi_t udata; /* User data for v2 B-tree record removal */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ /* Open the fractal heap */ if (NULL == (fheap = H5HF_open(f, ainfo->fheap_addr))) @@ -1596,7 +1596,7 @@ H5A__dense_remove_by_idx(H5F_t *f, const H5O_ainfo_t *ainfo, H5_index_t idx_type if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ @@ -1676,7 +1676,7 @@ H5A__dense_exists(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name) H5HF_t *fheap = NULL; /* Fractal heap handle */ H5HF_t *shared_fheap = NULL; /* Fractal heap handle for shared header messages */ H5B2_t *bt2_name = NULL; /* v2 B-tree handle for name index */ - htri_t attr_sharable; /* Flag indicating attributes are sharable */ + htri_t attr_sharable; /* Flag indicating attributes are shareable */ htri_t ret_value = TRUE; /* Return value */ FUNC_ENTER_PACKAGE @@ -1694,7 +1694,7 @@ H5A__dense_exists(H5F_t *f, const H5O_ainfo_t *ainfo, const char *name) if ((attr_sharable = H5SM_type_shared(f, H5O_ATTR_ID)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_CANTGET, FAIL, "can't determine if attributes are shared") - /* Get handle for shared message heap, if attributes are sharable */ + /* Get handle for shared message heap, if attributes are shareable */ if (attr_sharable) { haddr_t shared_fheap_addr; /* Address of fractal heap to use */ diff --git a/src/H5Adeprec.c b/src/H5Adeprec.c index 6ac66c8..7245890 100644 --- a/src/H5Adeprec.c +++ b/src/H5Adeprec.c @@ -112,7 +112,7 @@ H5Acreate1(hid_t loc_id, const char *name, hid_t type_id, hid_t space_id, hid_t FUNC_ENTER_API(FAIL) H5TRACE5("i", "i*siii", loc_id, name, type_id, space_id, acpl_id); - /* check arguments */ + /* Check arguments */ if (H5I_ATTR == H5I_get_type(loc_id)) HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "location is not valid for an attribute") if (H5G_loc(loc_id, &loc) < 0) @@ -235,7 +235,7 @@ H5Aopen_idx(hid_t loc_id, unsigned idx) FUNC_ENTER_API(FAIL) H5TRACE2("i", "iIu", loc_id, idx); - /* check arguments */ + /* Check arguments */ if (H5I_ATTR == H5I_get_type(loc_id)) HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "location is not valid for an attribute") if (H5G_loc(loc_id, &loc) < 0) @@ -350,7 +350,7 @@ done: last operator if it was non-zero, or zero if all attributes were processed. DESCRIPTION - This function interates over the attributes of dataset or group + This function iterates over the attributes of dataset or group specified with 'loc_id'. For each attribute of the object, the 'op_data' and some additional information (specified below) are passed to the 'op' function. The iteration begins with the '*attr_number' diff --git a/src/H5Aint.c b/src/H5Aint.c index d1b74fa..e5aca7c 100644 --- a/src/H5Aint.c +++ b/src/H5Aint.c @@ -852,7 +852,7 @@ H5A__get_name(H5A_t *attr, size_t buf_size, char *buf) nbytes = HDstrlen(attr->shared->name); HDassert((ssize_t)nbytes >= 0); /*overflow, pretty unlikely --rpm*/ - /* compute the string length which will fit into the user's buffer */ + /* Compute the string length which will fit into the user's buffer */ copy_len = MIN(buf_size - 1, nbytes); /* Copy all/some of the name */ @@ -960,13 +960,10 @@ done: NAME H5A__get_create_plist PURPOSE - private version of H5Aget_create_plist + Private version of H5Aget_create_plist RETURNS This function returns the ID of a copy of the attribute's creation property list, or negative on failure. - - ERRORS - DESCRIPTION This function returns a copy of the creation property list for an attribute. The resulting ID must be closed with H5Pclose() or @@ -2173,7 +2170,7 @@ H5A__attr_copy_file(const H5A_t *attr_src, H5F_t *file_dst, hbool_t *recompute_s if ((tid_mem = H5I_register(H5I_DATATYPE, dt_mem, FALSE)) < 0) HGOTO_ERROR(H5E_DATATYPE, H5E_CANTREGISTER, NULL, "unable to register memory datatype") - /* create variable-length datatype at the destinaton file */ + /* create variable-length datatype at the destination file */ if ((tid_dst = H5I_register(H5I_DATATYPE, attr_dst->shared->dt, FALSE)) < 0) HGOTO_ERROR(H5E_DATATYPE, H5E_CANTREGISTER, NULL, "unable to register destination file datatype") diff --git a/src/H5Amodule.h b/src/H5Amodule.h index e3bfe6f..0c31f71 100644 --- a/src/H5Amodule.h +++ b/src/H5Amodule.h @@ -364,7 +364,7 @@ * will be ignored by HDF5. * * The use of ASCII or UTF-8 characters is determined by the character encoding property. See - * #H5Pset_char_encoding in the \ref RM. + * #H5Pset_char_encoding in the \ref RM. * * <h4>No Special I/O or Storage</h4> * @@ -1457,7 +1457,7 @@ H5B__remove_helper(H5F_t *f, haddr_t addr, const H5B_class_t *type, int level, u */ if (type->critical_key == H5B_LEFT) /* Slide the rightmost key down one, overwriting the left key of - * the deleted (righmost) child */ + * the deleted (rightmost) child */ HDmemmove(H5B_NKEY(bt, shared, bt->nchildren - 1), H5B_NKEY(bt, shared, bt->nchildren), type->sizeof_nkey); else { diff --git a/src/H5B2internal.c b/src/H5B2internal.c index 6836aba..13c4fa1 100644 --- a/src/H5B2internal.c +++ b/src/H5B2internal.c @@ -17,7 +17,7 @@ * Dec 01 2016 * Quincey Koziol * - * Purpose: Routines for managing v2 B-tree internal ndoes. + * Purpose: Routines for managing v2 B-tree internal nodes. * *------------------------------------------------------------------------- */ diff --git a/src/H5B2leaf.c b/src/H5B2leaf.c index 3603fc3..6f73ce8 100644 --- a/src/H5B2leaf.c +++ b/src/H5B2leaf.c @@ -17,7 +17,7 @@ * Dec 01 2016 * Quincey Koziol * - * Purpose: Routines for managing v2 B-tree leaf ndoes. + * Purpose: Routines for managing v2 B-tree leaf nodes. * *------------------------------------------------------------------------- */ @@ -389,7 +389,7 @@ H5C_create(size_t max_cache_size, size_t min_clean_size, int max_type_id, cache_ptr->slist_size = (size_t)0; /* slist_ring_len, slist_ring_size, and - * slist_ptr initializaed above. + * slist_ptr initialized above. */ #if H5C_DO_SANITY_CHECKS @@ -798,7 +798,7 @@ H5C_prep_for_file_close(H5F_t *f) * * 2) Since the FSM settle routines are only invoked once during * file close, invoking them now will prevent their invocation - * during a flush, and thus avoid any resulting entrie dirties, + * during a flush, and thus avoid any resulting entry dirties, * deletions, insertion, or moves during the flush. */ if (H5C__serialize_cache(f) < 0) @@ -1505,7 +1505,7 @@ H5C_insert_entry(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *thing, u * Finally, we usually don't check to see if the cache is * oversized at the end of an unprotect. As a result, it is * possible to have a vastly oversized cache with no protected - * entries as long as all the protects preceed the unprotects. + * entries as long as all the protects precede the unprotects. * * Since items 1 and 2 are not changing any time soon, I see * no point in worrying about the third. @@ -2477,7 +2477,7 @@ H5C_protect(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *udata, unsign * Finally, we usually don't check to see if the cache is * oversized at the end of an unprotect. As a result, it is * possible to have a vastly oversized cache with no protected - * entries as long as all the protects preceed the unprotects. + * entries as long as all the protects precede the unprotects. * * Since items 1, 2, and 3 are not changing any time soon, I * see no point in worrying about the fourth. @@ -2592,7 +2592,7 @@ H5C_protect(H5F_t *f, const H5C_class_t *type, haddr_t addr, void *udata, unsign * * Also, if the min_clean_size requirement is not met, we * should also call H5C__make_space_in_cache() to bring us - * into complience. + * into compliance. */ if (cache_ptr->index_size >= cache_ptr->max_cache_size) @@ -3010,7 +3010,7 @@ H5C_set_slist_enabled(H5C_t *cache_ptr, hbool_t slist_enabled, hbool_t clear_sli } /* set cache_ptr->slist_enabled to TRUE so that the slist - * mainenance macros will be enabled. + * maintenance macros will be enabled. */ cache_ptr->slist_enabled = TRUE; @@ -4116,7 +4116,7 @@ done: "can't notify parent about child entry serialized flag set") } /* end if */ - /* Shrink or free the parent array if apporpriate */ + /* Shrink or free the parent array if appropriate */ if (child_entry->flush_dep_nparents == 0) { child_entry->flush_dep_parent = H5FL_SEQ_FREE(H5C_cache_entry_ptr_t, child_entry->flush_dep_parent); @@ -4883,7 +4883,7 @@ done: /* for now at least, don't bother to maintain the minimum clean size, * as the cache should now be less than its maximum size. Due to - * the vaguries of the cache size reduction algorthim, we may not + * the vaguries of the cache size reduction algorithm, we may not * reduce the size of the cache. * * If we do, we will calculate a new minimum clean size, which will @@ -5892,7 +5892,7 @@ done: * If either of these happen, and one of the target * or proxy entries happens to be the next entry in * the hash bucket, we could either find ourselves - * either scanning a non-existant entry, scanning + * either scanning a non-existent entry, scanning * through a different bucket, or skipping an entry. * * Neither of these are good, so restart the @@ -8202,7 +8202,7 @@ done: * To cork the object * Return error if the object is already corked * H5C__UNCORK: - * To uncork the obejct + * To uncork the object * Return error if the object is not corked * H5C__GET_CORKED: * To retrieve the cork status of an object in @@ -8627,7 +8627,7 @@ done: * each entry in the cache to zero before we start the serialization. * This allows us to detect the case in which any entry is serialized * more than once (a performance issues), and more importantly, the - * case is which any flush depencency parent is serializes more than + * case is which any flush dependency parent is serializes more than * once (a correctness issue). */ { @@ -8783,8 +8783,8 @@ done: * are serialized correctly, it doesn't matter if we have to go back * and serialize an entry a second time. * - * These possible actions result in the following modfications to - * tha basic algorithm: + * These possible actions result in the following modifications to + * the basic algorithm: * * 1) In the event of an entry expunge, eviction or removal, we must * restart the scan as it is possible that the next entry in our @@ -8834,7 +8834,7 @@ done: * condition appears. * * Observe that either eviction or removal of entries as a result of - * a serialization is not a problem as long as the flush depencency + * a serialization is not a problem as long as the flush dependency * tree does not change beyond the removal of a leaf. */ while (!done) { @@ -47,7 +47,7 @@ typedef struct H5CS_t { * each thread individually. The association of stacks to threads will * be handled by the pthread library. * - * In order for this macro to work, H5CS_get_my_stack() must be preceeded + * In order for this macro to work, H5CS_get_my_stack() must be preceded * by "H5CS_t *fstack =". */ static H5CS_t *H5CS__get_stack(void); @@ -127,7 +127,7 @@ H5CS__get_stack(void) herr_t H5CS_print_stack(const H5CS_t *fstack, FILE *stream) { - const int indent = 2; /* Indention level */ + const int indent = 2; /* Indentation level */ int i; /* Local index ariable */ /* Don't push this function on the function stack... :-) */ @@ -50,7 +50,7 @@ * each thread individually. The association of contexts to threads will * be handled by the pthread library. * - * In order for this macro to work, H5CX_get_my_context() must be preceeded + * In order for this macro to work, H5CX_get_my_context() must be preceded * by "H5CX_node_t *ctx =". */ #define H5CX_get_my_context() H5CX__get_context() @@ -237,7 +237,7 @@ typedef struct H5CX_t { H5FD_mpio_chunk_opt_t mpio_chunk_opt_mode; /* Collective chunk option (H5D_XFER_MPIO_CHUNK_OPT_HARD_NAME) */ hbool_t mpio_chunk_opt_mode_valid; /* Whether collective chunk option is valid */ - unsigned mpio_chunk_opt_num; /* Collective chunk thrreshold (H5D_XFER_MPIO_CHUNK_OPT_NUM_NAME) */ + unsigned mpio_chunk_opt_num; /* Collective chunk threshold (H5D_XFER_MPIO_CHUNK_OPT_NUM_NAME) */ hbool_t mpio_chunk_opt_num_valid; /* Whether collective chunk threshold is valid */ unsigned mpio_chunk_opt_ratio; /* Collective chunk ratio (H5D_XFER_MPIO_CHUNK_OPT_RATIO_NAME) */ hbool_t mpio_chunk_opt_ratio_valid; /* Whether collective chunk ratio is valid */ @@ -365,7 +365,7 @@ typedef struct H5CX_dxpl_cache_t { (H5D_MPIO_GLOBAL_NO_COLLECTIVE_CAUSE_NAME) */ H5FD_mpio_chunk_opt_t mpio_chunk_opt_mode; /* Collective chunk option (H5D_XFER_MPIO_CHUNK_OPT_HARD_NAME) */ - unsigned mpio_chunk_opt_num; /* Collective chunk thrreshold (H5D_XFER_MPIO_CHUNK_OPT_NUM_NAME) */ + unsigned mpio_chunk_opt_num; /* Collective chunk threshold (H5D_XFER_MPIO_CHUNK_OPT_NUM_NAME) */ unsigned mpio_chunk_opt_ratio; /* Collective chunk ratio (H5D_XFER_MPIO_CHUNK_OPT_RATIO_NAME) */ #endif /* H5_HAVE_PARALLEL */ H5Z_EDC_t err_detect; /* Error detection info (H5D_XFER_EDC_NAME) */ diff --git a/src/H5Cdbg.c b/src/H5Cdbg.c index a82d018..ec66fb1 100644 --- a/src/H5Cdbg.c +++ b/src/H5Cdbg.c @@ -915,9 +915,9 @@ H5C__dump_entry(H5C_t *cache_ptr, const H5C_cache_entry_t *entry_ptr, hbool_t du * If either the parent or the child is not in the metadata * cache, the function sets *fd_exists_ptr to FALSE. * - * If both are in the cache, the childs list of parents is + * If both are in the cache, the child's list of parents is * searched for the proposed parent. If the proposed parent - * is found in the childs parent list, the function sets + * is found in the child's parent list, the function sets * *fd_exists_ptr to TRUE. In all other non-error cases, * the function sets *fd_exists_ptr FALSE. * diff --git a/src/H5Cimage.c b/src/H5Cimage.c index 18e35f3..dddc89c 100644 --- a/src/H5Cimage.c +++ b/src/H5Cimage.c @@ -591,7 +591,7 @@ H5C__deserialize_prefetched_entry(H5F_t *f, H5C_t *cache_ptr, H5C_cache_entry_t * * Note that at present, dirty can't be set to true with prefetched * entries. However this may change, so include this functionality - * against that posibility. + * against that possibility. * * Also, note that it is possible for a prefetched entry to be dirty -- * hence the value assigned to ds_entry_ptr->is_dirty below. @@ -1179,7 +1179,7 @@ done: * image superblock extension message must be deleted from * the superblock extension and the image block freed * - * Contrawise, if the file is openened R/O, the metadata + * Contrawise, if the file is opened R/O, the metadata * cache image superblock extension message and image block * must be left as is. Further, any dirty entries in the * cache image block must be marked as clean to avoid @@ -1872,7 +1872,7 @@ done: * Purpose: Decode the metadata cache image entry from the supplied * buffer into the supplied instance of H5C_image_entry_t. * This includes allocating a buffer for the entry image, - * loading it, and seting ie_ptr->image_ptr to point to + * loading it, and setting ie_ptr->image_ptr to point to * the buffer. * * Advances the buffer pointer to the first byte @@ -2389,7 +2389,7 @@ done: * also be a flush dependency child. * * Finally, note that for purposes of the cache image, flush - * dependency height ends when a flush dependecy relation + * dependency height ends when a flush dependency relation * passes off the cache image. * * On exit, the flush dependency height of each entry in the @@ -2551,7 +2551,7 @@ H5C__prep_for_file_close__compute_fd_heights(const H5C_t *cache_ptr) entry_ptr = entry_ptr->il_next; } /* while (entry_ptr != NULL) */ - /* At present, no extenal parent or child flush dependency links + /* At present, no external parent or child flush dependency links * should exist -- hence the following assertions. This will change * if we support ageout of entries in the cache image. */ @@ -3433,7 +3433,7 @@ done: * creating if specified. * * In general, the size and location of the cache image block - * will be unknow at the time that the cache image superblock + * will be unknown at the time that the cache image superblock * message is created. A subsequent call to this routine will * be used to write the correct data. * diff --git a/src/H5Cmpio.c b/src/H5Cmpio.c index 7989794..be6a842 100644 --- a/src/H5Cmpio.c +++ b/src/H5Cmpio.c @@ -1105,7 +1105,7 @@ done: * * Note that this function is a modified version of * H5C_flush_cache() -- any changes there may need to be - * reflected here and vise versa. + * reflected here and vice versa. * * Return: Non-negative on success/Negative on failure. * diff --git a/src/H5Cpkg.h b/src/H5Cpkg.h index 75b39d1..e7d997e 100644 --- a/src/H5Cpkg.h +++ b/src/H5Cpkg.h @@ -4726,7 +4726,7 @@ typedef struct H5C_tag_info_t { * * Fields for tracking prefetched entries. Note that flushes and evictions * of prefetched entries are tracked in the flushes and evictions arrays - * discused above. + * discussed above. * * prefetches: Number of prefetched entries that are loaded to the * cache. diff --git a/src/H5Cprefetched.c b/src/H5Cprefetched.c index fa5d82e..c652aaf 100644 --- a/src/H5Cprefetched.c +++ b/src/H5Cprefetched.c @@ -302,7 +302,7 @@ H5C__prefetched_entry_free_icr(void *_thing) HDassert(entry_ptr->fd_parent_count == 0); if (entry_ptr->image_ptr != NULL) - HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "prefetched entry image buffer still attatched?") + HGOTO_ERROR(H5E_CACHE, H5E_SYSTEM, FAIL, "prefetched entry image buffer still attached?") entry_ptr = H5FL_FREE(H5C_cache_entry_t, entry_ptr); diff --git a/src/H5Cprivate.h b/src/H5Cprivate.h index 9aa53a3..e513850 100644 --- a/src/H5Cprivate.h +++ b/src/H5Cprivate.h @@ -924,7 +924,7 @@ typedef herr_t (*H5C_log_flush_func_t)(H5C_t *cache_ptr, haddr_t addr, hbool_t w * instead would apply unnecessary constraints on flushes under normal * operating circumstances. * - * As of this writing, all metadata entries pretaining to data sets and + * As of this writing, all metadata entries pertaining to data sets and * groups must be flushed first, and are thus assigned to the outermost * ring. * @@ -2161,7 +2161,7 @@ typedef struct H5C_auto_size_ctl_t { * H5AC__CACHE_IMAGE__ENTRY_AGEOUT__MAX (100). * * H5AC__CACHE_IMAGE__ENTRY_AGEOUT__NONE means that no limit - * is imposed on number of times a prefeteched entry can appear + * is imposed on number of times a prefetched entry can appear * in subsequent cache images. * * A value of 0 prevents prefetched entries from being included diff --git a/src/H5Dchunk.c b/src/H5Dchunk.c index f08e664..e9ca0fc 100644 --- a/src/H5Dchunk.c +++ b/src/H5Dchunk.c @@ -966,7 +966,7 @@ H5D__chunk_init(H5F_t *f, const H5D_t *const dset, hid_t dapl_id) if (!(scaled_power2up = H5VM_power2up(rdcc->scaled_dims[u]))) HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "unable to get the next power of 2") - /* Inital 'power2up' values for scaled dimensions */ + /* Initial 'power2up' values for scaled dimensions */ rdcc->scaled_power2up[u] = scaled_power2up; /* Number of bits required to encode scaled dimension size */ @@ -1030,7 +1030,7 @@ H5D__chunk_is_space_alloc(const H5O_storage_t *storage) * Return: Non-negative on success/Negative on failure * * Programmer: Neil Fortner - * Wednessday, March 6, 2016 + * Wednesday, March 6, 2016 * *------------------------------------------------------------------------- */ @@ -1633,7 +1633,7 @@ H5D__create_chunk_file_map_all(H5D_chunk_map_t *fm, const H5D_io_info_t coords[u] = 0; end[u] = fm->chunk_dim[u] - 1; - /* Iniitialize partial chunk dimension information */ + /* Initialize partial chunk dimension information */ partial_dim_size[u] = file_dims[u] % fm->chunk_dim[u]; if (file_dims[u] < fm->chunk_dim[u]) { curr_partial_clip[u] = partial_dim_size[u]; @@ -2537,7 +2537,7 @@ H5D__chunk_read(H5D_io_info_t *io_info, const H5D_type_info_t *type_info, hsize_ HDassert((H5F_addr_defined(udata.chunk_block.offset) && udata.chunk_block.length > 0) || (!H5F_addr_defined(udata.chunk_block.offset) && udata.chunk_block.length == 0)); - /* Check for non-existant chunk & skip it if appropriate */ + /* Check for non-existnt chunk & skip it if appropriate */ if (H5F_addr_defined(udata.chunk_block.offset) || UINT_MAX != udata.idx_hint || !skip_missing_chunks) { H5D_io_info_t *chk_io_info; /* Pointer to I/O info object for this chunk */ @@ -3604,7 +3604,7 @@ H5D__chunk_cache_prune(const H5D_t *dset, size_t size) * traversing the list when pointer pN reaches wN percent of the original * list. In other words, preemption method N gets to consider entries in * approximate least recently used order w0 percent before method N+1 - * where 100% means tha method N will run to completion before method N+1 + * where 100% means the method N will run to completion before method N+1 * begins. The pointers participating in the list traversal are each * given a chance at preemption before any of the pointers are advanced. */ @@ -6039,7 +6039,7 @@ H5D__chunk_copy_cb(const H5D_chunk_rec_t *chunk_rec, void *_udata) size_t buf_size = udata->buf_size; /* Size of chunk buffer */ const H5O_pline_t *pline = udata->pline; /* I/O pipeline for applying filters */ - /* needed for commpressed variable length data */ + /* needed for compressed variable length data */ hbool_t must_filter = FALSE; /* Whether chunk must be filtered during copy */ size_t nbytes; /* Size of chunk in file (in bytes) */ H5Z_cb_t filter_cb; /* Filter failure callback struct */ @@ -6383,7 +6383,7 @@ H5D__chunk_copy(H5F_t *f_src, H5O_storage_chunk_t *storage_src, H5O_layout_chunk HGOTO_ERROR(H5E_DATATYPE, H5E_CANTREGISTER, FAIL, "unable to register memory datatype") } /* end if */ - /* create variable-length datatype at the destinaton file */ + /* create variable-length datatype at the destination file */ if (NULL == (dt_dst = H5T_copy(dt_src, H5T_COPY_TRANSIENT))) HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "unable to copy") if (H5T_set_loc(dt_dst, f_dst, H5T_LOC_DISK) < 0) { diff --git a/src/H5Dcompact.c b/src/H5Dcompact.c index f48c7e1..dfe199b 100644 --- a/src/H5Dcompact.c +++ b/src/H5Dcompact.c @@ -461,7 +461,7 @@ H5D__compact_copy(H5F_t *f_src, H5O_storage_compact_t *_storage_src, H5F_t *f_ds HGOTO_ERROR(H5E_DATATYPE, H5E_CANTREGISTER, FAIL, "unable to register memory datatype") } /* end if */ - /* create variable-length datatype at the destinaton file */ + /* create variable-length datatype at the destination file */ if (NULL == (dt_dst = H5T_copy(dt_src, H5T_COPY_TRANSIENT))) HGOTO_ERROR(H5E_DATATYPE, H5E_CANTINIT, FAIL, "unable to copy") if (H5T_set_loc(dt_dst, f_dst, H5T_LOC_DISK) < 0) { diff --git a/src/H5Dcontig.c b/src/H5Dcontig.c index 7a87699..19476ba2 100644 --- a/src/H5Dcontig.c +++ b/src/H5Dcontig.c @@ -538,7 +538,7 @@ H5D__contig_is_space_alloc(const H5O_storage_t *storage) * Return: Non-negative on success/Negative on failure * * Programmer: Neil Fortner - * Wednessday, March 6, 2016 + * Wednesday, March 6, 2016 * *------------------------------------------------------------------------- */ @@ -1404,7 +1404,7 @@ H5D__contig_copy(H5F_t *f_src, const H5O_storage_contig_t *storage_src, H5F_t *f HGOTO_ERROR(H5E_DATASET, H5E_CANTREGISTER, FAIL, "unable to register memory datatype") } /* end if */ - /* create variable-length datatype at the destinaton file */ + /* create variable-length datatype at the destination file */ if (NULL == (dt_dst = H5T_copy(dt_src, H5T_COPY_TRANSIENT))) HGOTO_ERROR(H5E_DATASET, H5E_CANTINIT, FAIL, "unable to copy") if (H5T_set_loc(dt_dst, f_dst, H5T_LOC_DISK) < 0) { diff --git a/src/H5Dint.c b/src/H5Dint.c index 16db10d..f9e7953 100644 --- a/src/H5Dint.c +++ b/src/H5Dint.c @@ -343,7 +343,7 @@ done: * * Return: * Success: Non-negative - * Failture: Negative + * Failure: Negative *------------------------------------------------------------------------- */ herr_t @@ -671,7 +671,7 @@ H5D__calculate_minimum_header_size(H5F_t *file, H5D_t *dset, H5O_t *ohdr) H5T_t *type = NULL; H5O_fill_t *fill_prop = NULL; hbool_t use_at_least_v18 = FALSE; - const char continuation[1] = ""; /* requred for work-around */ + const char continuation[1] = ""; /* required for work-around */ size_t get_value = 0; size_t ret_value = 0; @@ -718,7 +718,7 @@ H5D__calculate_minimum_header_size(H5F_t *file, H5D_t *dset, H5O_t *ohdr) HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, 0, "can't get size of continuation message") ret_value += get_value; - /* Fill Value (backwards compatability) message size */ + /* Fill Value (backwards compatibility) message size */ if (fill_prop->buf && !use_at_least_v18) { H5O_fill_t old_fill_prop; /* Copy for writing "old" fill value */ @@ -807,7 +807,7 @@ H5D__prepare_minimized_oh(H5F_t *file, H5D_t *dset, H5O_loc_t *oloc) if (ohdr_size == 0) HGOTO_ERROR(H5E_OHDR, H5E_BADVALUE, FAIL, "computed header size is invalid") - /* Special allocation of space for compact datsets is handled by the call here. */ + /* Special allocation of space for compact datasets is handled by the call here. */ if (H5O__apply_ohdr(file, oh, dset->shared->dcpl_id, ohdr_size, (size_t)1, oloc) == FAIL) HGOTO_ERROR(H5E_OHDR, H5E_BADVALUE, FAIL, "can't apply object header to file") @@ -2916,7 +2916,7 @@ H5D__set_extent(H5D_t *dset, const hsize_t *size) } /*------------------------------------------------------------------------- * Remove chunk information in the case of chunked datasets - * This removal takes place only in case we are shrinking the dateset + * This removal takes place only in case we are shrinking the dataset * and if the chunks are written *------------------------------------------------------------------------- */ diff --git a/src/H5Dmpio.c b/src/H5Dmpio.c index c136af6..6ae3f12 100644 --- a/src/H5Dmpio.c +++ b/src/H5Dmpio.c @@ -664,7 +664,7 @@ H5D__mpio_opt_possible(const H5D_io_info_t *io_info, const H5S_t *file_space, co /* Check to see if the process is reading the entire dataset */ if (H5S_GET_SELECT_TYPE(file_space) != H5S_SEL_ALL) local_cause[1] |= H5D_MPIO_RANK0_NOT_H5S_ALL; - /* Only perform this optimization for contigous datasets, currently */ + /* Only perform this optimization for contiguous datasets, currently */ else if (H5D_CONTIGUOUS != io_info->dset->shared->layout.type) /* Flag to do a MPI_Bcast of the data from one proc instead of * having all the processes involved in the collective I/O. @@ -1329,9 +1329,9 @@ H5D__link_chunk_collective_io(H5D_io_info_t *io_info, const H5D_type_info_t *typ int sum_chunk, int mpi_rank, int mpi_size) { H5D_chunk_addr_info_t *chunk_addr_info_array = NULL; - MPI_Datatype chunk_final_mtype; /* Final memory MPI datatype for all chunks with seletion */ + MPI_Datatype chunk_final_mtype; /* Final memory MPI datatype for all chunks with selection */ hbool_t chunk_final_mtype_is_derived = FALSE; - MPI_Datatype chunk_final_ftype; /* Final file MPI datatype for all chunks with seletion */ + MPI_Datatype chunk_final_ftype; /* Final file MPI datatype for all chunks with selection */ hbool_t chunk_final_ftype_is_derived = FALSE; H5D_storage_t ctg_store; /* Storage info for "fake" contiguous dataset */ size_t total_chunks; @@ -2895,7 +2895,7 @@ done: * * 1) Each process provides two piece of information for all chunks having selection * a) chunk index - * b) wheather this chunk is regular(for MPI derived datatype not working case) + * b) whether this chunk is regular(for MPI derived datatype not working case) * * 2) Gather all the information to the root process * diff --git a/src/H5Dvirtual.c b/src/H5Dvirtual.c index 28a49b1..89dc630 100644 --- a/src/H5Dvirtual.c +++ b/src/H5Dvirtual.c @@ -205,7 +205,7 @@ H5D_virtual_check_mapping_pre(const H5S_t *vspace, const H5S_t *src_space, "can't get number of elements in non-unlimited dimension") if (nenu_vs != nenu_ss) HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, - "numbers of elemnts in the non-unlimited dimensions is different for source and " + "numbers of elements in the non-unlimited dimensions is different for source and " "virtual spaces") } /* end if */ /* We will handle the printf case after parsing the source names */ @@ -1492,7 +1492,7 @@ H5D__virtual_set_extent_unlim(const H5D_t *dset) storage->list[i].source_dset.dset->shared->space) < 0) HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, FAIL, "can't copy source dataspace extent") - /* Get source space dimenstions */ + /* Get source space dimensions */ if (H5S_get_simple_extent_dims(storage->list[i].source_select, curr_dims, NULL) < 0) HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get source space dimensions") @@ -1970,7 +1970,7 @@ H5D__virtual_init_all(const H5D_t *dset) storage->list[i].source_dset.dset->shared->space) < 0) HGOTO_ERROR(H5E_DATASET, H5E_CANTCOPY, FAIL, "can't copy source dataspace extent") - /* Get source space dimenstions */ + /* Get source space dimensions */ if (H5S_get_simple_extent_dims(storage->list[i].source_select, source_dims, NULL) < 0) HGOTO_ERROR(H5E_DATASET, H5E_CANTGET, FAIL, "can't get source space dimensions") @@ -2331,7 +2331,7 @@ H5D__virtual_is_space_alloc(const H5O_storage_t H5_ATTR_UNUSED *storage) * Return: Non-negative on success/Negative on failure * * Programmer: Neil Fortner - * Wednessday, March 6, 2016 + * Wednesday, March 6, 2016 * *------------------------------------------------------------------------- */ @@ -2354,7 +2354,7 @@ H5D__virtual_is_data_cached(const H5D_shared_t *shared_dset) if (storage->list[i].psfn_nsubs || storage->list[i].psdn_nsubs) { /* Iterate over sub-source dsets */ for (j = storage->list[i].sub_dset_io_start; j < storage->list[i].sub_dset_io_end; j++) - /* Check for cahced data in source dset */ + /* Check for cached data in source dset */ if (storage->list[i].sub_dset[j].dset && storage->list[i].sub_dset[j].dset->shared->layout.ops->is_data_cached && storage->list[i].sub_dset[j].dset->shared->layout.ops->is_data_cached( @@ -2688,7 +2688,7 @@ H5D__virtual_post_io(H5O_storage_virtual_t *storage) /*------------------------------------------------------------------------- * Function: H5D__virtual_read_one * - * Purpose: Read from a singe source dataset in a virtual dataset. + * Purpose: Read from a single source dataset in a virtual dataset. * * Return: Non-negative on success/Negative on failure * @@ -2878,7 +2878,7 @@ done: /*------------------------------------------------------------------------- * Function: H5D__virtual_write_one * - * Purpose: Write to a singe source dataset in a virtual dataset. + * Purpose: Write to a single source dataset in a virtual dataset. * * Return: Non-negative on success/Negative on failure * @@ -706,7 +706,7 @@ done: /*------------------------------------------------------------------------- * Function: H5E__close_msg * - * Purpose: Private function to close an error messge. + * Purpose: Private function to close an error message. * * Return: SUCCEED/FAIL * @@ -95,7 +95,7 @@ H5FL_BLK_DEFINE(fa_native_elmt); /*------------------------------------------------------------------------- * Function: H5FA__new * - * Purpose: Allocate and initialize a new fixe array wrapper in memory + * Purpose: Allocate and initialize a new fixed array wrapper in memory * * Return: Pointer to farray wrapper success * NULL on failure diff --git a/src/H5FAcache.c b/src/H5FAcache.c index a4c5f74..987b29d 100644 --- a/src/H5FAcache.c +++ b/src/H5FAcache.c @@ -943,7 +943,7 @@ done: * file space block set. * * This function is needed when the data block is paged, as - * the datablock header and all its pages are allocted as a + * the datablock header and all its pages are allocated as a * single contiguous chunk of file space, and must be * deallocated the same way. * @@ -1324,7 +1324,7 @@ H5FD_get_fs_type_map(const H5FD_t *file, H5FD_mem_t *type_map) HDassert(file->cls); HDassert(type_map); - /* Check for VFD class providing a type map retrieval rouine */ + /* Check for VFD class providing a type map retrieval routine */ if (file->cls->get_type_map) { /* Retrieve type mapping for this file */ if ((file->cls->get_type_map)(file, type_map) < 0) diff --git a/src/H5FDcore.c b/src/H5FDcore.c index db2b546..8ef3532 100644 --- a/src/H5FDcore.c +++ b/src/H5FDcore.c @@ -1411,7 +1411,7 @@ done: * Addendum -- 12/2/11 * For file images opened with the core file driver, it is * necessary that we avoid reallocating the core file driver's - * buffer uneccessarily. + * buffer unnecessarily. * * To this end, I have made the following functional changes * to this function. diff --git a/src/H5FDhdfs.c b/src/H5FDhdfs.c index d653b19..4c0dd0c 100644 --- a/src/H5FDhdfs.c +++ b/src/H5FDhdfs.c @@ -567,16 +567,16 @@ done: * Function: H5FD__hdfs_validate_config() * * Purpose: Test to see if the supplied instance of H5FD_hdfs_fapl_t - * contains internally consistant data. Return SUCCEED if so, + * contains internally consistent data. Return SUCCEED if so, * and FAIL otherwise. * - * Note the difference between internally consistant and + * Note the difference between internally consistent and * correct. As we will have to try to access the target * object to determine whether the supplied data is correct, - * we will settle for internal consistancy at this point + * we will settle for internal consistency at this point * * Return: SUCCEED if instance of H5FD_hdfs_fapl_t contains internally - * consistant data, FAIL otherwise. + * consistent data, FAIL otherwise. * * Programmer: Jacob Smith * 9/10/17 diff --git a/src/H5FDlog.c b/src/H5FDlog.c index bf6e387..49fd26e 100644 --- a/src/H5FDlog.c +++ b/src/H5FDlog.c @@ -1559,7 +1559,7 @@ H5FD__log_write(H5FD_t *_file, H5FD_mem_t type, hid_t H5_ATTR_UNUSED dxpl_id, ha HDfprintf(file->logfp, "%10" PRIuHADDR "-%10" PRIuHADDR " (%10zu bytes) (%s) Written", orig_addr, (orig_addr + orig_size) - 1, orig_size, flavors[type]); - /* Check if this is the first write into a "default" section, grabbed by the metadata agregation + /* Check if this is the first write into a "default" section, grabbed by the metadata aggregation * algorithm */ if (file->fa.flags & H5FD_LOG_FLAVOR) { if ((H5FD_mem_t)file->flavor[orig_addr] == H5FD_MEM_DEFAULT) { diff --git a/src/H5FDmirror.c b/src/H5FDmirror.c index d43c404..4b9991f 100644 --- a/src/H5FDmirror.c +++ b/src/H5FDmirror.c @@ -335,9 +335,9 @@ H5FD__mirror_xmit_decode_uint32(uint32_t *out, const unsigned char *_buf) /* --------------------------------------------------------------------------- * Function: is_host_little_endian * - * Purpose: Determine whether the host machine is is little-endian. + * Purpose: Determine whether the host machine is little-endian. * - * Store an intger with a known value, re-map the memory to a + * Store an integer with a known value, re-map the memory to a * character array, and inspect the array's contents. * * Return: The number of bytes written to the buffer (8). @@ -1568,7 +1568,7 @@ H5FD__mirror_query(const H5FD_t H5_ATTR_UNUSED *_file, unsigned long *flags) LOG_OP_CALL(FUNC); /* Notice: the Mirror VFD Writer currently uses only the Sec2 driver as - * the underying driver -- as such, the Mirror VFD implementation copies + * the underlying driver -- as such, the Mirror VFD implementation copies * the Sec2 feature flags as its own. * * File pointer is always NULL/unused -- the H5FD_FEAT_IGNORE_DRVRINFO flag diff --git a/src/H5FDmpio.c b/src/H5FDmpio.c index ca0a2ea..a7c904b 100644 --- a/src/H5FDmpio.c +++ b/src/H5FDmpio.c @@ -1840,11 +1840,11 @@ done: * easy for us to track the current EOF by extracting it from * write calls. * - * Instead, we first check to see if the eoa has changed since + * Instead, we first check to see if the EOA has changed since * the last call to this function. If it has, we call * MPI_File_get_size() to determine the current EOF, and * only call MPI_File_set_size() if this value disagrees - * with the current eoa. + * with the current EOA. * * Return: SUCCEED/FAIL * @@ -1884,7 +1884,7 @@ H5FD__mpio_truncate(H5FD_t *_file, hid_t H5_ATTR_UNUSED dxpl_id, hbool_t H5_ATTR * call. * * In practice, most (all?) truncate calls will come after a barrier - * and with no interviening writes to the file (with the possible + * and with no intervening writes to the file (with the possible * exception of sueprblock / superblock extension message updates). * * Check the "MPI file closing" flag in the API context to determine @@ -1911,7 +1911,7 @@ H5FD__mpio_truncate(H5FD_t *_file, hid_t H5_ATTR_UNUSED dxpl_id, hbool_t H5_ATTR if (H5FD_mpi_haddr_to_MPIOff(file->eoa, &needed_eof) < 0) HGOTO_ERROR(H5E_INTERNAL, H5E_BADRANGE, FAIL, "cannot convert from haddr_t to MPI_Offset") - /* eoa != eof. Set eof to eoa */ + /* EOA != EOF. Set EOF to EOA */ if (size != needed_eof) { /* Extend the file's size */ if (MPI_SUCCESS != (mpi_code = MPI_File_set_size(file->f, needed_eof))) diff --git a/src/H5FDros3.c b/src/H5FDros3.c index 90c5798..18d602b 100644 --- a/src/H5FDros3.c +++ b/src/H5FDros3.c @@ -379,7 +379,7 @@ H5FD__ros3_term(void) * Function: H5Pset_fapl_ros3 * * Purpose: Modify the file access property list to use the H5FD_ROS3 - * driver defined in this source file. All driver specfic + * driver defined in this source file. All driver specific * properties are passed in as a pointer to a suitably * initialized instance of H5FD_ros3_fapl_t * @@ -422,16 +422,16 @@ done: * Function: H5FD__ros3_validate_config() * * Purpose: Test to see if the supplied instance of H5FD_ros3_fapl_t - * contains internally consistant data. Return SUCCEED if so, + * contains internally consistent data. Return SUCCEED if so, * and FAIL otherwise. * - * Note the difference between internally consistant and + * Note the difference between internally consistent and * correct. As we will have to try to access the target * object to determine whether the supplied data is correct, - * we will settle for internal consistancy at this point + * we will settle for internal consistency at this point * * Return: SUCCEED if instance of H5FD_ros3_fapl_t contains internally - * consistant data, FAIL otherwise. + * consistent data, FAIL otherwise. * * Programmer: Jacob Smith * 9/10/17 diff --git a/src/H5FDsec2.c b/src/H5FDsec2.c index 94610d3..16f4874 100644 --- a/src/H5FDsec2.c +++ b/src/H5FDsec2.c @@ -549,7 +549,7 @@ H5FD__sec2_query(const H5FD_t *_file, unsigned long *flags /* out */) /* Set the VFL feature flags that this driver supports */ /* Notice: the Mirror VFD Writer currently uses only the Sec2 driver as - * the underying driver -- as such, the Mirror VFD implementation copies + * the underlying driver -- as such, the Mirror VFD implementation copies * these feature flags as its own. Any modifications made here must be * reflected in H5FDmirror.c * -- JOS 2020-01-13 diff --git a/src/H5FSprivate.h b/src/H5FSprivate.h index b50056d..3d11aa9 100644 --- a/src/H5FSprivate.h +++ b/src/H5FSprivate.h @@ -116,7 +116,7 @@ typedef struct H5FS_section_class_t { const struct H5FS_section_class_t *, const uint8_t *, haddr_t, hsize_t, unsigned *); /* Routine to deserialize a buffer into a "live" section */ htri_t (*can_merge)(const H5FS_section_info_t *, const H5FS_section_info_t *, - void *); /* Routine to determine if two nodes are mergable */ + void *); /* Routine to determine if two nodes are mergeable */ herr_t (*merge)(H5FS_section_info_t **, H5FS_section_info_t *, void *); /* Routine to merge two nodes */ htri_t (*can_shrink)(const H5FS_section_info_t *, void *); /* Routine to determine if node can shrink container */ diff --git a/src/H5FSsection.c b/src/H5FSsection.c index 0efd750..9ac28e1 100644 --- a/src/H5FSsection.c +++ b/src/H5FSsection.c @@ -2050,11 +2050,11 @@ H5FS_sect_change_class(H5F_t *f, H5FS_t *fspace, H5FS_section_info_t *sect, uint } /* end else */ } /* end if */ - /* Check if the section's class change will affect the mergable list */ + /* Check if the section's class change will affect the mergeable list */ if ((old_cls->flags & H5FS_CLS_SEPAR_OBJ) != (new_cls->flags & H5FS_CLS_SEPAR_OBJ)) { - hbool_t to_mergable; /* Flag if the section is changing to a mergable section */ + hbool_t to_mergable; /* Flag if the section is changing to a mergeable section */ - /* Determine if this section is becoming mergable or is becoming separate */ + /* Determine if this section is becoming mergeable or is becoming separate */ if (old_cls->flags & H5FS_CLS_SEPAR_OBJ) to_mergable = TRUE; else @@ -2390,7 +2390,7 @@ done: * next file open, we will read the self referential free space * managers, float them and reduce the EOA to its value prior * to allocation of file space for the self referential free - * space managers on the preceeding file close. This EOA value + * space managers on the preceding file close. This EOA value * is stored in the free space manager superblock extension * message. * diff --git a/src/H5Fefc.c b/src/H5Fefc.c index 14dd69e..4c0bc79 100644 --- a/src/H5Fefc.c +++ b/src/H5Fefc.c @@ -422,7 +422,7 @@ H5F__efc_release_real(H5F_efc_t *efc) /* Sanity checks */ HDassert(efc); - /* Lock the EFC to prevent manipulation of the EFC wile we are releasing it. + /* Lock the EFC to prevent manipulation of the EFC while we are releasing it. * The EFC should never be locked when we enter this function because that * would require a cycle, a cycle would necessarily invoke * H5F__efc_try_close(), and that function checks the status of the lock diff --git a/src/H5Fint.c b/src/H5Fint.c index 24b306e..5bf2202 100644 --- a/src/H5Fint.c +++ b/src/H5Fint.c @@ -920,7 +920,7 @@ H5F__is_hdf5(const char *name) { H5FD_t *file = NULL; /* Low-level file struct */ H5F_shared_t *shared = NULL; /* Shared part of file */ - haddr_t sig_addr = HADDR_UNDEF; /* Addess of hdf5 file signature */ + haddr_t sig_addr = HADDR_UNDEF; /* Address of hdf5 file signature */ htri_t ret_value = FAIL; /* Return value */ FUNC_ENTER_PACKAGE @@ -1014,7 +1014,7 @@ H5F__new(H5F_shared_t *shared, unsigned flags, hid_t fcpl_id, hid_t fapl_id, H5F /* Initialization for handling file space (for paged aggregation) */ f->shared->pgend_meta_thres = H5F_FILE_SPACE_PGEND_META_THRES; - /* intialize point of no return */ + /* initialize point of no return */ f->shared->point_of_no_return = FALSE; /* Copy the file creation and file access property lists into the @@ -1151,7 +1151,7 @@ H5F__new(H5F_shared_t *shared, unsigned flags, hid_t fcpl_id, hid_t fapl_id, H5F f->shared->read_attempts = H5F_METADATA_READ_ATTEMPTS; } - /* Determine the # of bins for metdata read retries */ + /* Determine the # of bins for metadata read retries */ if (H5F_set_retries(f) < 0) HGOTO_ERROR(H5E_FILE, H5E_CANTINIT, NULL, "can't set retries and retries_nbins") @@ -1638,7 +1638,7 @@ H5F_open(const char *name, unsigned flags, hid_t fcpl_id, hid_t fapl_id) * or creating it) so we can compare it with files that are already * open. If that fails then we try again with the full set of flags * (only if they're different than the original failed attempt). - * However, if the file driver can't distinquish between files then + * However, if the file driver can't distinguish between files then * there's no reason to open the file tentatively because it's the * application's responsibility to prevent this situation (there's no * way for us to detect it here anyway). @@ -2939,7 +2939,7 @@ H5F__get_file_image(H5F_t *file, void *buf_ptr, size_t buf_len) * file driver. However, this test will not work if there is some * other file driver sitting on top of the multi file driver. * - * I'm not sure if this is possible at present, but in all likelyhood, + * I'm not sure if this is possible at present, but in all likelihood, * it will become possible in the future. On the other hand, we may * remove the split/multi file drivers before then. * @@ -2957,7 +2957,7 @@ H5F__get_file_image(H5F_t *file, void *buf_ptr, size_t buf_len) * driver other than the family file driver. Needless to say, this * rather defeats the purpose of the get file image operation. * - * While this problem is quire solvable, the required time and + * While this problem is quite solvable, the required time and * resources are lacking at present. Hence, for now, we don't * allow the get file image operation to be performed on files * opened with the family file driver. @@ -3418,7 +3418,7 @@ done: * 1) The file being opened has v3 superblock * 2) The file is opened with H5F_ACC_RDWR * 3) The file is not already marked for SWMR writing - * 4) Current implementaion for opened objects: + * 4) Current implementation for opened objects: * --only allow datasets and groups without attributes * --disallow named datatype with/without attributes * --disallow opened attributes attached to objects diff --git a/src/H5Fmount.c b/src/H5Fmount.c index e4ae406..3dba221 100644 --- a/src/H5Fmount.c +++ b/src/H5Fmount.c @@ -103,7 +103,7 @@ H5F__mount(H5G_loc_t *loc, const char *name, H5F_t *child, hid_t H5_ATTR_UNUSED H5F_t *parent = NULL; /*file containing mount point */ unsigned lt, rt, md; /*binary search indices */ int cmp; /*binary search comparison value*/ - H5G_loc_t mp_loc; /* entry of moint point to be opened */ + H5G_loc_t mp_loc; /* entry of mount point to be opened */ H5G_name_t mp_path; /* Mount point group hier. path */ H5O_loc_t mp_oloc; /* Mount point object location */ H5G_loc_t root_loc; /* Group location of root of file to mount */ @@ -483,7 +483,7 @@ done: /*------------------------------------------------------------------------- * Function: H5Funmount * - * Purpose: Given a mount point, dissassociate the mount point's file + * Purpose: Given a mount point, disassociate the mount point's file * from the file mounted there. Do not close either file. * * The mount point can either be the group in the parent or the diff --git a/src/H5Fsuper.c b/src/H5Fsuper.c index 4c4f087..8e5fe42 100644 --- a/src/H5Fsuper.c +++ b/src/H5Fsuper.c @@ -383,7 +383,7 @@ H5F__super_read(H5F_t *f, H5P_genplist_t *fa_plist, hbool_t initial_read) if (0 == mpi_rank) { herr_t status; - /* Try detecting file's siganture */ + /* Try detecting file's signature */ /* (Don't leave before Bcast, to avoid hang on error) */ H5E_BEGIN_TRY { @@ -475,7 +475,7 @@ H5F__super_read(H5F_t *f, H5P_genplist_t *fa_plist, hbool_t initial_read) * * After upgrading low_bound, the library will check to ensure that the * superblock version does not exceed the version allowed by high_bound. - * Otherise fail file open. + * Otherwise fail file open. * * For details, please see RFC:Setting Bounds for Object Creation in HDF5 1.10.0. */ diff --git a/src/H5Fsuper_cache.c b/src/H5Fsuper_cache.c index c2c3ada..cc8f97d 100644 --- a/src/H5Fsuper_cache.c +++ b/src/H5Fsuper_cache.c @@ -698,7 +698,7 @@ H5F__cache_superblock_serialize(const H5F_t *f, void *_image, size_t H5_ATTR_UNU HGOTO_ERROR(H5E_RESOURCE, H5E_CANTGET, FAIL, "driver get_eoa request failed") H5F_addr_encode(f, &image, (rel_eof + sblock->base_addr)); - /* Encode the driver informaton block address */ + /* Encode the driver information block address */ H5F_addr_encode(f, &image, sblock->driver_addr); /* Encode the root group object entry, including the cached stab info */ @@ -797,7 +797,7 @@ done: /*------------------------------------------------------------------------- * Function: H5F__cache_drvrinfo_get_initial_load_size * - * Purpose: Compute the intiial size of the data structure on disk. + * Purpose: Compute the initial size of the data structure on disk. * * Return: Non-negative on success/Negative on failure * diff --git a/src/H5Gcache.c b/src/H5Gcache.c index e9c60dd..340a1e4 100644 --- a/src/H5Gcache.c +++ b/src/H5Gcache.c @@ -134,7 +134,7 @@ H5G__cache_node_get_initial_load_size(void *_udata, size_t *image_len) * Function: H5G__cache_node_deserialize * * Purpose: Given a buffer containing the on disk image of a symbol table - * node, allocate an instance of H5G_node_t, load the contence of the + * node, allocate an instance of H5G_node_t, load the contents of the * image into it, and return a pointer to the instance. * * Note that deserializing the image requires access to the file @@ -240,7 +240,7 @@ H5G__cache_node_image_len(const void *_thing, size_t *image_len) /*------------------------------------------------------------------------- * Function: H5G__cache_node_serialize * - * Purpose: Given a correctly sized buffer and an instace of H5G_node_t, + * Purpose: Given a correctly sized buffer and an instance of H5G_node_t, * serialize the contents of the instance of H5G_node_t, and write * this data into the supplied buffer. This buffer will be written * to disk. diff --git a/src/H5Gcompact.c b/src/H5Gcompact.c index 350b4ae..6bc92ce 100644 --- a/src/H5Gcompact.c +++ b/src/H5Gcompact.c @@ -419,7 +419,7 @@ done: * Function: H5G__compact_lookup_cb * * Purpose: Callback routine for searching 'link' messages for a particular - * name & gettting object location for it + * name & getting object location for it * * Return: SUCCEED/FAIL * diff --git a/src/H5Gnode.c b/src/H5Gnode.c index a7b71c9..5964293 100644 --- a/src/H5Gnode.c +++ b/src/H5Gnode.c @@ -1207,7 +1207,7 @@ H5G__node_copy(H5F_t *f, const void H5_ATTR_UNUSED *_lt_key, haddr_t addr, const &(sn->entry[i]); /* Convenience variable to refer to current source group entry */ H5O_link_t lnk; /* Link to insert */ const char *name; /* Name of source object */ - H5G_entry_t tmp_src_ent; /* Temperary copy. Change will not affect the cache */ + H5G_entry_t tmp_src_ent; /* Temporary copy. Change will not affect the cache */ H5O_type_t obj_type = H5O_TYPE_UNKNOWN; /* Target object type */ H5G_copy_file_ud_t *cpy_udata; /* Copy file udata */ H5G_obj_create_t gcrt_info; /* Group creation info */ diff --git a/src/H5Gtraverse.c b/src/H5Gtraverse.c index 7ea6454..552e0b0 100644 --- a/src/H5Gtraverse.c +++ b/src/H5Gtraverse.c @@ -190,7 +190,7 @@ H5G__traverse_ud(const H5G_loc_t *grp_loc /*in,out*/, const H5O_link_t *lnk, H5G HGOTO_ERROR(H5E_SYM, H5E_CANTREGISTER, FAIL, "unable to register group") /* User-defined callback function */ - /* (Backwardly compatible with v0 H5L_class_t traverssal callback) */ + /* (Backwardly compatible with v0 H5L_class_t traversal callback) */ if (link_class->version == H5L_LINK_CLASS_T_VERS_0) cb_return = (((const H5L_class_0_t *)link_class)->trav_func)(lnk->name, cur_grp, lnk->u.ud.udata, lnk->u.ud.size, H5CX_get_lapl()); diff --git a/src/H5HFcache.c b/src/H5HFcache.c index 3053abc..bfb72d4 100644 --- a/src/H5HFcache.c +++ b/src/H5HFcache.c @@ -796,7 +796,7 @@ H5HF__cache_hdr_serialize(const H5F_t *f, void *_image, size_t H5_ATTR_NDEBUG_UN /* Encode I/O filter information */ if (H5O_msg_encode(hdr->f, H5O_PLINE_ID, FALSE, image, &(hdr->pline)) < 0) - HGOTO_ERROR(H5E_HEAP, H5E_CANTENCODE, FAIL, "can't encode I/O pipeline fiters") + HGOTO_ERROR(H5E_HEAP, H5E_CANTENCODE, FAIL, "can't encode I/O pipeline filters") image += hdr->filter_len; } /* end if */ @@ -895,7 +895,7 @@ H5HF__cache_iblock_get_initial_load_size(void *_udata, size_t *image_len) * Function: H5HF__cache_iblock_verify_chksum * * Purpose: Verify the computed checksum of the data structure is the - * same as the stored chksum. + * same as the stored checksum. * * Return: Success: TRUE/FALSE * Failure: Negative @@ -2627,7 +2627,7 @@ H5HF__cache_dblock_fsf_size(const void *_thing, hsize_t *fsf_size) * * For example, it is now possible for a dirty fractal heap * header to be flushed before a dirty dblock, as long as the - * there in an interviening iblock, and the header has no + * there in an intervening iblock, and the header has no * dirty immediate flush dependency children. * * Also, I gather that under some circumstances, a dblock @@ -2992,7 +2992,7 @@ done: * * For example, it is now possible for a dirty fractal heap * header to be flushed before a dirty dblock, as long as the - * there in an interviening iblock, and the header has no + * there in an intervening iblock, and the header has no * dirty immediate flush dependency children. * * Also, I gather that under some circumstances, a dblock @@ -3109,7 +3109,7 @@ done: * * For example, it is now possible for a dirty fractal heap * header to be flushed before a dirty dblock, as long as the - * there in an interviening iblock, and the header has no + * there in an intervening iblock, and the header has no * dirty immediate flush dependency children. * * Also, I gather that under some circumstances, a dblock @@ -3272,7 +3272,7 @@ done: * * For example, it is now possible for a dirty fractal heap * header to be flushed before a dirty dblock, as long as the - * there in an interviening iblock, and the header has no + * there in an intervening iblock, and the header has no * dirty immediate flush dependency children. * * Also, I gather that under some circumstances, a dblock diff --git a/src/H5HFdbg.c b/src/H5HFdbg.c index edc9612..a3cd84f 100644 --- a/src/H5HFdbg.c +++ b/src/H5HFdbg.c @@ -50,7 +50,7 @@ /* User data for direct block debugging iterator callback */ typedef struct { FILE *stream; /* Stream for output */ - int indent; /* Indention amount */ + int indent; /* Indentation amount */ int fwidth; /* Field width mount */ haddr_t dblock_addr; /* Direct block's address */ hsize_t dblock_size; /* Direct block's size */ @@ -63,7 +63,7 @@ typedef struct { typedef struct { H5FS_t *fspace; /* Free space manager */ FILE *stream; /* Stream for output */ - int indent; /* Indention amount */ + int indent; /* Indentation amount */ int fwidth; /* Field width mount */ } H5HF_debug_iter_ud2_t; diff --git a/src/H5HFdblock.c b/src/H5HFdblock.c index a20e2b8..3211501 100644 --- a/src/H5HFdblock.c +++ b/src/H5HFdblock.c @@ -432,7 +432,7 @@ H5HF__man_dblock_protect(H5HF_hdr_t *hdr, haddr_t dblock_addr, size_t dblock_siz H5HF_indirect_t *par_iblock, unsigned par_entry, unsigned flags) { H5HF_direct_t *dblock; /* Direct block from cache */ - H5HF_dblock_cache_ud_t udata; /* parent and other infor for deserializing direct block */ + H5HF_dblock_cache_ud_t udata; /* parent and other info for deserializing direct block */ H5HF_direct_t *ret_value = NULL; /* Return value */ FUNC_ENTER_PACKAGE diff --git a/src/H5HFsection.c b/src/H5HFsection.c index aa189a9..df41cf7 100644 --- a/src/H5HFsection.c +++ b/src/H5HFsection.c @@ -2291,7 +2291,7 @@ H5HF_sect_indirect_for_row(H5HF_hdr_t *hdr, H5HF_indirect_t *iblock, H5HF_free_s (sect->u.indirect.dir_rows = (H5HF_free_section_t **)H5MM_malloc(sizeof(H5HF_free_section_t *)))) HGOTO_ERROR(H5E_HEAP, H5E_NOSPACE, NULL, "allocation failed for row section pointer array") - /* Atatch the new row section to indirect section */ + /* Attach the new row section to indirect section */ sect->u.indirect.dir_rows[0] = row_sect; sect->u.indirect.rc = 1; diff --git a/src/H5HFspace.c b/src/H5HFspace.c index b5622d6..530d49c 100644 --- a/src/H5HFspace.c +++ b/src/H5HFspace.c @@ -383,7 +383,7 @@ H5HF__space_create_root(const H5HF_hdr_t *hdr, H5HF_indirect_t *root_iblock) /* Only need to scan the sections if the free space has been initialized */ if (hdr->fspace) - /* Iterate over all sections, seting the parent pointers in 'single' sections to the new indirect + /* Iterate over all sections, setting the parent pointers in 'single' sections to the new indirect * block */ if (H5FS_sect_iterate(hdr->f, hdr->fspace, H5HF_space_create_root_cb, root_iblock) < 0) HGOTO_ERROR(H5E_FSPACE, H5E_BADITER, FAIL, "can't iterate over sections to set parent pointers") @@ -2727,7 +2727,7 @@ H5L__exists_inter_cb(H5G_loc_t H5_ATTR_UNUSED *grp_loc /*in*/, const char H5_ATT if (lnk != NULL) { /* Check for more components to the path */ if (udata->sep) { - H5G_traverse_t cb_func; /* Callback function for tranversal */ + H5G_traverse_t cb_func; /* Callback function for traversal */ char *next; /* Pointer to next component name */ /* Look for another separator */ @@ -2778,7 +2778,7 @@ htri_t H5L_exists_tolerant(const H5G_loc_t *loc, const char *name) { H5L_trav_le_t udata; /* User data for traversal */ - H5G_traverse_t cb_func; /* Callback function for tranversal */ + H5G_traverse_t cb_func; /* Callback function for traversal */ char *name_copy = NULL; /* Duplicate of name */ char *name_trav; /* Name to traverse */ htri_t ret_value = FAIL; /* Return value */ diff --git a/src/H5MFaggr.c b/src/H5MFaggr.c index ba1f2c6..8df9f09 100644 --- a/src/H5MFaggr.c +++ b/src/H5MFaggr.c @@ -180,7 +180,7 @@ H5MF__aggr_alloc(H5F_t *f, H5F_blk_aggr_t *aggr, H5F_blk_aggr_t *other_aggr, H5F haddr_t aggr_frag_addr = HADDR_UNDEF; /* Address of aggregrator fragment */ hsize_t aggr_frag_size = 0; /* Size of aggregator fragment */ hsize_t alignment; /* Alignment of this section */ - hsize_t aggr_mis_align = 0; /* Mis-alignment of aggregator */ + hsize_t aggr_mis_align = 0; /* Misalignment of aggregator */ H5FD_mem_t alloc_type, other_alloc_type; /* Current aggregator & 'other' aggregator types */ #ifdef H5MF_AGGR_DEBUG diff --git a/src/H5MFdbg.c b/src/H5MFdbg.c index 1afbdc8..c84f173 100644 --- a/src/H5MFdbg.c +++ b/src/H5MFdbg.c @@ -50,7 +50,7 @@ typedef struct { H5FS_t *fspace; /* Free space manager */ FILE *stream; /* Stream for output */ - int indent; /* Indention amount */ + int indent; /* Indentation amount */ int fwidth; /* Field width amount */ } H5MF_debug_iter_ud_t; diff --git a/src/H5Oainfo.c b/src/H5Oainfo.c index c3eb1c9..39ece0b 100644 --- a/src/H5Oainfo.c +++ b/src/H5Oainfo.c @@ -52,7 +52,7 @@ const H5O_msg_class_t H5O_MSG_AINFO[1] = {{ H5O_AINFO_ID, /*message id number */ "ainfo", /*message name for debugging */ sizeof(H5O_ainfo_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__ainfo_decode, /*decode message */ H5O_ainfo_encode, /*encode message */ H5O_ainfo_copy, /*copy the native value */ diff --git a/src/H5Oattr.c b/src/H5Oattr.c index a41571d..c75f51c 100644 --- a/src/H5Oattr.c +++ b/src/H5Oattr.c @@ -65,7 +65,7 @@ const H5O_msg_class_t H5O_MSG_ATTR[1] = {{ H5O_ATTR_ID, /* message id number */ "attribute", /* message name for debugging */ sizeof(H5A_t), /* native message size */ - H5O_SHARE_IS_SHARABLE, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE, /* messages are shareable? */ H5O_attr_shared_decode, /* decode message */ H5O_attr_shared_encode, /* encode message */ H5O_attr_copy, /* copy the native value */ diff --git a/src/H5Oattribute.c b/src/H5Oattribute.c index 0b1d1e8..1b75cc3 100644 --- a/src/H5Oattribute.c +++ b/src/H5Oattribute.c @@ -24,7 +24,7 @@ /* Module Setup */ /****************/ -#define H5A_FRIEND /*suppress error about including H5Apkg */ +#define H5A_FRIEND /* Suppress error about including H5Apkg.h */ #include "H5Omodule.h" /* This source code file is part of the H5O module */ /***********/ @@ -243,19 +243,19 @@ H5O__attr_create(const H5O_loc_t *loc, H5A_t *attr) /* Check if switching to "dense" attribute storage is possible */ if (!H5F_addr_defined(ainfo.fheap_addr)) { - htri_t sharable; /* Whether the attribute will be shared */ + htri_t shareable; /* Whether the attribute will be shared */ size_t raw_size = 0; /* Raw size of message */ - /* Check for attribute being sharable */ - if ((sharable = H5SM_can_share(loc->file, NULL, NULL, H5O_ATTR_ID, attr)) < 0) + /* Check for attribute being shareable */ + if ((shareable = H5SM_can_share(loc->file, NULL, NULL, H5O_ATTR_ID, attr)) < 0) HGOTO_ERROR(H5E_ATTR, H5E_BADMESG, FAIL, "can't determine attribute sharing status") - else if (sharable == FALSE) { + else if (shareable == FALSE) { /* Compute the size needed to encode the attribute */ raw_size = (H5O_MSG_ATTR->raw_size)(loc->file, FALSE, attr); } /* end if */ /* Check for condititions for switching to "dense" attribute storage are met */ - if (ainfo.nattrs == oh->max_compact || (!sharable && raw_size >= H5O_MESG_MAX_SIZE)) { + if (ainfo.nattrs == oh->max_compact || (!shareable && raw_size >= H5O_MESG_MAX_SIZE)) { H5O_iter_cvt_t udata; /* User data for callback */ H5O_mesg_operator_t op; /* Wrapper for operator */ diff --git a/src/H5Obogus.c b/src/H5Obogus.c index b98f058..2c01f30 100644 --- a/src/H5Obogus.c +++ b/src/H5Obogus.c @@ -47,7 +47,7 @@ const H5O_msg_class_t H5O_MSG_BOGUS_VALID[1] = {{ H5O_BOGUS_VALID_ID, /*message id number */ "bogus valid", /*message name for debugging */ 0, /*native message size */ - H5O_SHARE_IS_SHARABLE, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE, /* messages are shareable? */ H5O__bogus_decode, /*decode message */ H5O__bogus_encode, /*encode message */ NULL, /*copy the native value */ @@ -71,7 +71,7 @@ const H5O_msg_class_t H5O_MSG_BOGUS_INVALID[1] = {{ H5O_BOGUS_INVALID_ID, /*message id number */ "bogus invalid", /*message name for debugging */ 0, /*native message size */ - H5O_SHARE_IS_SHARABLE, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE, /* messages are shareable? */ H5O__bogus_decode, /*decode message */ H5O__bogus_encode, /*encode message */ NULL, /*copy the native value */ diff --git a/src/H5Obtreek.c b/src/H5Obtreek.c index 5bb7be8..c635a21 100644 --- a/src/H5Obtreek.c +++ b/src/H5Obtreek.c @@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_BTREEK[1] = {{ H5O_BTREEK_ID, /*message id number */ "v1 B-tree 'K' values", /*message name for debugging */ sizeof(H5O_btreek_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O_btreek_decode, /*decode message */ H5O_btreek_encode, /*encode message */ H5O_btreek_copy, /*copy the native value */ diff --git a/src/H5Ocache.c b/src/H5Ocache.c index 1487e19..09b3cc5 100644 --- a/src/H5Ocache.c +++ b/src/H5Ocache.c @@ -275,7 +275,7 @@ H5O__cache_verify_chksum(const void *_image, size_t len, void *_udata) * * Note that the object header is read with with a speculative read. * If the initial read is too small, make note of this fact and return - * without error. H5C_load_entry() will note the size discrepency + * without error. H5C_load_entry() will note the size discrepancy * and retry the deserialize operation with the correct size read. * * Return: Success: Pointer to in core representation diff --git a/src/H5Ocache_image.c b/src/H5Ocache_image.c index bb603d5..bda94dc 100644 --- a/src/H5Ocache_image.c +++ b/src/H5Ocache_image.c @@ -50,7 +50,7 @@ const H5O_msg_class_t H5O_MSG_MDCI[1] = {{ H5O_MDCI_MSG_ID, /* message id number */ "mdci", /* message name for debugging */ sizeof(H5O_mdci_t), /* native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__mdci_decode, /* decode message */ H5O__mdci_encode, /* encode message */ H5O__mdci_copy, /* copy method */ @@ -285,7 +285,7 @@ H5O__mdci_delete(H5F_t *f, H5O_t *open_oh, void *_mesg) * manager issue is in use, file space for the non-empty self * referential free space managers was also allocated from VFD * layer at the end of file. Since these allocations directly - * preceeded the cache image allocation they should be directly + * preceded the cache image allocation they should be directly * adjacent to the cache image block at the end of file. * * In this case, just call H5MF_tidy_self_referential_fsm_hack(). @@ -302,7 +302,7 @@ H5O__mdci_delete(H5F_t *f, H5O_t *open_oh, void *_mesg) * * If the hack to work around the persistent self referential * free space manager issue is NOT in use, just call H5MF_xfree() - * to release the cache iamge. In principle, we should be able + * to release the cache image. In principle, we should be able * to just reduce the EOA to the base address of the cache * image block, as there shouldn't be any file space allocation * before the first metadata cache access. However, given diff --git a/src/H5Ocont.c b/src/H5Ocont.c index 19e5a0b..3805368 100644 --- a/src/H5Ocont.c +++ b/src/H5Ocont.c @@ -47,7 +47,7 @@ const H5O_msg_class_t H5O_MSG_CONT[1] = {{ H5O_CONT_ID, /*message id number */ "hdr continuation", /*message name for debugging */ sizeof(H5O_cont_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__cont_decode, /*decode message */ H5O__cont_encode, /*encode message */ NULL, /*no copy method */ diff --git a/src/H5Ocopy.c b/src/H5Ocopy.c index 8e576d5..8a11a75 100644 --- a/src/H5Ocopy.c +++ b/src/H5Ocopy.c @@ -883,7 +883,7 @@ H5O__copy_header_real(const H5O_loc_t *oloc_src, H5O_loc_t *oloc_dst /*out*/, H5 oh_dst = NULL; inserted = TRUE; - /* Reset metadat tag */ + /* Reset metadata tag */ H5_END_TAG /* Set obj_type and udata, if requested */ @@ -1096,7 +1096,7 @@ H5O__copy_header(const H5O_loc_t *oloc_src, H5O_loc_t *oloc_dst /*out */, hid_t if (H5P_get(ocpy_plist, H5O_CPY_OPTION_NAME, &cpy_option) < 0) HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get object copy flag") - /* Retrieve the marge committed datatype list */ + /* Retrieve the merge committed datatype list */ if (H5P_peek(ocpy_plist, H5O_CPY_MERGE_COMM_DT_LIST_NAME, &dt_list) < 0) HGOTO_ERROR(H5E_PLIST, H5E_CANTGET, FAIL, "can't get merge committed datatype list") diff --git a/src/H5Odrvinfo.c b/src/H5Odrvinfo.c index 6f8a4d0..2846c42 100644 --- a/src/H5Odrvinfo.c +++ b/src/H5Odrvinfo.c @@ -38,7 +38,7 @@ const H5O_msg_class_t H5O_MSG_DRVINFO[1] = {{ H5O_DRVINFO_ID, /*message id number */ "driver info", /*message name for debugging */ sizeof(H5O_drvinfo_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O_drvinfo_decode, /*decode message */ H5O_drvinfo_encode, /*encode message */ H5O_drvinfo_copy, /*copy the native value */ diff --git a/src/H5Odtype.c b/src/H5Odtype.c index 947d80b..34c14c7 100644 --- a/src/H5Odtype.c +++ b/src/H5Odtype.c @@ -89,7 +89,7 @@ const H5O_msg_class_t H5O_MSG_DTYPE[1] = {{ H5O_DTYPE_ID, /* message id number */ "datatype", /* message name for debugging */ sizeof(H5T_t), /* native message size */ - H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */ H5O_dtype_shared_decode, /* decode message */ H5O_dtype_shared_encode, /* encode message */ H5O_dtype_copy, /* copy the native value */ diff --git a/src/H5Oefl.c b/src/H5Oefl.c index 7f5f524..7cedea4 100644 --- a/src/H5Oefl.c +++ b/src/H5Oefl.c @@ -41,7 +41,7 @@ const H5O_msg_class_t H5O_MSG_EFL[1] = {{ H5O_EFL_ID, /*message id number */ "external file list", /*message name for debugging */ sizeof(H5O_efl_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O_efl_decode, /*decode message */ H5O_efl_encode, /*encode message */ H5O_efl_copy, /*copy native value */ diff --git a/src/H5Ofill.c b/src/H5Ofill.c index 6ba7a11..8313955 100644 --- a/src/H5Ofill.c +++ b/src/H5Ofill.c @@ -107,7 +107,7 @@ const H5O_msg_class_t H5O_MSG_FILL[1] = {{ H5O_FILL_ID, /*message id number */ "fill", /*message name for debugging */ sizeof(H5O_fill_t), /*native message size */ - H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */ H5O_fill_shared_decode, /*decode message */ H5O_fill_shared_encode, /*encode message */ H5O_fill_copy, /*copy the native value */ @@ -131,7 +131,7 @@ const H5O_msg_class_t H5O_MSG_FILL_NEW[1] = {{ H5O_FILL_NEW_ID, /*message id number */ "fill_new", /*message name for debugging */ sizeof(H5O_fill_t), /*native message size */ - H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */ H5O_fill_new_shared_decode, /*decode message */ H5O_fill_new_shared_encode, /*encode message */ H5O_fill_copy, /*copy the native value */ diff --git a/src/H5Ofsinfo.c b/src/H5Ofsinfo.c index 541f0f5..dab34d0 100644 --- a/src/H5Ofsinfo.c +++ b/src/H5Ofsinfo.c @@ -44,7 +44,7 @@ const H5O_msg_class_t H5O_MSG_FSINFO[1] = {{ H5O_FSINFO_ID, /* message id number */ "fsinfo", /* message name for debugging */ sizeof(H5O_fsinfo_t), /* native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O_fsinfo_decode, /* decode message */ H5O_fsinfo_encode, /* encode message */ H5O_fsinfo_copy, /* copy the native value */ @@ -165,7 +165,7 @@ H5O_fsinfo_decode(H5F_t *f, H5O_t H5_ATTR_UNUSED *open_oh, unsigned H5_ATTR_UNUS H5F_DECODE_LENGTH(f, p, fsinfo->threshold); /* Free-space section threshold */ H5F_DECODE_LENGTH(f, p, fsinfo->page_size); /* File space page size */ - UINT16DECODE(p, fsinfo->pgend_meta_thres); /* Page end metdata threshold */ + UINT16DECODE(p, fsinfo->pgend_meta_thres); /* Page end metadata threshold */ H5F_addr_decode(f, &p, &(fsinfo->eoa_pre_fsm_fsalloc)); /* EOA before free-space header and section info */ diff --git a/src/H5Oginfo.c b/src/H5Oginfo.c index 7879e25..d116a9f 100644 --- a/src/H5Oginfo.c +++ b/src/H5Oginfo.c @@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_GINFO[1] = {{ H5O_GINFO_ID, /*message id number */ "ginfo", /*message name for debugging */ sizeof(H5O_ginfo_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O_ginfo_decode, /*decode message */ H5O_ginfo_encode, /*encode message */ H5O_ginfo_copy, /*copy the native value */ diff --git a/src/H5Oint.c b/src/H5Oint.c index 4448277..93a0f95 100644 --- a/src/H5Oint.c +++ b/src/H5Oint.c @@ -1425,7 +1425,7 @@ H5O_touch(const H5O_loc_t *loc, hbool_t force) /* Create/Update the modification time message */ if (H5O_touch_oh(loc->file, oh, force) < 0) - HGOTO_ERROR(H5E_OHDR, H5E_CANTSET, FAIL, "unable to update object modificaton time") + HGOTO_ERROR(H5E_OHDR, H5E_CANTSET, FAIL, "unable to update object modification time") /* Mark object header as changed */ oh_flags |= H5AC__DIRTIED_FLAG; diff --git a/src/H5Olayout.c b/src/H5Olayout.c index dec7444..bc75f88 100644 --- a/src/H5Olayout.c +++ b/src/H5Olayout.c @@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_LAYOUT[1] = {{ H5O_LAYOUT_ID, /* message id number */ "layout", /* message name for debugging */ sizeof(H5O_layout_t), /* native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__layout_decode, /* decode message */ H5O__layout_encode, /* encode message */ H5O__layout_copy, /* copy the native value */ diff --git a/src/H5Olinfo.c b/src/H5Olinfo.c index 0978d25..3c87500 100644 --- a/src/H5Olinfo.c +++ b/src/H5Olinfo.c @@ -53,7 +53,7 @@ const H5O_msg_class_t H5O_MSG_LINFO[1] = {{ H5O_LINFO_ID, /*message id number */ "linfo", /*message name for debugging */ sizeof(H5O_linfo_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__linfo_decode, /*decode message */ H5O__linfo_encode, /*encode message */ H5O__linfo_copy, /*copy the native value */ diff --git a/src/H5Olink.c b/src/H5Olink.c index 6b87fce..767e97f 100644 --- a/src/H5Olink.c +++ b/src/H5Olink.c @@ -56,7 +56,7 @@ const H5O_msg_class_t H5O_MSG_LINK[1] = {{ H5O_LINK_ID, /*message id number */ "link", /*message name for debugging */ sizeof(H5O_link_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__link_decode, /*decode message */ H5O_link_encode, /*encode message */ H5O_link_copy, /*copy the native value */ diff --git a/src/H5Omessage.c b/src/H5Omessage.c index 76a3d06..6a39ef6 100644 --- a/src/H5Omessage.c +++ b/src/H5Omessage.c @@ -371,7 +371,7 @@ H5O__msg_write_real(H5F_t *f, H5O_t *oh, const H5O_msg_class_t *type, unsigned m HDassert(((H5O_shared_t *)idx_msg->native)->type != H5O_SHARE_TYPE_COMMITTED); /* Also, sanity check that a message doesn't switch status from being - * shared (or sharable) to being unsharable. (Which could cause + * shared (or shareable) to being unshareable. (Which could cause * a message to increase in size in the object header) */ HDassert(!(mesg_flags & H5O_MSG_FLAG_DONTSHARE)); @@ -693,7 +693,7 @@ H5O__msg_free_real(const H5O_msg_class_t *type, void *msg_native) /*------------------------------------------------------------------------- * Function: H5O_msg_copy * - * Purpose: Copies a message. If MESG is is the null pointer then a null + * Purpose: Copies a message. If MESG is the null pointer then a null * pointer is returned with no error. * * Return: Success: Ptr to the new message @@ -1131,7 +1131,7 @@ done: * Nov 19 2004 * * Description: - * This function interates over the object headers of an object + * This function iterates over the object headers of an object * specified with 'loc' of type 'type_id'. For each object header of the * object, the 'op_data' and some additional information (specified below) are * passed to the 'op' function. @@ -1194,7 +1194,7 @@ done: * Sep 6 2005 * * Description: - * This function interates over the object headers of an object + * This function iterates over the object headers of an object * specified with 'ent' of type 'type_id'. For each object header of the * object, the 'op_data' and some additional information (specified below) are * passed to the 'op' function. @@ -1544,7 +1544,7 @@ H5O_msg_is_shared(unsigned type_id, const void *mesg) HDassert(type); HDassert(mesg); - /* If messages in a class aren't sharable, then obviously this message isn't shared! :-) */ + /* If messages in a class aren't shareable, then obviously this message isn't shared! :-) */ if (type->share_flags & H5O_SHARE_IS_SHARABLE) ret_value = H5O_IS_STORED_SHARED(((const H5O_shared_t *)mesg)->type); else @@ -1755,7 +1755,7 @@ done: /*------------------------------------------------------------------------- * Function: H5O__msg_copy_file * - * Purpose: Copies a message to file. If MESG is is the null pointer then a null + * Purpose: Copies a message to file. If MESG is the null pointer then a null * pointer is returned with no error. * * Attempts to share the message in the destination and sets diff --git a/src/H5Omtime.c b/src/H5Omtime.c index b52e65f..36995b0 100644 --- a/src/H5Omtime.c +++ b/src/H5Omtime.c @@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_MTIME[1] = {{ H5O_MTIME_ID, /*message id number */ "mtime", /*message name for debugging */ sizeof(time_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__mtime_decode, /*decode message */ H5O_mtime_encode, /*encode message */ H5O_mtime_copy, /*copy the native value */ @@ -68,7 +68,7 @@ const H5O_msg_class_t H5O_MSG_MTIME_NEW[1] = {{ H5O_MTIME_NEW_ID, /*message id number */ "mtime_new", /*message name for debugging */ sizeof(time_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__mtime_new_decode, /*decode message */ H5O_mtime_new_encode, /*encode message */ H5O_mtime_copy, /*copy the native value */ diff --git a/src/H5Oname.c b/src/H5Oname.c index e9b5c5c..0c1923a 100644 --- a/src/H5Oname.c +++ b/src/H5Oname.c @@ -43,7 +43,7 @@ const H5O_msg_class_t H5O_MSG_NAME[1] = {{ H5O_NAME_ID, /*message id number */ "name", /*message name for debugging */ sizeof(H5O_name_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__name_decode, /*decode message */ H5O_name_encode, /*encode message */ H5O_name_copy, /*copy the native value */ diff --git a/src/H5Onull.c b/src/H5Onull.c index 9b377fb..0f3143c 100644 --- a/src/H5Onull.c +++ b/src/H5Onull.c @@ -31,7 +31,7 @@ const H5O_msg_class_t H5O_MSG_NULL[1] = {{ H5O_NULL_ID, /*message id number */ "null", /*message name for debugging */ 0, /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ NULL, /*no decode method */ NULL, /*no encode method */ NULL, /*no copy method */ diff --git a/src/H5Opkg.h b/src/H5Opkg.h index 0b1f410..52b7ee8 100644 --- a/src/H5Opkg.h +++ b/src/H5Opkg.h @@ -378,7 +378,7 @@ typedef struct H5O_chunk_proxy_t { H5O_t *oh; /* Object header for this chunk */ unsigned chunkno; /* Chunk number for this chunk */ - /* Flush depencency parent information (not stored) + /* Flush dependency parent information (not stored) * * The following field is used to store a pointer * to the in-core representation of a new chunk proxy's flush dependency diff --git a/src/H5Opline.c b/src/H5Opline.c index de9c99d..b278509 100644 --- a/src/H5Opline.c +++ b/src/H5Opline.c @@ -67,7 +67,7 @@ const H5O_msg_class_t H5O_MSG_PLINE[1] = {{ H5O_PLINE_ID, /* message id number */ "filter pipeline", /* message name for debugging */ sizeof(H5O_pline_t), /* native message size */ - H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */ H5O_pline_shared_decode, /* decode message */ H5O_pline_shared_encode, /* encode message */ H5O_pline_copy, /* copy the native value */ @@ -86,7 +86,7 @@ const H5O_msg_class_t H5O_MSG_PLINE[1] = {{ H5O_pline_shared_debug /* debug the message */ }}; -/* Format version bounds for filter pipleline */ +/* Format version bounds for filter pipeline */ const unsigned H5O_pline_ver_bounds[] = { H5O_PLINE_VERSION_1, /* H5F_LIBVER_EARLIEST */ H5O_PLINE_VERSION_2, /* H5F_LIBVER_V18 */ diff --git a/src/H5Orefcount.c b/src/H5Orefcount.c index 4bf54c1..f3e3bc0 100644 --- a/src/H5Orefcount.c +++ b/src/H5Orefcount.c @@ -45,7 +45,7 @@ const H5O_msg_class_t H5O_MSG_REFCOUNT[1] = {{ H5O_REFCOUNT_ID, /*message id number */ "refcount", /*message name for debugging */ sizeof(H5O_refcount_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__refcount_decode, /*decode message */ H5O_refcount_encode, /*encode message */ H5O_refcount_copy, /*copy the native value */ diff --git a/src/H5Osdspace.c b/src/H5Osdspace.c index 6c2b856..99689cc 100644 --- a/src/H5Osdspace.c +++ b/src/H5Osdspace.c @@ -61,7 +61,7 @@ const H5O_msg_class_t H5O_MSG_SDSPACE[1] = {{ H5O_SDSPACE_ID, /* message id number */ "dataspace", /* message name for debugging */ sizeof(H5S_extent_t), /* native message size */ - H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are sharable? */ + H5O_SHARE_IS_SHARABLE | H5O_SHARE_IN_OHDR, /* messages are shareable? */ H5O_sdspace_shared_decode, /* decode message */ H5O_sdspace_shared_encode, /* encode message */ H5O_sdspace_copy, /* copy the native value */ @@ -225,19 +225,6 @@ done: This function encodes the native memory form of the simple dimensionality message in the "raw" disk form. - MODIFICATIONS - Robb Matzke, 1998-04-09 - The current and maximum dimensions are now H5F_SIZEOF_SIZET bytes - instead of just four bytes. - - Robb Matzke, 1998-07-20 - Added a version number and reformatted the message for aligment. - - Raymond Lu - April 8, 2004 - Added the type of dataspace into this header message using a reserved - byte. - --------------------------------------------------------------------------*/ static herr_t H5O_sdspace_encode(H5F_t *f, uint8_t *p, const void *_mesg) diff --git a/src/H5Oshared.c b/src/H5Oshared.c index a222bde..37250b1 100644 --- a/src/H5Oshared.c +++ b/src/H5Oshared.c @@ -621,7 +621,7 @@ done: * Function: H5O__shared_post_copy_file * * Purpose: Delete a shared message and replace with a new one. - * The function is needed at cases such as coping a shared reg_ref attribute. + * The function is needed at cases such as copying a shared reg_ref attribute. * When a shared reg_ref attribute is copied from one file to * another, the values in file need to be replaced. The only way * to complish that is to delete the old message and write the diff --git a/src/H5Oshmesg.c b/src/H5Oshmesg.c index c8fd06c..a0bad93 100644 --- a/src/H5Oshmesg.c +++ b/src/H5Oshmesg.c @@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_SHMESG[1] = {{ H5O_SHMESG_ID, /*message id number */ "shared message table", /*message name for debugging */ sizeof(H5O_shmesg_table_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__shmesg_decode, /*decode message */ H5O_shmesg_encode, /*encode message */ H5O_shmesg_copy, /*copy the native value */ diff --git a/src/H5Ostab.c b/src/H5Ostab.c index 9e5c87f..e95b89d 100644 --- a/src/H5Ostab.c +++ b/src/H5Ostab.c @@ -51,7 +51,7 @@ const H5O_msg_class_t H5O_MSG_STAB[1] = {{ H5O_STAB_ID, /*message id number */ "stab", /*message name for debugging */ sizeof(H5O_stab_t), /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ H5O__stab_decode, /*decode message */ H5O__stab_encode, /*encode message */ H5O__stab_copy, /*copy the native value */ diff --git a/src/H5Ounknown.c b/src/H5Ounknown.c index 2b8a7bb..dc43a6c 100644 --- a/src/H5Ounknown.c +++ b/src/H5Ounknown.c @@ -37,7 +37,7 @@ const H5O_msg_class_t H5O_MSG_UNKNOWN[1] = {{ H5O_UNKNOWN_ID, /*message id number */ "unknown", /*message name for debugging */ 0, /*native message size */ - 0, /* messages are sharable? */ + 0, /* messages are shareable? */ NULL, /*decode message */ NULL, /*encode message */ NULL, /*copy the native value */ @@ -664,7 +664,7 @@ H5Pset(hid_t plist_id, const char *name, const void *value) if (!name || !*name) HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property name"); if (value == NULL) - HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalied property value"); + HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property value"); /* Go set the value */ if (H5P_set(plist, name, value) < 0) @@ -678,7 +678,7 @@ done: NAME H5Pexist PURPOSE - Routine to query the existance of a property in a property object. + Routine to query the existence of a property in a property object. USAGE htri_t H5P_exist(id, name) hid_t id; IN: Property object ID to check @@ -712,7 +712,7 @@ H5Pexist(hid_t id, const char *name) if (!name || !*name) HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property name"); - /* Check for the existance of the property in the list or class */ + /* Check for the existence of the property in the list or class */ if (H5I_GENPROP_LST == H5I_get_type(id)) { if (NULL == (plist = (H5P_genplist_t *)H5I_object(id))) HGOTO_ERROR(H5E_ARGS, H5E_BADTYPE, FAIL, "not a property list"); @@ -856,7 +856,7 @@ done: Failure: H5I_INVALID_HID (negative) DESCRIPTION Decodes a property list from a binary buffer. The contents of the buffer - contain the values for the correponding properties of the plist. The decode + contain the values for the corresponding properties of the plist. The decode callback of a certain property decodes its value from the buffer and sets it in the property list. GLOBAL VARIABLES @@ -1274,7 +1274,7 @@ H5Pget(hid_t plist_id, const char *name, void *value) if (!name || !*name) HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property name"); if (value == NULL) - HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalied property value"); + HGOTO_ERROR(H5E_ARGS, H5E_BADVALUE, FAIL, "invalid property value"); /* Go get the value */ if (H5P_get(plist, name, value) < 0) @@ -1297,7 +1297,7 @@ done: Returns non-negative on success, negative on failure. DESCRIPTION Removes a property from a property list. Both properties which were - in existance when the property list was created (i.e. properties registered + in existence when the property list was created (i.e. properties registered with H5Pregister2) and properties added to the list after it was created (i.e. added with H5Pinsert2) may be removed from a property list. Properties do not need to be removed a property list before the list itself @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * @@ -67,7 +67,7 @@ * the special "::" string. * * Return: Success: Non-negative - * Failture: Negative + * Failure: Negative * *------------------------------------------------------------------------- */ @@ -98,7 +98,7 @@ done: * Positive if one or more of the plugin types are enabled * * Return: Success: Non-negative - * Failture: Negative + * Failure: Negative * *------------------------------------------------------------------------- */ @@ -127,7 +127,7 @@ done: * Purpose: Insert a plugin search path at the end of the list. * * Return: Success: Non-negative - * Failture: Negative + * Failure: Negative * *------------------------------------------------------------------------- */ @@ -159,7 +159,7 @@ done: * Purpose: Insert a plugin search path at the beginning of the list. * * Return: Success: Non-negative - * Failture: Negative + * Failure: Negative * *------------------------------------------------------------------------- */ @@ -233,7 +233,7 @@ done: * other paths after the index. * * Return: Success: Non-negative - * Failture: Negative + * Failure: Negative * *------------------------------------------------------------------------- */ @@ -269,11 +269,13 @@ done: /*------------------------------------------------------------------------- * Function: H5PLremove * - * Purpose: Remove the plugin path at the specifed index and compact + * Purpose: Remove the plugin path at the specified index and compact * the list. * * Return: Success: Non-negative - * Failture: Negative + * Failure: Negative + * + * Return: Non-negative or success. * *------------------------------------------------------------------------- */ diff --git a/src/H5PLextern.h b/src/H5PLextern.h index 7f3df5e..d136051 100644 --- a/src/H5PLextern.h +++ b/src/H5PLextern.h @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/src/H5PLint.c b/src/H5PLint.c index 3c60f68..ff0be4e 100644 --- a/src/H5PLint.c +++ b/src/H5PLint.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/src/H5PLmodule.h b/src/H5PLmodule.h index 66a24fd..fff1e95 100644 --- a/src/H5PLmodule.h +++ b/src/H5PLmodule.h @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/src/H5PLpath.c b/src/H5PLpath.c index 16b4566..96bbfa2 100644 --- a/src/H5PLpath.c +++ b/src/H5PLpath.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * @@ -320,7 +320,7 @@ H5PL__close_path_table(void) * Purpose: Gets the number of plugin paths that have been stored. * * Return: Success: The number of paths - * Failture: Can't fail + * Failure: Can't fail *------------------------------------------------------------------------- */ unsigned diff --git a/src/H5PLplugin_cache.c b/src/H5PLplugin_cache.c index b348738..1d267a2 100644 --- a/src/H5PLplugin_cache.c +++ b/src/H5PLplugin_cache.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/src/H5PLprivate.h b/src/H5PLprivate.h index 5406ad7..8f871c9 100644 --- a/src/H5PLprivate.h +++ b/src/H5PLprivate.h @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/src/H5PLpublic.h b/src/H5PLpublic.h index 9014131..f5332e6 100644 --- a/src/H5PLpublic.h +++ b/src/H5PLpublic.h @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/src/H5Pdapl.c b/src/H5Pdapl.c index 37a3be8..0ce0efd 100644 --- a/src/H5Pdapl.c +++ b/src/H5Pdapl.c @@ -739,7 +739,7 @@ H5P__dapl_efile_pref_close(const char H5_ATTR_UNUSED *name, size_t H5_ATTR_UNUSE * Purpose: Set the number of objects in the meta data cache and the * maximum number of chunks and bytes in the raw data chunk cache. * Once set, these values will override the values in the file access - * property list. Each of thhese values can be individually unset + * property list. Each of these values can be individually unset * (or not set at all) by passing the macros: * H5D_CHUNK_CACHE_NCHUNKS_DEFAULT, * H5D_CHUNK_CACHE_NSLOTS_DEFAULT, and/or diff --git a/src/H5Pdcpl.c b/src/H5Pdcpl.c index f6b2a7a..b3d7b44 100644 --- a/src/H5Pdcpl.c +++ b/src/H5Pdcpl.c @@ -1717,7 +1717,7 @@ done: * Failure: Negative * * Programmer: Neil Fortner - * Thurday, Feb 26, 2015 + * Thursday, Feb 26, 2015 * *-------------------------------------------------------------------------- */ @@ -1951,7 +1951,7 @@ H5P__init_def_layout(void) FUNC_ENTER_STATIC_NOERR - /* Initialize the default layout info for non-contigous layouts */ + /* Initialize the default layout info for non-contiguous layouts */ H5D_def_layout_compact_g.storage.type = H5D_COMPACT; H5D_def_layout_compact_g.storage.u.compact = def_store_compact; H5D_def_layout_chunk_g.u.chunk = def_layout_chunk; diff --git a/src/H5Pdxpl.c b/src/H5Pdxpl.c index de16c3e..77e8aa7 100644 --- a/src/H5Pdxpl.c +++ b/src/H5Pdxpl.c @@ -1968,7 +1968,7 @@ done: /*------------------------------------------------------------------------- * Function: H5Pget_mpio_actual_io_mode * - * Purpose: Retrieves the type of I/O actually preformed when collective I/O + * Purpose: Retrieves the type of I/O actually performed when collective I/O * is requested. * * Return: Non-negative on success/Negative on failure diff --git a/src/H5Pencdec.c b/src/H5Pencdec.c index 07f2a86..e670d0d 100644 --- a/src/H5Pencdec.c +++ b/src/H5Pencdec.c @@ -663,7 +663,7 @@ done: on failure. DESCRIPTION Decodes a property list from a binary buffer. The contents of the buffer - contain the values for the correponding properties of the plist. The decode + contain the values for the corresponding properties of the plist. The decode callback of a certain property decodes its value from the buffer and sets it in the property list. GLOBAL VARIABLES diff --git a/src/H5Pfapl.c b/src/H5Pfapl.c index e4e742c..b3886da 100644 --- a/src/H5Pfapl.c +++ b/src/H5Pfapl.c @@ -450,7 +450,7 @@ static const size_t H5F_def_page_buf_size_g = H5F_ACS_PAGE_BUFFER_SIZE_DEF; /* static const unsigned H5F_def_page_buf_min_meta_perc_g = H5F_ACS_PAGE_BUFFER_MIN_META_PERC_DEF; /* Default page buffer minimum metadata size */ static const unsigned H5F_def_page_buf_min_raw_perc_g = - H5F_ACS_PAGE_BUFFER_MIN_RAW_PERC_DEF; /* Default page buffer mininum raw data size */ + H5F_ACS_PAGE_BUFFER_MIN_RAW_PERC_DEF; /* Default page buffer minimum raw data size */ static const hbool_t H5F_def_use_file_locking_g = H5F_ACS_USE_FILE_LOCKING_DEF; /* Default use file locking flag */ static const hbool_t H5F_def_ignore_disabled_file_locks_g = @@ -841,7 +841,7 @@ done: * * Purpose: Set the file driver (DRIVER_ID) for a file access * property list (PLIST_ID) and supply an optional - * struct containing the driver-specific properites + * struct containing the driver-specific properties * (DRIVER_INFO). The driver properties will be copied into the * property list and the reference count on the driver will be * incremented, allowing the caller to close the driver ID but @@ -888,7 +888,7 @@ done: * * Purpose: Set the file driver (DRIVER_ID) for a file access * property list (PLIST_ID) and supply an optional - * struct containing the driver-specific properites + * struct containing the driver-specific properties * (DRIVER_INFO). The driver properties will be copied into the * property list and the reference count on the driver will be * incremented, allowing the caller to close the driver ID but @@ -1095,7 +1095,7 @@ done: * * Note: This is an "in-place" copy, since this routine gets called * after the top-level copy has been performed and this routine - * finishes the "deep" part of the copy. + * finishes the "deep" part of the copy. * * Return: Success: Non-negative * Failure: Negative @@ -2299,7 +2299,7 @@ done: * * Currently, the only two valid combinations for this routine are: * LOW = H5F_FORMAT_EARLIEST and HIGH = H5F_FORMAT_LATEST (the default - * setting, which creates objects with the ealiest version possible for + * setting, which creates objects with the earliest version possible for * each object, but no upper limit on the version allowed to be created if * a newer version of an object's format is required to support a feature * requested with an HDF5 library API routine), and LOW = H5F_FORMAT_LATEST @@ -2315,7 +2315,7 @@ done: * Note: Eventually we want to add more values to the H5F_libver_t * enumerated type that indicate library release values where the file * format was changed (like "H5F_FORMAT_1_2_0" for the file format changes - * in the 1.2.x release branch and possily even "H5F_FORMAT_1_4_2" for + * in the 1.2.x release branch and possibly even "H5F_FORMAT_1_4_2" for * a change mid-way through the 1.4.x release branch, etc). * * Adding more values will allow applications to make settings like the @@ -2538,7 +2538,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ @@ -2635,7 +2635,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ @@ -2710,7 +2710,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ @@ -2764,7 +2764,7 @@ H5Pset_file_image_callbacks(hid_t fapl_id, H5FD_file_image_callbacks_t *callback HDassert(callbacks_ptr->udata_copy); HDassert(callbacks_ptr->udata_free); if ((info.callbacks.udata = callbacks_ptr->udata_copy(callbacks_ptr->udata)) == NULL) - HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't copy the suppplied udata") + HGOTO_ERROR(H5E_PLIST, H5E_CANTSET, FAIL, "can't copy the supplied udata") } /* end if */ /* Set values */ @@ -2776,7 +2776,7 @@ done: } /* end H5Pset_file_image_callbacks() */ /*------------------------------------------------------------------------- - * Function: H5Pget_file_image_callbacks + * Function: H5Pget_file_image_callbacks * * Purpose: Sets the callbacks for file images. Some file drivers allow * the use of user-defined callbacks for allocating, freeing and @@ -2786,7 +2786,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ @@ -3188,7 +3188,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ @@ -3218,7 +3218,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ @@ -3317,7 +3317,7 @@ done: * Return: Non-negative on success/Negative on failure * * Programmer: Jacob Gruber - * Thurday, August 11, 2011 + * Thursday, August 11, 2011 * *------------------------------------------------------------------------- */ diff --git a/src/H5Pint.c b/src/H5Pint.c index 1811f94..0d8e09d 100644 --- a/src/H5Pint.c +++ b/src/H5Pint.c @@ -2991,7 +2991,7 @@ done: NAME H5P_exist_plist PURPOSE - Internal routine to query the existance of a property in a property list. + Internal routine to query the existence of a property in a property list. USAGE htri_t H5P_exist_plist(plist, name) const H5P_genplist_t *plist; IN: Property list to check @@ -3050,7 +3050,7 @@ done: NAME H5P__exist_pclass PURPOSE - Internal routine to query the existance of a property in a property class. + Internal routine to query the existence of a property in a property class. USAGE herr_t H5P__exist_pclass(pclass, name) H5P_genclass_t *pclass; IN: Property class to check @@ -3566,7 +3566,7 @@ H5P__cmp_plist_cb(H5P_genprop_t *prop, void *_udata) /* Check if the property exists in the second property list */ if ((prop2_exist = H5P_exist_plist(udata->plist2, prop->name)) < 0) - HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, H5_ITER_ERROR, "can't lookup existance of property?") + HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, H5_ITER_ERROR, "can't lookup existence of property?") if (prop2_exist) { const H5P_genprop_t *prop2; /* Pointer to property in second plist */ @@ -4517,7 +4517,7 @@ done: Returns non-negative on success, negative on failure. DESCRIPTION Removes a property from a property list. Both properties which were - in existance when the property list was created (i.e. properties registered + in existence when the property list was created (i.e. properties registered with H5Pregister2) and properties added to the list after it was created (i.e. added with H5Pinsert2) may be removed from a property list. Properties do not need to be removed a property list before the list itself @@ -4598,7 +4598,7 @@ H5P__copy_prop_plist(hid_t dst_id, hid_t src_id, const char *name) NULL == (dst_plist = (H5P_genplist_t *)H5I_object(dst_id))) HGOTO_ERROR(H5E_PLIST, H5E_NOTFOUND, FAIL, "property object doesn't exist") - /* If the property exists in the destination alread */ + /* If the property exists in the destination already */ if (NULL != H5P__find_prop_plist(dst_plist, name)) { /* Delete the property from the destination list, calling the 'close' callback if necessary */ if (H5P_remove(dst_plist, name) < 0) diff --git a/src/H5Pocpl.c b/src/H5Pocpl.c index c3095f8..42b93d2 100644 --- a/src/H5Pocpl.c +++ b/src/H5Pocpl.c @@ -482,7 +482,7 @@ done: * transfer property list. The FLAGS argument specifies certain * general properties of the filter and is documented below. * The CD_VALUES is an array of CD_NELMTS integers which are - * auxiliary data for the filter. The integer vlues will be + * auxiliary data for the filter. The integer values will be * stored in the dataset object header as part of the filter * information. * @@ -543,7 +543,7 @@ done: * transfer property list. The FLAGS argument specifies certain * general properties of the filter and is documented below. * The CD_VALUES is an array of CD_NELMTS integers which are - * auxiliary data for the filter. The integer vlues will be + * auxiliary data for the filter. The integer values will be * stored in the dataset object header as part of the filter * information. * @@ -609,7 +609,7 @@ done: * creation property list. The FLAGS argument specifies certain * general properties of the filter and is documented below. * The CD_VALUES is an array of CD_NELMTS integers which are - * auxiliary data for the filter. The integer vlues will be + * auxiliary data for the filter. The integer values will be * stored in the dataset object header as part of the filter * information. * @@ -671,7 +671,7 @@ done: * creation property list. The FLAGS argument specifies certain * general properties of the filter and is documented below. * The CD_VALUES is an array of CD_NELMTS integers which are - * auxiliary data for the filter. The integer vlues will be + * auxiliary data for the filter. The integer values will be * stored in the dataset object header as part of the filter * information. * @@ -782,7 +782,7 @@ done: * CD_NELMTS indicates the number of entries in the CD_VALUES * array allocated by the caller while on exit it contains the * number of values defined by the filter. FILTER_CONFIG is a bit - * field contaning encode/decode flags from H5Zpublic.h. The IDX + * field containing encode/decode flags from H5Zpublic.h. The IDX * should be a value between zero and N-1 as described for * H5Pget_nfilters() and the function will return failure if the * filter number is out of range. @@ -867,7 +867,7 @@ done: * CD_NELMTS indicates the number of entries in the CD_VALUES * array allocated by the caller while on exit it contains the * number of values defined by the filter. FILTER_CONFIG is a bit - * field contaning encode/decode flags from H5Zpublic.h. The ID + * field containing encode/decode flags from H5Zpublic.h. The ID * should be the filter ID to retrieve the parameters for. If the * filter is not set for the property list, an error will be returned. * @@ -916,7 +916,7 @@ done: * CD_NELMTS indicates the number of entries in the CD_VALUES * array allocated by the caller while on exit it contains the * number of values defined by the filter. FILTER_CONFIG is a bit - * field contaning encode/decode flags from H5Zpublic.h. The ID + * field containing encode/decode flags from H5Zpublic.h. The ID * should be the filter ID to retrieve the parameters for. If the * filter is not set for the property list, an error will be returned. * diff --git a/src/H5Rdeprec.c b/src/H5Rdeprec.c index cb54248..1fb32b0 100644 --- a/src/H5Rdeprec.c +++ b/src/H5Rdeprec.c @@ -85,7 +85,7 @@ * ref_type IN: Type of reference to query * ref IN: Reference to query * - * Return: Success: An object type (as defined in H5Gpublic.h) + * Return: Success: An object type (as defined in H5Gpublic.h) * Failure: H5G_UNKNOWN * *------------------------------------------------------------------------- @@ -1962,7 +1962,7 @@ done: htri_t H5S_extent_equal(ds1, ds2) H5S_t *ds1, *ds2; IN: Dataspace objects to compare RETURNS - TRUE if equal, FALSE if unequal on succeess/Negative on failure + TRUE if equal, FALSE if unequal on success/Negative on failure DESCRIPTION Compare two dataspaces if their extents are identical. --------------------------------------------------------------------------*/ @@ -289,7 +289,7 @@ } /* Macro used to reduce the level of a node by 1. Does not update the head node - * "current level". PREV is the previous node of the currrent height of X. */ + * "current level". PREV is the previous node of the current height of X. */ #define H5SL_DEMOTE(X, PREV) \ { \ size_t _lvl = X->level; \ @@ -305,7 +305,7 @@ #define H5SL_INSERT(CMP, SLIST, X, TYPE, KEY, HASHVAL) \ { \ H5SL_node_t *_last = X; /* Lowest node in the current gap */ \ - H5SL_node_t *_next = NULL; /* Highest node in the currect gap */ \ + H5SL_node_t *_next = NULL; /* Highest node in the current gap */ \ H5SL_node_t *_drop; /* Low node of the gap to drop into */ \ int _count; /* Number of nodes in the current gap */ \ int _i; \ @@ -365,7 +365,7 @@ else { \ H5SL_node_t *_last = X; /* Lowest node in the current gap */ \ H5SL_node_t *_llast = X; /* Lowest node in the previous gap */ \ - H5SL_node_t *_next = NULL; /* Highest node in the currect gap */ \ + H5SL_node_t *_next = NULL; /* Highest node in the current gap */ \ H5SL_node_t *_drop = NULL; /* Low node of the gap to drop into */ \ H5SL_node_t *_ldrop = NULL; /* Low node of gap before the one to drop into */ \ H5SL_node_t *_head = SLIST->header; /* Head of the skip list */ \ @@ -2028,7 +2028,7 @@ H5SL_next(H5SL_node_t *slist_node) NAME H5SL_prev PURPOSE - Gets a pointer to the previos node in a skip list + Gets a pointer to the previous node in a skip list USAGE H5SL_node_t *H5SL_prev(slist_node) H5SL_node_t *slist_node; IN: Pointer to skip list node @@ -116,7 +116,7 @@ H5SM_init(H5F_t *f, H5P_genplist_t *fc_plist, const H5O_loc_t *ext_loc) haddr_t table_addr = HADDR_UNDEF; /* Address of SOHM master table in file */ unsigned list_max, btree_min; /* Phase change limits for SOHM indices */ unsigned index_type_flags[H5O_SHMESG_MAX_NINDEXES]; /* Messages types stored in each index */ - unsigned minsizes[H5O_SHMESG_MAX_NINDEXES]; /* Message size sharing threshhold for each index */ + unsigned minsizes[H5O_SHMESG_MAX_NINDEXES]; /* Message size sharing threshold for each index */ unsigned type_flags_used; /* Message type flags used, for sanity checking */ unsigned x; /* Local index variable */ herr_t ret_value = SUCCEED; /* Return value */ @@ -1560,7 +1560,7 @@ H5SM_delete(H5F_t *f, H5O_t *open_oh, H5O_shared_t *sh_mesg) */ if (H5SM__delete_from_index(f, open_oh, &(table->indexes[index_num]), sh_mesg, &cache_flags, &mesg_size, &mesg_buf) < 0) - HGOTO_ERROR(H5E_SOHM, H5E_CANTDELETE, FAIL, "unable to delete mesage from SOHM index") + HGOTO_ERROR(H5E_SOHM, H5E_CANTDELETE, FAIL, "unable to delete message from SOHM index") /* Release the master SOHM table */ if (H5AC_unprotect(f, H5AC_SOHM_TABLE, H5F_SOHM_ADDR(f), table, cache_flags) < 0) diff --git a/src/H5SMcache.c b/src/H5SMcache.c index 5e66dbc..e9c6a41 100644 --- a/src/H5SMcache.c +++ b/src/H5SMcache.c @@ -15,7 +15,7 @@ * * Created: H5SMcache.c * Nov 13 2006 - * James Laird <jlaird@hdfgroup.org> + * James Laird * * Purpose: Implement shared message metadata cache methods. * diff --git a/src/H5Shyper.c b/src/H5Shyper.c index 5652a60..b07696f 100644 --- a/src/H5Shyper.c +++ b/src/H5Shyper.c @@ -2059,7 +2059,7 @@ H5S__hyper_iter_get_seq_list_opt(H5S_sel_iter_t *iter, size_t maxseq, size_t max /* Increment the offset and count for the other dimensions */ temp_dim = (int)fast_dim - 1; while (temp_dim >= 0) { - /* Move to the next row in the curent dimension */ + /* Move to the next row in the current dimension */ offset[temp_dim]++; tmp_block[temp_dim]++; @@ -2170,7 +2170,7 @@ H5S__hyper_iter_get_seq_list_opt(H5S_sel_iter_t *iter, size_t maxseq, size_t max /* Increment the offset and count for the other dimensions */ temp_dim = (int)fast_dim - 1; while (temp_dim >= 0) { - /* Move to the next row in the curent dimension */ + /* Move to the next row in the current dimension */ offset[temp_dim]++; tmp_block[temp_dim]++; @@ -2484,7 +2484,7 @@ H5S__hyper_iter_get_seq_list_single(H5S_sel_iter_t *iter, size_t maxseq, size_t /* Increment the offset and count for the other dimensions */ while (temp_dim >= 0) { - /* Move to the next row in the curent dimension */ + /* Move to the next row in the current dimension */ offset[temp_dim]++; tmp_block[temp_dim]++; @@ -3531,9 +3531,9 @@ H5S__hyper_get_version_enc_size(H5S_t *space, hsize_t block_count, uint32_t *ver hbool_t bound_up_version = FALSE; /* Whether high bounds exceed (2^32 - 1) */ H5F_libver_t low_bound; /* The 'low' bound of library format versions */ H5F_libver_t high_bound; /* The 'high' bound of library format versions */ - unsigned u; /* Local index veriable */ - uint32_t tmp_version; /* Temporay version */ - herr_t ret_value = SUCCEED; /* return value */ + uint32_t tmp_version; /* Local temporary version */ + unsigned u; /* Local index variable */ + herr_t ret_value = SUCCEED; /* Return value */ FUNC_ENTER_STATIC @@ -3831,7 +3831,7 @@ H5S__hyper_serialize(H5S_t *space, uint8_t **p) pp = (*p); HDassert(pp); - /* Set some convienence values */ + /* Set some convenience values */ ndims = space->extent.rank; diminfo = space->select.sel_info.hslab->diminfo.opt; @@ -3896,7 +3896,7 @@ H5S__hyper_serialize(H5S_t *space, uint8_t **p) else { HDassert(version == H5S_HYPER_VERSION_1); - /* Set some convienence values */ + /* Set some convenience values */ fast_dim = ndims - 1; /* Encode number of hyperslabs */ @@ -4034,7 +4034,7 @@ H5S__hyper_deserialize(H5S_t **space, const uint8_t **p) { H5S_t *tmp_space = NULL; /* Pointer to actual dataspace to use, either *space or a newly allocated one */ - hsize_t dims[H5S_MAX_RANK]; /* Dimenion sizes */ + hsize_t dims[H5S_MAX_RANK]; /* Dimension sizes */ hsize_t start[H5S_MAX_RANK]; /* hyperslab start information */ hsize_t block[H5S_MAX_RANK]; /* hyperslab block information */ uint32_t version; /* Version number */ @@ -4396,7 +4396,7 @@ H5S__get_select_hyper_blocklist(H5S_t *space, hsize_t startblock, hsize_t numblo hbool_t done; /* Whether we are done with the iteration */ unsigned u; /* Counter */ - /* Set some convienence values */ + /* Set some convenience values */ ndims = space->extent.rank; fast_dim = ndims - 1; @@ -4890,7 +4890,7 @@ H5S__hyper_is_contiguous(const H5S_t *space) large_contiguous = TRUE; /* assume true and reset if the dimensions don't match */ small_contiguous = FALSE; /* assume false initially */ - /* Check for a "large contigous" block */ + /* Check for a "large contiguous" block */ for (u = 0; u < space->extent.rank; u++) { if (diminfo[u].count > 1) { large_contiguous = FALSE; @@ -11295,7 +11295,7 @@ H5S__hyper_proj_int_iterate(H5S_hyper_span_info_t *ss_span_info, const H5S_hyper udata->skip += (count - 1) * (udata->skip - old_skip); } /* end if */ else { - /* Third case: agorithm added skip and nelem (in that + /* Third case: algorithm added skip and nelem (in that * order). Add the same skip and nelem once for each item * remaining in count. */ hsize_t skip_add; diff --git a/src/H5Smpio.c b/src/H5Smpio.c index 3f9137b..7d8f396 100644 --- a/src/H5Smpio.c +++ b/src/H5Smpio.c @@ -1170,7 +1170,7 @@ H5S__obtain_datatype(H5S_hyper_span_info_t *spans, const hsize_t *down, size_t e /* If this is the fastest changing dimension, it is the base case for derived datatype. */ span = spans->head; if (NULL == span->down) { - hbool_t large_block = FALSE; /* Wether the block length is larger than 32 bit integer */ + hbool_t large_block = FALSE; /* Whether the block length is larger than 32 bit integer */ outercount = 0; while (span) { diff --git a/src/H5Spoint.c b/src/H5Spoint.c index 19f3ec9..f1540f2 100644 --- a/src/H5Spoint.c +++ b/src/H5Spoint.c @@ -1037,7 +1037,7 @@ H5S__point_get_version_enc_size(const H5S_t *space, uint32_t *version, uint8_t * hbool_t exceed = FALSE; hsize_t bounds_start[H5S_MAX_RANK]; /* Starting coordinate of bounding box */ hsize_t bounds_end[H5S_MAX_RANK]; /* Opposite coordinate of bounding box */ - unsigned u; /* Local index veriable */ + unsigned u; /* Local index variable */ herr_t ret_value = SUCCEED; /* Return value */ FUNC_ENTER_STATIC @@ -1350,7 +1350,7 @@ done: dataspace they are located within. The list of points is formatted as follows: <coordinate> followed by the next coordinate, etc. until all the point information in the selection have been put into the user's buffer. - The points are returned in the order they will be interated through + The points are returned in the order they will be iterated through when a selection is read/written from/to disk. GLOBAL VARIABLES COMMENTS, BUGS, ASSUMPTIONS @@ -1424,7 +1424,7 @@ H5S__get_select_elem_pointlist(const H5S_t *space, hsize_t startpoint, hsize_t n dataspace they are located within. The list of points is formatted as follows: <coordinate> followed by the next coordinate, etc. until all the point information in the selection have been put into the user's buffer. - The points are returned in the order they will be interated through + The points are returned in the order they will be iterated through when a selection is read/written from/to disk. GLOBAL VARIABLES COMMENTS, BUGS, ASSUMPTIONS diff --git a/src/H5Sselect.c b/src/H5Sselect.c index 894c307..9456e58 100644 --- a/src/H5Sselect.c +++ b/src/H5Sselect.c @@ -2239,7 +2239,7 @@ H5S_select_construct_projection(H5S_t *base_space, H5S_t **new_space_ptr, unsign if (new_space_rank == 0) { hssize_t npoints; /* Number of points selected */ - /* Retreve the number of elements selected */ + /* Retrieve the number of elements selected */ if ((npoints = (hssize_t)H5S_GET_SELECT_NPOINTS(base_space)) < 0) HGOTO_ERROR(H5E_DATASPACE, H5E_CANTGET, FAIL, "unable to get number of points selected") HDassert(npoints <= 1); @@ -2308,12 +2308,12 @@ H5S_select_construct_projection(H5S_t *base_space, H5S_t **new_space_ptr, unsign * true on selections of different rank iff: * * 1) the selection in the lower rank dataspace matches that - * in the dimensions with the fastest changing indicies in + * in the dimensions with the fastest changing indices in * the larger rank dataspace, and * * 2) the selection has thickness 1 in all ranks that appear * only in the higher rank dataspace (i.e. those with - * more slowly changing indicies). + * more slowly changing indices). */ if (new_space_rank > base_space_rank) { hsize_t tmp_dim_size = 1; /* Temporary dimension value, for filling arrays */ @@ -2476,7 +2476,7 @@ H5S_select_fill(const void *fill, size_t fill_size, H5S_t *space, void *_buf) /* Loop, while elements left in selection */ while (max_elem > 0) { size_t nseq; /* Number of sequences generated */ - size_t curr_seq; /* Current sequnce being worked on */ + size_t curr_seq; /* Current sequence being worked on */ size_t nelem; /* Number of elements used in sequences */ /* Get the sequences of bytes */ diff --git a/src/H5Stest.c b/src/H5Stest.c index a851549..6a31f3c 100644 --- a/src/H5Stest.c +++ b/src/H5Stest.c @@ -185,12 +185,13 @@ H5S__check_spans_tail_ptr(const H5S_hyper_span_info_t *span_lst) if (NULL != cur_elem->down) if ((ret_value = H5S__check_spans_tail_ptr(cur_elem->down)) < 0) HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, - "the seletion has inconsistent tail pointers") + "the selection has inconsistent tail pointers") cur_elem = cur_elem->next; } /* end while */ if (actual_tail != span_lst->tail) - HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the seletion has inconsistent tail pointers") + HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, + "the selection has inconsistent tail pointers") done: FUNC_LEAVE_NOAPI(ret_value) @@ -231,7 +232,8 @@ H5S__check_points_tail_ptr(const H5S_pnt_list_t *pnt_lst) cur_elem = cur_elem->next; } /* end while */ if (actual_tail != pnt_lst->tail) - HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, "the seletion has inconsistent tail pointers") + HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, + "the selection has inconsistent tail pointers") done: FUNC_LEAVE_NOAPI(ret_value) @@ -314,7 +316,7 @@ H5S__check_internal_consistency(const H5S_t *space) if ((NULL != hslab) && (NULL != hslab->span_lst)) if (H5S__check_spans_tail_ptr(hslab->span_lst) < 0) HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, - "the seletion has inconsistent tail pointers") + "the selection has inconsistent tail pointers") } /* end if */ else if (space->select.type->type == H5S_SEL_POINTS) { H5S_pnt_list_t *pnt_lst = space->select.sel_info.pnt_lst; @@ -322,7 +324,7 @@ H5S__check_internal_consistency(const H5S_t *space) if (NULL != pnt_lst) if (H5S__check_points_tail_ptr(pnt_lst) < 0) HGOTO_ERROR(H5E_DATASPACE, H5E_INCONSISTENTSTATE, FAIL, - "the seletion has inconsistent tail pointers") + "the selection has inconsistent tail pointers") } /* end else-if */ done: @@ -2646,7 +2646,7 @@ done: * All arguments are optional. Missing arguments are wild cards. * The special no-op path cannot be removed. * - * Return: Succeess: non-negative + * Return: Success: non-negative * Failure: negative * * Programmer: Robb Matzke @@ -2750,7 +2750,7 @@ H5T__unregister(H5T_pers_t pers, const char *name, H5T_t *src, H5T_t *dst, H5T_c * All arguments are optional. Missing arguments are wild cards. * The special no-op path cannot be removed. * - * Return: Succeess: non-negative + * Return: Success: non-negative * * Failure: negative * @@ -4064,7 +4064,7 @@ done: * * Return: Success: non-negative * - * Failure: nagative + * Failure: negative * * Programmer: Robb Matzke * Tuesday, December 22, 1998 @@ -20,7 +20,7 @@ /* Module specific data structures */ -/* cancelability structure */ +/* Cancellability structure */ typedef struct H5TS_cancel_struct { int previous_state; unsigned int cancel_count; @@ -375,12 +375,12 @@ H5TS_mutex_unlock(H5TS_mutex_t *mutex) * 0 on success non-zero error code on error. * * DESCRIPTION - * Creates a cancelation counter for a thread if it is the first time + * Creates a cancellation counter for a thread if it is the first time * the thread is entering the library. * - * if counter value is zero, then set cancelability type of the thread + * if counter value is zero, then set cancellability type of the thread * to PTHREAD_CANCEL_DISABLE as thread is entering the library and store - * the previous cancelability type into cancelation counter. + * the previous cancellability type into cancellation counter. * Increase the counter value by 1. * * PROGRAMMER: Chee Wai LEE @@ -439,8 +439,8 @@ H5TS_cancel_count_inc(void) * 0 on success and a non-zero error code on error. * * DESCRIPTION - * If counter value is one, then set cancelability type of the thread - * to the previous cancelability type stored in the cancelation counter. + * If counter value is one, then set cancellability type of the thread + * to the previous cancellability type stored in the cancellation counter. * (the thread is leaving the library). * * Decrement the counter value by 1. diff --git a/src/H5Tbit.c b/src/H5Tbit.c index 04e2a97..5a3ce1e 100644 --- a/src/H5Tbit.c +++ b/src/H5Tbit.c @@ -228,7 +228,7 @@ done: * Purpose: Return a small bit sequence as a number. Bit vector starts * at OFFSET and is SIZE bits long. * - * Return: The bit sequence interpretted as an unsigned integer + * Return: The bit sequence interpreted as an unsigned integer * *------------------------------------------------------------------------- */ @@ -372,7 +372,7 @@ H5T__bit_set(uint8_t *buf, size_t offset, size_t size, hbool_t value) * Purpose: Finds the first bit with the specified VALUE within a region * of a bit vector. The region begins at OFFSET and continues * for SIZE bits, but the region can be searched from the least - * significat end toward the most significant end(H5T_BIT_LSB + * significant end toward the most significant end(H5T_BIT_LSB * as DIRECTION), or from the most significant end to the least * significant end(H5T_BIT_MSB as DIRECTION). * @@ -538,7 +538,7 @@ H5T__bit_inc(uint8_t *buf, size_t start, size_t size) /*------------------------------------------------------------------------- * Function: H5T__bit_dec * - * Purpose: Decrement part of a bit field by substracting 1. The bit + * Purpose: Decrement part of a bit field by subtracting 1. The bit * field starts with bit position START and is SIZE bits long. * * Return: The "borrow-in" value. It's TRUE if underflows, FALSE @@ -564,9 +564,9 @@ H5T__bit_dec(uint8_t *buf, size_t start, size_t size) if ((size + start - 1) / 8 > idx) { /* The bit sequence doesn't end in the same byte as starts */ - /* Example: a sequence like 11000100 and start = 3. We substract 00001000 from + /* Example: a sequence like 11000100 and start = 3. We subtract 00001000 from * it and get 10111100. If a sequence is 00000111, we do right shift for START - * bits and get 00000000. So we need to borrow from higher byte when we substract + * bits and get 00000000. So we need to borrow from higher byte when we subtract * 00001000. */ if (!(buf[idx] >> pos)) @@ -597,7 +597,7 @@ H5T__bit_dec(uint8_t *buf, size_t start, size_t size) else { /* The bit sequence ends in the same byte as starts */ - /* Example: a sequence like 11000100 and pos=3, size=3. We substract 00001000 + /* Example: a sequence like 11000100 and pos=3, size=3. We subtract 00001000 * and get 10111100. A bit is borrowed from 6th bit(buf[idx]>>6=00000010, tmp>>6=00000011, * not equal). We need to put this bit back by increment 1000000. */ diff --git a/src/H5Tconv.c b/src/H5Tconv.c index aff6762..5e4c9fe 100644 --- a/src/H5Tconv.c +++ b/src/H5Tconv.c @@ -2008,7 +2008,7 @@ H5T_conv_struct_init(H5T_t *src, H5T_t *dst, H5T_cdata_t *cdata) priv->subset_info.copy_size = 0; /* - * Insure that members are sorted. + * Ensure that members are sorted. */ H5T__sort_value(src, NULL); H5T__sort_value(dst, NULL); @@ -3105,7 +3105,7 @@ H5T__conv_vlen(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelmts, si H5T_path_t *tpath = NULL; /* Type conversion path */ hbool_t noop_conv = FALSE; /* Flag to indicate a noop conversion */ hbool_t write_to_file = FALSE; /* Flag to indicate writing to file */ - htri_t parent_is_vlen; /* Flag to indicate parent is vlen datatyp */ + htri_t parent_is_vlen; /* Flag to indicate parent is vlen datatype */ hid_t tsrc_id = -1, tdst_id = -1; /*temporary type atoms */ H5T_t *src = NULL; /*source datatype */ H5T_t *dst = NULL; /*destination datatype */ @@ -4644,7 +4644,7 @@ H5T__conv_s_s(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelmts, siz size_t olap; /*num overlapping elements */ size_t nchars = 0; /*number of characters copied */ uint8_t *s, *sp, *d, *dp; /*src and dst traversal pointers*/ - uint8_t *dbuf = NULL; /*temp buf for overlap convers. */ + uint8_t *dbuf = NULL; /*temp buf for overlap converts. */ herr_t ret_value = SUCCEED; /* Return value */ FUNC_ENTER_PACKAGE @@ -4854,7 +4854,7 @@ H5T__conv_s_s(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelmts, siz break; default: - HGOTO_ERROR(H5E_DATATYPE, H5E_UNSUPPORTED, FAIL, "unknown converson command") + HGOTO_ERROR(H5E_DATATYPE, H5E_UNSUPPORTED, FAIL, "unknown conversion command") } /* end switch */ done: @@ -6930,7 +6930,7 @@ H5T__conv_schar_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -6990,7 +6990,7 @@ H5T__conv_uchar_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7050,7 +7050,7 @@ H5T__conv_short_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7110,7 +7110,7 @@ H5T__conv_ushort_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7170,7 +7170,7 @@ H5T__conv_int_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nelm * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7230,7 +7230,7 @@ H5T__conv_uint_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7290,7 +7290,7 @@ H5T__conv_long_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7350,7 +7350,7 @@ H5T__conv_ulong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7410,7 +7410,7 @@ H5T__conv_llong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7472,7 +7472,7 @@ H5T__conv_ullong_double(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7582,7 +7582,7 @@ H5T__conv_double_uchar(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7604,7 +7604,7 @@ H5T__conv_ldouble_schar(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7714,7 +7714,7 @@ H5T__conv_double_ushort(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7736,7 +7736,7 @@ H5T__conv_ldouble_short(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7846,7 +7846,7 @@ H5T__conv_double_uint(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7868,7 +7868,7 @@ H5T__conv_ldouble_int(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t nel * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -7978,7 +7978,7 @@ H5T__conv_double_ulong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -8000,7 +8000,7 @@ H5T__conv_ldouble_long(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t ne * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -8110,7 +8110,7 @@ H5T__conv_double_ullong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ @@ -8134,7 +8134,7 @@ H5T__conv_ldouble_llong(hid_t src_id, hid_t dst_id, H5T_cdata_t *cdata, size_t n * Return: Non-negative on success/Negative on failure * * Programmer: Raymond Lu - * Tuesday, Febuary 1, 2005 + * Tuesday, February 1, 2005 * *------------------------------------------------------------------------- */ diff --git a/src/H5Tfields.c b/src/H5Tfields.c index 8b830fc..249a025 100644 --- a/src/H5Tfields.c +++ b/src/H5Tfields.c @@ -12,7 +12,7 @@ * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ /* - * Module Info: This module contains commond functionality for fields in + * Module Info: This module contains command functionality for fields in * enumerated & compound datatypes in the H5T interface. */ diff --git a/src/H5Tnative.c b/src/H5Tnative.c index dc9c64f..d4a44de 100644 --- a/src/H5Tnative.c +++ b/src/H5Tnative.c @@ -43,7 +43,7 @@ static herr_t H5T__cmp_offset(size_t *comp_size, size_t *offset, size_t elem_siz * * Purpose: High-level API to return the native type of a datatype. * The native type is chosen by matching the size and class of - * querried datatype from the following native premitive + * queried datatype from the following native primitive * datatypes: * H5T_NATIVE_CHAR H5T_NATIVE_UCHAR * H5T_NATIVE_SHORT H5T_NATIVE_USHORT @@ -56,7 +56,7 @@ static herr_t H5T__cmp_offset(size_t *comp_size, size_t *offset, size_t elem_siz * H5T_NATIVE_LDOUBLE * * Compound, array, enum, and VL types all choose among these - * types for theire members. Time, Bifield, Opaque, Reference + * types for their members. Time, Bitfield, Opaque, Reference * types are only copy out. * * Return: Success: Returns the native data type if successful. @@ -687,7 +687,7 @@ H5_GCC_DIAG_OFF("duplicated-branches") /*------------------------------------------------------------------------- * Function: H5T__get_native_float * - * Purpose: Returns the native floatt type of a datatype. + * Purpose: Returns the native float type of a datatype. * * Return: Success: Returns the native data type if successful. * diff --git a/src/H5Toffset.c b/src/H5Toffset.c index b206dca..cccede6 100644 --- a/src/H5Toffset.c +++ b/src/H5Toffset.c @@ -70,7 +70,7 @@ H5Tget_offset(hid_t type_id) /* Get offset */ if ((ret_value = H5T_get_offset(dt)) < 0) - HGOTO_ERROR(H5E_DATATYPE, H5E_UNSUPPORTED, FAIL, "cant't get offset for specified datatype") + HGOTO_ERROR(H5E_DATATYPE, H5E_UNSUPPORTED, FAIL, "can't get offset for specified datatype") done: FUNC_LEAVE_API(ret_value) diff --git a/src/H5Topaque.c b/src/H5Topaque.c index 710121b..15753fe 100644 --- a/src/H5Topaque.c +++ b/src/H5Topaque.c @@ -70,7 +70,7 @@ done: /*------------------------------------------------------------------------- * Function: H5Tget_tag * - * Purpose: Get tha tag associated with an opaque datatype. + * Purpose: Get the tag associated with an opaque datatype. * * Return: A pointer to an allocated string. The caller should free * the string. NULL is returned for errors. diff --git a/src/H5Tprecis.c b/src/H5Tprecis.c index 03d20f2..07b868e 100644 --- a/src/H5Tprecis.c +++ b/src/H5Tprecis.c @@ -59,7 +59,7 @@ H5Tget_precision(hid_t type_id) /* Get precision */ if ((ret_value = H5T_get_precision(dt)) == 0) - HGOTO_ERROR(H5E_DATATYPE, H5E_UNSUPPORTED, 0, "cant't get precision for specified datatype") + HGOTO_ERROR(H5E_DATATYPE, H5E_UNSUPPORTED, 0, "can't get precision for specified datatype") done: FUNC_LEAVE_API(ret_value) diff --git a/src/H5Tpublic.h b/src/H5Tpublic.h index 49a2d60..39d24ed 100644 --- a/src/H5Tpublic.h +++ b/src/H5Tpublic.h @@ -1081,7 +1081,7 @@ H5_DLLVAR hid_t H5T_NATIVE_UINT_FAST64_g; * When creating a fixed-length string datatype, \p size will * be the length of the string in bytes. The length of the * string in characters will depend on i the encoding used; see - * H5Pset_char_encoding(). + * #H5Pset_char_encoding. * * ENUMs created with this function have a signed native integer * base datatype. Use H5Tenum_create() if a different integer base @@ -1253,7 +1253,7 @@ H5VM_chunk_index_scaled(unsigned ndims, const hsize_t *coord, const uint32_t *ch * Function: H5VM_opvv * * Purpose: Perform an operation on a source & destination sequences - * of offset/length pairs. Each set of sequnces has an array + * of offset/length pairs. Each set of sequences has an array * of lengths, an array of offsets, the maximum number of * sequences and the current sequence to start at in the sequence. * diff --git a/src/H5VMprivate.h b/src/H5VMprivate.h index 0d3bd0f..6ea0d5a 100644 --- a/src/H5VMprivate.h +++ b/src/H5VMprivate.h @@ -435,7 +435,7 @@ static const unsigned MultiplyDeBruijnBitPosition[32] = {0, 1, 28, 2, 29, 14, * Return: log2(n) (always - no failure condition) * * Programmer: Quincey Koziol - * Monday, Febraury 27, 2006 + * Monday, February 27, 2006 * *------------------------------------------------------------------------- */ @@ -18,7 +18,7 @@ * Quincey Koziol * * Purpose: Implements the "wrapped buffer" code for wrapping - * an existing [staticly sized] buffer, in order to + * an existing [statically sized] buffer, in order to * avoid lots of memory allocation calls. * *------------------------------------------------------------------------- @@ -1502,7 +1502,7 @@ done: /*------------------------------------------------------------------------- * Function: H5Z_filter_in_pline * - * Purpose: Check wheter a filter is in the filter pipeline using the + * Purpose: Check whether a filter is in the filter pipeline using the * filter ID. This function is very similar to H5Z_filter_info * * Return: TRUE - found filter diff --git a/src/H5Znbit.c b/src/H5Znbit.c index c5d3250..b787c8e 100644 --- a/src/H5Znbit.c +++ b/src/H5Znbit.c @@ -1015,7 +1015,7 @@ done: * assume one byte has 8 bit * assume padding bit is 0 * assume size of unsigned char is one byte - * assume one data item of certain datatype is stored continously in bytes + * assume one data item of certain datatype is stored continuously in bytes * atomic datatype is treated on byte basis */ diff --git a/src/H5Zscaleoffset.c b/src/H5Zscaleoffset.c index 2c4ee28..ce43369 100644 --- a/src/H5Zscaleoffset.c +++ b/src/H5Zscaleoffset.c @@ -660,7 +660,7 @@ H5Z_class2_t H5Z_SCALEOFFSET[1] = {{ buf[i] = (type)(buf[i] + (type)(minval)); \ } while (0) -/* Retrive minimum value of floating-point type */ +/* Retrieve minimum value of floating-point type */ #define H5Z_scaleoffset_get_min(type, minval, min) \ { \ if (sizeof(type) <= sizeof(long long)) \ @@ -1391,7 +1391,7 @@ done: * assume one byte has 8 bit * assume padding bit is 0 * assume size of unsigned char is one byte - * assume one data item of certain datatype is stored continously in bytes + * assume one data item of certain datatype is stored continuously in bytes * atomic datatype is treated on byte basis */ diff --git a/src/H5Ztrans.c b/src/H5Ztrans.c index 09dabb1..fd759c2 100644 --- a/src/H5Ztrans.c +++ b/src/H5Ztrans.c @@ -423,7 +423,7 @@ H5Z__unget_token(H5Z_token *current) * kept internal to the H5Z_token and handled by this and the * unget_H5Z_token function. * - * Return: Succeess: The passed in H5Z_token with a valid tok_type + * Return: Success: The passed in H5Z_token with a valid tok_type * field. * Failure: The passed in H5Z_token but with the tok_type * field set to ERROR. @@ -1416,7 +1416,7 @@ H5Z__op_is_numbs2(H5Z_node *_tree) * Function: H5Z__xform_reduce_tree * * Purpose: Simplifies parse tree passed in by performing any obvious - * and trivial arithemtic calculations. + * and trivial arithmetic calculations. * * Return: None. * @@ -1756,7 +1756,7 @@ H5Z_xform_noop(const H5Z_data_xform_t *data_xform_prop) if (data_xform_prop) { ret_value = FALSE; - /* Check for trivial data tranformation: expression = "x" */ + /* Check for trivial data transformation: expression = "x" */ if ((HDstrlen(data_xform_prop->xform_exp) == 1) && data_xform_prop->dat_val_pointers && (data_xform_prop->dat_val_pointers->num_ptrs == 1)) { ret_value = TRUE; @@ -1787,7 +1787,7 @@ H5Z_xform_extract_xform_str(const H5Z_data_xform_t *data_xform_prop) /* There should be no way that this can be NULL since the function * that calls this one checks to make sure it isn't before - * pasing them */ + * passing them */ HDassert(data_xform_prop); FUNC_LEAVE_NOAPI(data_xform_prop->xform_exp) diff --git a/src/H5checksum.c b/src/H5checksum.c index 0de6c83..ec6e323 100644 --- a/src/H5checksum.c +++ b/src/H5checksum.c @@ -110,7 +110,7 @@ H5_checksum_fletcher32(const void *_data, size_t _len) HDassert(_len > 0); /* Compute checksum for pairs of bytes */ - /* (the magic "360" value is is the largest number of sums that can be + /* (the magic "360" value is the largest number of sums that can be * performed without numeric overflow) */ while (len) { diff --git a/src/H5detect.c b/src/H5detect.c index 90d4622..6fddb54 100644 --- a/src/H5detect.c +++ b/src/H5detect.c @@ -774,7 +774,7 @@ done:\n\ fprintf(rawoutstream, "/* sigprocmask() support: no */\n"); #endif - /* Print the statics of signal handlers called for debugging */ + /* Print the statistics of signal handlers called for debugging */ fprintf(rawoutstream, "\n" "/******************************/\n" "/* signal handlers statistics */\n" @@ -1565,7 +1565,7 @@ verify_signal_handlers(int signum, void (*handler)(int)) } else { if (val == signum) { - /* return from signum handler. Record a sucess. */ + /* return from signum handler. Record a success. */ nsuccesses++; } else { diff --git a/src/H5make_libsettings.c b/src/H5make_libsettings.c index c65f7f3..34d35b4 100644 --- a/src/H5make_libsettings.c +++ b/src/H5make_libsettings.c @@ -11,7 +11,7 @@ * help@hdfgroup.org. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ -/*keep this declaration near the top of this file -RPM*/ +/* Keep this declaration near the top of this file */ static const char *FileHeader = "\n\ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * *\n\ * Copyright by The HDF Group. *\n\ @@ -101,7 +101,7 @@ insert_libhdf5_settings(FILE *flibinfo) HDexit(EXIT_FAILURE); } #else - /* print variable definition and an empty string */ + /* Print variable definition and an empty string */ /* Do not use const else AIX strings does not show it. */ fprintf(flibinfo, "char H5libhdf5_settings[]=\"\";\n"); #endif @@ -120,7 +120,7 @@ insert_libhdf5_settings(FILE *flibinfo) static void make_libinfo(void) { - /* print variable definition and then the string as a macro. */ + /* Print variable definition and then the string as a macro */ insert_libhdf5_settings(rawoutstream); } diff --git a/src/H5mpi.c b/src/H5mpi.c index 013fc83..0f0c05f 100644 --- a/src/H5mpi.c +++ b/src/H5mpi.c @@ -38,7 +38,7 @@ static hsize_t bigio_count = H5_MAX_MPI_COUNT; /*------------------------------------------------------------------------- * Function: H5_mpi_set_bigio_count * - * Purpose: Allow us to programatically change the switch point + * Purpose: Allow us to programmatically change the switch point * when we utilize derived datatypes. This is of * particular interest for allowing nightly testing * diff --git a/src/H5private.h b/src/H5private.h index a94049f..1a20b62 100644 --- a/src/H5private.h +++ b/src/H5private.h @@ -60,7 +60,7 @@ /* * The `struct stat' data type for stat() and fstat(). This is a POSIX file - * but often apears on non-POSIX systems also. The `struct stat' is required + * but often appears on non-POSIX systems also. The `struct stat' is required * for HDF5 to compile, although only a few fields are actually used. */ #ifdef H5_HAVE_SYS_STAT_H @@ -87,7 +87,7 @@ #endif /* - * Longjumps are used to detect alignment constrants + * Longjumps are used to detect alignment constraints */ #ifdef H5_HAVE_SETJMP_H #include <setjmp.h> diff --git a/src/H5public.h b/src/H5public.h index ba65a83..d03bf76 100644 --- a/src/H5public.h +++ b/src/H5public.h @@ -674,7 +674,7 @@ H5_DLL herr_t H5get_libversion(unsigned *majnum, unsigned *minnum, unsigned *rel * currently linked. If this check fails, H5check_version() causes the * application to abort (by means of a standard C abort() call) and * prints information that is usually useful for debugging. This - * precaution is is taken to avoid the risks of data corruption or + * precaution is taken to avoid the risks of data corruption or * segmentation faults. * * The most common cause of this failure is that an application was diff --git a/test/AtomicWriterReader.txt b/test/AtomicWriterReader.txt index dc0a3bd..064ba39 100644 --- a/test/AtomicWriterReader.txt +++ b/test/AtomicWriterReader.txt @@ -11,7 +11,7 @@ atomic_reader.c: is the "read" part of the test. Building the Tests ------------------ -The two test parts are automically built during configure and make process. +The two test parts are automatically built during configure and make process. But to build them individually, you can do in test/ directory: $ gcc atomic_writer $ gcc atomic_reader diff --git a/test/SWMR_POSIX_Order_UG.txt b/test/SWMR_POSIX_Order_UG.txt index 2771af1..a4190bc 100644 --- a/test/SWMR_POSIX_Order_UG.txt +++ b/test/SWMR_POSIX_Order_UG.txt @@ -3,7 +3,7 @@ POSIX Write Order Test Instructions Purpose ------- -This documents shows the requirments, implementaion design and instructions +This documents shows the requirements, implementation design and instructions of building and running the POSIX Write Order test. The name of the test is twriteorder and it resides in the test/ directory. @@ -44,7 +44,7 @@ and they are ready for the reader processes to access. If the system, in which the writer and reader processes run, the readers will always get all chain-linked blocks correctly. If the order of write -is not maintained, some reader processes may found unexpect block data. +is not maintained, some reader processes may found unexpected block data. Building the Tests ------------------ diff --git a/test/SWMR_UseCase_UG.txt b/test/SWMR_UseCase_UG.txt index 1e3d1e6..677fdc4 100644 --- a/test/SWMR_UseCase_UG.txt +++ b/test/SWMR_UseCase_UG.txt @@ -2,7 +2,7 @@ User Guide for SWMR Use Case Programs 2. Purpose: - This is a User Guide of the SWMR Use Case programs. It descibes the use + This is a User Guide of the SWMR Use Case programs. It describes the use case program and explain how to run them. 2.1. Author and Dates: @@ -87,7 +87,7 @@ It opens the HDF5 data file without the SWMR access mode (0 means off). This likely will result in error. This option is provided for - users to see the effect of the neede SWMR access mode for concurrent + users to see the effect of the needed SWMR access mode for concurrent access. 3.4. Test Shell Script: @@ -176,7 +176,7 @@ It opens the HDF5 data file without the SWMR access mode (0 means off). This likely will result in error. This option is provided for - users to see the effect of the neede SWMR access mode for concurrent + users to see the effect of the needed SWMR access mode for concurrent access. 4.4. Test Shell Script: diff --git a/test/accum.c b/test/accum.c index f098021..56080a3 100644 --- a/test/accum.c +++ b/test/accum.c @@ -1,4 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * + * Copyright by The HDF Group. * * Copyright by the Board of Trustees of the University of Illinois. * * All rights reserved. * * * @@ -1340,7 +1341,7 @@ test_accum_adjust(H5F_t *f) /* ==> Accumulator will try to resize, but see that it's getting too big */ /* ==> Size of new block is less than than half maximum size of accumulator */ /* ==> New block being appended to accumulator */ - /* ==> We can slide the dirty region down, to accomodate the request */ + /* ==> We can slide the dirty region down, to accommodate the request */ /* ==> Max Buffer Size - (dirty offset + adjust size) >= 2 * size) */ /* ==> Need to adjust location of accumulator while appending */ /* ==> Accumulator will need to be reallocated */ @@ -1386,7 +1387,7 @@ test_accum_adjust(H5F_t *f) /* ==> Accumulator will try to resize, but see that it's getting too big */ /* ==> Size of new block is less than than half maximum size of accumulator */ /* ==> New block being appended to accumulator */ - /* ==> We can slide the dirty region down, to accomodate the request */ + /* ==> We can slide the dirty region down, to accommodate the request */ /* ==> Max Buffer Size - (dirty offset + adjust size) < 2 * size) */ /* ==> Need to adjust location of accumulator while appending */ if (accum_write((1024 * 1024) - 5, 10, wbuf) < 0) diff --git a/test/accum_swmr_reader.c b/test/accum_swmr_reader.c index 5d78f1c..3dd80dd 100644 --- a/test/accum_swmr_reader.c +++ b/test/accum_swmr_reader.c @@ -1,4 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * + * Copyright by The HDF Group. * * Copyright by the Board of Trustees of the University of Illinois. * * All rights reserved. * * * diff --git a/test/atomic_writer.c b/test/atomic_writer.c index 5a6310a..33252f1 100644 --- a/test/atomic_writer.c +++ b/test/atomic_writer.c @@ -112,7 +112,7 @@ int main(int argc, char *argv[]) { int fd = -1; /* file descriptor */ - ssize_t bytes_wrote; /* the nubmer of bytes written */ + ssize_t bytes_wrote; /* the number of bytes written */ unsigned int *buf = NULL; /* buffer to hold written data */ unsigned int n, u, i; /* local index variable */ int temp; /* temporary variable */ @@ -43,7 +43,7 @@ * that disk space is allocated only when the contents are actually written. * E.g., If one creates a new file, seeks forward 10 million bytes, writes * 1 bytes and closes the file, then a sparse file, will show file size of - * 10 million bytes but actaully uses only couple disk blocks, much smaller + * 10 million bytes but actually uses only couple disk blocks, much smaller * than the formal file size.) * * One more consideration is that we want to distinguish an HDF5 library @@ -70,7 +70,7 @@ #endif /* Define Small, Large, Extra Large, Huge File which - * corrspond to less than 2GB, 2GB, 4GB, and tens of GB file size. + * correspond to less than 2GB, 2GB, 4GB, and tens of GB file size. * NO_FILE stands for "no file" to be tested. */ typedef enum fsizes_t { SFILE, LFILE, XLFILE, HUGEFILE, NO_FILE } fsizes_t; @@ -607,9 +607,9 @@ usage(void) "\t-fsize\tChange family size default to <fsize> where <fsize> is\n" "\t\ta positive float point number. Default value is %" PRIuHSIZE ".\n" "Examples:\n" - "\tbig -fsize 2.1e9 \t# test with file size just under 2GB\n" - "\tbig -fsize 2.2e9 \t# test with file size just above 2GB\n" - "\tBe sure the file system can support the file size requested\n", + "\t big -fsize 2.1e9 \t# test with file size just under 2GB\n" + "\t big -fsize 2.2e9 \t# test with file size just above 2GB\n" + "\t Be sure the file system can support the file size requested\n", (hsize_t)FAMILY_SIZE); } diff --git a/test/btree2.c b/test/btree2.c index ee43f5f..e04a5b4 100644 --- a/test/btree2.c +++ b/test/btree2.c @@ -696,7 +696,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_insert(bt2, &record) < 0) FAIL_STACK_ERROR - /* Attempt to find non-existant record in B-tree with 1 record */ + /* Attempt to find non-existent record in B-tree with 1 record */ /* (Should not be found, but not fail) */ idx = 41; if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE) @@ -707,7 +707,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_find(bt2, &idx, NULL, NULL) != FALSE) TEST_ERROR - /* Attempt to find existant record in B-tree with 1 record */ + /* Attempt to find existent record in B-tree with 1 record */ idx = 42; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) TEST_ERROR @@ -716,7 +716,7 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_find(bt2, &idx, NULL, NULL) != TRUE) TEST_ERROR - /* Attempt to index non-existant record in B-tree with 1 record */ + /* Attempt to index non-existent record in B-tree with 1 record */ idx = 0; H5E_BEGIN_TRY { @@ -764,18 +764,18 @@ test_insert_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_insert(bt2, &record) < 0) FAIL_STACK_ERROR - /* Attempt to find non-existant record in level-0 B-tree with several records */ + /* Attempt to find non-existent record in level-0 B-tree with several records */ /* (Should not be found, but not fail) */ idx = 41; if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE) TEST_ERROR - /* Attempt to find existant record in level-0 B-tree with several record */ + /* Attempt to find existent record in level-0 B-tree with several record */ idx = 56; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) TEST_ERROR - /* Attempt to index non-existant record in B-tree with several records */ + /* Attempt to index non-existent record in B-tree with several records */ idx = 0; H5E_BEGIN_TRY { @@ -935,23 +935,23 @@ test_insert_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p if (idx != (INSERT_SPLIT_ROOT_NREC + 2)) TEST_ERROR - /* Attempt to find non-existant record in level-1 B-tree */ + /* Attempt to find non-existent record in level-1 B-tree */ /* (Should not be found, but not fail) */ idx = INSERT_SPLIT_ROOT_NREC + 10; if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE) TEST_ERROR - /* Attempt to find existant record in root of level-1 B-tree */ + /* Attempt to find existent record in root of level-1 B-tree */ idx = 33; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) FAIL_STACK_ERROR - /* Attempt to find existant record in leaf of level-1 B-tree */ + /* Attempt to find existent record in leaf of level-1 B-tree */ idx = 56; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) FAIL_STACK_ERROR - /* Attempt to index non-existant record in level-1 B-tree */ + /* Attempt to index non-existent record in level-1 B-tree */ idx = 0; H5E_BEGIN_TRY { @@ -1671,13 +1671,13 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (idx != ((INSERT_SPLIT_ROOT_NREC * 29) + 5)) TEST_ERROR - /* Attempt to find non-existant record in level-2 B-tree */ + /* Attempt to find non-existent record in level-2 B-tree */ /* (Should not be found, but not fail) */ idx = INSERT_SPLIT_ROOT_NREC * 30; if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE) TEST_ERROR - /* Attempt to find existant record in root of level-2 B-tree */ + /* Attempt to find existent record in root of level-2 B-tree */ idx = 948; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) FAIL_STACK_ERROR @@ -1687,7 +1687,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)2) < 0) TEST_ERROR - /* Attempt to find existant record in internal node of level-2 B-tree */ + /* Attempt to find existent record in internal node of level-2 B-tree */ idx = 505; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) FAIL_STACK_ERROR @@ -1697,7 +1697,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)1) < 0) TEST_ERROR - /* Attempt to find existant record in leaf of level-2 B-tree */ + /* Attempt to find existent record in leaf of level-2 B-tree */ idx = 555; if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) FAIL_STACK_ERROR @@ -1707,7 +1707,7 @@ test_insert_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)0) < 0) TEST_ERROR - /* Attempt to index non-existant record in level-2 B-tree */ + /* Attempt to index non-existent record in level-2 B-tree */ idx = 0; H5E_BEGIN_TRY { @@ -3060,7 +3060,7 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time); if (idx != INSERT_MANY) TEST_ERROR - /* Attempt to find non-existant record in level-4 B-tree */ + /* Attempt to find non-existent record in level-4 B-tree */ /* (Should not be found, but not fail) */ idx = INSERT_MANY * 2; if (H5B2_find(bt2, &idx, find_cb, &idx) != FALSE) @@ -3071,12 +3071,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time); /* Pick random record */ idx = (hsize_t)(HDrandom() % INSERT_MANY); - /* Attempt to find existant record in root of level-4 B-tree */ + /* Attempt to find existent record in root of level-4 B-tree */ if (H5B2_find(bt2, &idx, find_cb, &idx) != TRUE) FAIL_STACK_ERROR } /* end for */ - /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */ + /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY * 3), find_cb, NULL); @@ -3099,12 +3099,12 @@ HDfprintf(stderr,"curr_time=%lu\n",(unsigned long)curr_time); /* Pick random record */ idx = (hsize_t)(HDrandom() % INSERT_MANY); - /* Attempt to find existant record in root of level-4 B-tree */ + /* Attempt to find existent record in root of level-4 B-tree */ /* (in increasing order) */ if (H5B2_index(bt2, H5_ITER_INC, idx, find_cb, &idx) < 0) FAIL_STACK_ERROR - /* Attempt to find existant record in root of level-4 B-tree */ + /* Attempt to find existent record in root of level-4 B-tree */ /* (in decreasing order) */ if (H5B2_index(bt2, H5_ITER_DEC, idx, find_dec_cb, &idx) < 0) FAIL_STACK_ERROR @@ -3210,7 +3210,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0) FAIL_STACK_ERROR - /* Attempt to find non-existant record in B-tree with 1 record */ + /* Attempt to find non-existent record in B-tree with 1 record */ /* (Should not be found, but not fail) */ find.key = 10; find.val = (hsize_t)-1; @@ -3226,7 +3226,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in B-tree with 1 record */ + /* Attempt to find existent record in B-tree with 1 record */ find.key = 42; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) FAIL_STACK_ERROR @@ -3240,7 +3240,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to index non-existant record in B-tree with 1 record */ + /* Attempt to index non-existent record in B-tree with 1 record */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL); @@ -3275,7 +3275,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_update(bt2, &modify, modify_rec_cb, &modify) < 0) FAIL_STACK_ERROR - /* Attempt to find non-existant record in B-tree with 1 record */ + /* Attempt to find non-existent record in B-tree with 1 record */ /* (Should not be found, but not fail) */ find.key = 10; find.val = (hsize_t)-1; @@ -3305,7 +3305,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to index non-existant record in B-tree with 1 record */ + /* Attempt to index non-existent record in B-tree with 1 record */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)1, index_rec_cb, NULL); @@ -3360,7 +3360,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_update(bt2, &record, no_modify_cb, NULL) < 0) FAIL_STACK_ERROR - /* Attempt to find non-existant record in level-0 B-tree with several records */ + /* Attempt to find non-existent record in level-0 B-tree with several records */ /* (Should not be found, but not fail) */ find.key = 10; find.val = (hsize_t)-1; @@ -3369,14 +3369,14 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in level-0 B-tree with several records */ + /* Attempt to find existent record in level-0 B-tree with several records */ find.key = 56; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) TEST_ERROR if (find.val != 12) TEST_ERROR - /* Attempt to index non-existant record in B-tree with several records */ + /* Attempt to index non-existent record in B-tree with several records */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL); @@ -3448,7 +3448,7 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (H5B2_update(bt2, &record, modify_rec_cb, &modify) < 0) FAIL_STACK_ERROR - /* Attempt to find non-existant record in level-0 B-tree with several records */ + /* Attempt to find non-existent record in level-0 B-tree with several records */ /* (Should not be found, but not fail) */ find.key = 41; find.val = (hsize_t)-1; @@ -3457,14 +3457,14 @@ test_update_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in level-0 B-tree with several record */ + /* Attempt to find existent record in level-0 B-tree with several record */ find.key = 56; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) TEST_ERROR if (find.val != 22) TEST_ERROR - /* Attempt to index non-existant record in B-tree with several records */ + /* Attempt to index non-existent record in B-tree with several records */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)4, index_rec_cb, NULL); @@ -3692,7 +3692,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p if (idx.key != (INSERT_SPLIT_ROOT_NREC_REC + 2)) TEST_ERROR - /* Attempt to find non-existant record in level-1 B-tree */ + /* Attempt to find non-existent record in level-1 B-tree */ /* (Should not be found, but not fail) */ find.key = 800; find.val = (hsize_t)-1; @@ -3701,7 +3701,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in root of level-1 B-tree */ + /* Attempt to find existent record in root of level-1 B-tree */ find.key = 33; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -3711,7 +3711,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p if (find.val != 67) TEST_ERROR - /* Attempt to find existant record in leaf of level-1 B-tree */ + /* Attempt to find existent record in leaf of level-1 B-tree */ find.key = 56; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -3721,7 +3721,7 @@ test_update_split_root(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_p if (find.val != 113) TEST_ERROR - /* Attempt to index non-existant record in level-1 B-tree */ + /* Attempt to index non-existent record in level-1 B-tree */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC + 2), index_rec_cb, NULL); @@ -4478,7 +4478,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4)) TEST_ERROR - /* Attempt to find non-existant record in level-2 B-tree */ + /* Attempt to find non-existent record in level-2 B-tree */ /* (Should not be found, but not fail) */ find.key = INSERT_SPLIT_ROOT_NREC_REC * 42; find.val = (hsize_t)-1; @@ -4487,7 +4487,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in root of level-2 B-tree */ + /* Attempt to find existent record in root of level-2 B-tree */ find.key = 1347; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4500,7 +4500,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)2) < 0) TEST_ERROR - /* Attempt to find existant record in internal node of level-2 B-tree */ + /* Attempt to find existent record in internal node of level-2 B-tree */ find.key = 513; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4513,7 +4513,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)1) < 0) TEST_ERROR - /* Attempt to find existant record in leaf of level-2 B-tree */ + /* Attempt to find existent record in leaf of level-2 B-tree */ find.key = 555; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4526,7 +4526,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)0) < 0) TEST_ERROR - /* Attempt to index non-existant record in level-2 B-tree */ + /* Attempt to index non-existent record in level-2 B-tree */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL); @@ -4643,7 +4643,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 4)) TEST_ERROR - /* Attempt to find non-existant record in level-2 B-tree */ + /* Attempt to find non-existent record in level-2 B-tree */ /* (Should not be found, but not fail) */ find.key = INSERT_SPLIT_ROOT_NREC_REC * 42; find.val = (hsize_t)-1; @@ -4652,7 +4652,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in root of level-2 B-tree */ + /* Attempt to find existent record in root of level-2 B-tree */ find.key = 1344; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4665,7 +4665,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)2) < 0) TEST_ERROR - /* Attempt to find existant record in internal node of level-2 B-tree */ + /* Attempt to find existent record in internal node of level-2 B-tree */ find.key = 512; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4678,7 +4678,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)1) < 0) TEST_ERROR - /* Attempt to find existant record in leaf of level-2 B-tree */ + /* Attempt to find existent record in leaf of level-2 B-tree */ find.key = 555; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4691,7 +4691,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)0) < 0) TEST_ERROR - /* Attempt to index non-existant record in level-2 B-tree */ + /* Attempt to index non-existent record in level-2 B-tree */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL); @@ -4818,7 +4818,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (idx.key != ((INSERT_SPLIT_ROOT_NREC_REC * 41) + 6)) TEST_ERROR - /* Attempt to find non-existant record in level-2 B-tree */ + /* Attempt to find non-existent record in level-2 B-tree */ /* (Should not be found, but not fail) */ find.key = INSERT_SPLIT_ROOT_NREC_REC * 42; find.val = (hsize_t)-1; @@ -4827,7 +4827,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (find.val != (hsize_t)-1) TEST_ERROR - /* Attempt to find existant record in root of level-2 B-tree */ + /* Attempt to find existent record in root of level-2 B-tree */ find.key = 1345; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4840,7 +4840,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)2) < 0) TEST_ERROR - /* Attempt to find existant record in internal node of level-2 B-tree */ + /* Attempt to find existent record in internal node of level-2 B-tree */ find.key = 513; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4853,7 +4853,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)1) < 0) TEST_ERROR - /* Attempt to find existant record in leaf of level-2 B-tree */ + /* Attempt to find existent record in leaf of level-2 B-tree */ find.key = 555; find.val = (hsize_t)-1; if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) @@ -4866,7 +4866,7 @@ test_update_make_level2(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_ if (check_node_depth(bt2, &record, (unsigned)0) < 0) TEST_ERROR - /* Attempt to index non-existant record in level-2 B-tree */ + /* Attempt to index non-existent record in level-2 B-tree */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_SPLIT_ROOT_NREC_REC * 42), index_rec_cb, NULL); @@ -5051,7 +5051,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time); if (iter.key != INSERT_MANY_REC) TEST_ERROR - /* Attempt to find non-existant record in level-4 B-tree */ + /* Attempt to find non-existent record in level-4 B-tree */ /* (Should not be found, but not fail) */ find.key = INSERT_MANY_REC * 2; find.val = (hsize_t)-1; @@ -5066,14 +5066,14 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time); find.key = (hsize_t)(HDrandom() % INSERT_MANY_REC); find.val = (hsize_t)-1; - /* Attempt to find existant record in level-4 B-tree */ + /* Attempt to find existent record in level-4 B-tree */ if (H5B2_find(bt2, &find, find_rec_cb, &find) != TRUE) FAIL_STACK_ERROR if (find.val != (find.key * 2)) TEST_ERROR } /* end for */ - /* Attempt to index non-existant record in level-4 B-tree, in increasing & decreasing order */ + /* Attempt to index non-existent record in level-4 B-tree, in increasing & decreasing order */ H5E_BEGIN_TRY { ret = H5B2_index(bt2, H5_ITER_INC, (hsize_t)(INSERT_MANY_REC * 3), find_rec_cb, NULL); @@ -5102,7 +5102,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time); find.key = (hsize_t)-1; find.val = (hsize_t)-1; - /* Attempt to find existant record in level-4 B-tree */ + /* Attempt to find existent record in level-4 B-tree */ /* (in increasing order) */ if (H5B2_index(bt2, H5_ITER_INC, idx, index_rec_cb, &find) < 0) FAIL_STACK_ERROR @@ -5115,7 +5115,7 @@ HDfprintf(stderr, "curr_time = %lu\n", (unsigned long)curr_time); find.key = (hsize_t)-1; find.val = (hsize_t)-1; - /* Attempt to find existant record in level-4 B-tree */ + /* Attempt to find existent record in level-4 B-tree */ /* (in decreasing order) */ if (H5B2_index(bt2, H5_ITER_DEC, idx, index_rec_cb, &find) < 0) FAIL_STACK_ERROR @@ -5234,7 +5234,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ PASSED(); - TESTING("B-tree remove: non-existant record from 1 record B-tree"); + TESTING("B-tree remove: non-existent record from 1 record B-tree"); /* Check for closing & re-opening the B-tree */ if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0) @@ -5261,7 +5261,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ if (!H5F_addr_defined(root_addr)) TEST_ERROR - /* Attempt to remove a non-existant record from a B-tree with 1 record */ + /* Attempt to remove a non-existent record from a B-tree with 1 record */ record = 0; H5E_BEGIN_TRY { @@ -5275,7 +5275,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ PASSED(); /* Attempt to remove a record from a B-tree with 1 record */ - TESTING("B-tree remove: existant record from 1 record B-tree"); + TESTING("B-tree remove: existent record from 1 record B-tree"); /* Check for closing & re-opening the B-tree */ if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0) @@ -5339,8 +5339,8 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ PASSED(); - /* Attempt to remove a non-existant record from a level-0 B-tree with mult. record */ - TESTING("B-tree remove: non-existant record from level-0 B-tree"); + /* Attempt to remove a non-existent record from a level-0 B-tree with mult. record */ + TESTING("B-tree remove: non-existent record from level-0 B-tree"); /* Check for closing & re-opening the B-tree */ if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0) @@ -5359,7 +5359,7 @@ test_remove_basic(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_ PASSED(); /* Attempt to remove a record from a level-0 B-tree with mult. record */ - TESTING("B-tree remove: mult. existant records from level-0 B-tree"); + TESTING("B-tree remove: mult. existent records from level-0 B-tree"); /* Check for closing & re-opening the B-tree */ if (reopen_btree(f, &bt2, bt2_addr, tparam) < 0) @@ -5517,7 +5517,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt herr_t ret; /* Generic error return value */ /* B-tree record removal tests */ - TESTING("B-tree remove: non-existant record from level-1 B-tree"); + TESTING("B-tree remove: non-existent record from level-1 B-tree"); /* Create the file for the test */ if (create_file(&file, &f, fapl) < 0) @@ -5550,7 +5550,7 @@ test_remove_level1_noredistrib(hid_t fapl, const H5B2_create_t *cparam, const bt if (!H5F_addr_defined(root_addr)) TEST_ERROR - /* Attempt to remove a non-existant record from a B-tree with 1 record */ + /* Attempt to remove a non-existent record from a B-tree with 1 record */ record = (INSERT_SPLIT_ROOT_NREC * 2) + 1; H5E_BEGIN_TRY { @@ -9605,7 +9605,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa /* * Test modifying records */ - TESTING("B-tree modify: attempt to modify non-existant record"); + TESTING("B-tree modify: attempt to modify non-existent record"); /* Create the file for the test */ if (create_file(&file, &f, fapl) < 0) @@ -9628,7 +9628,7 @@ test_modify(hid_t fapl, const H5B2_create_t *cparam, const bt2_test_param_t *tpa if (bt2_stat.depth != 2) TEST_ERROR - /* Attempt to modify a non-existant record */ + /* Attempt to modify a non-existent record */ record = 3; modify = 4; H5E_BEGIN_TRY diff --git a/test/cache.c b/test/cache.c index 8ab5838..41be0d4 100644 --- a/test/cache.c +++ b/test/cache.c @@ -2789,8 +2789,8 @@ write_permitted_check(int * Test the behaviour with different flags. * * This test was added primarily to test basic insert - * pinned entry functionallity, but I through in explicit - * tests for other functionallity that is tested implicitly + * pinned entry functionality, but I through in explicit + * tests for other functionality that is tested implicitly * elsewhere. * * Return: void @@ -2908,7 +2908,7 @@ check_insert_entry(unsigned paged) } } - /* Thats all we can get from H5C_get_entry_status(). + /* That's all we can get from H5C_get_entry_status(). * Now start looking at the cache data structures directly. */ @@ -5186,7 +5186,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr) * * 10/8/07 -- JRM * Added a resize operation to this test to satisfy the new - * requiremnt that any resize of an entry on flush will always + * requirement that any resize of an entry on flush will always * be accompanied by a resize. Note that as a result, this * test becomes redundant with later tests. */ @@ -5283,7 +5283,7 @@ check_flush_cache__flush_ops(H5F_t *file_ptr) * * 10/8/07 -- JRM * Added a resize operation to this test to satisfy the new - * requiremnt that any resize of an entry on flush will always + * requirement that any resize of an entry on flush will always * be accompanied by a resize. Note that as a result, this * test becomes redundant with later tests. */ @@ -8892,7 +8892,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr) * * (VET, 9) Y 10 KB N N - dirty (VET, 8) * - * Recall that in this test bed, flush operations are excuted the + * Recall that in this test bed, flush operations are executed the * first time the associated entry is flushed, and are then * deleted. */ @@ -9445,7 +9445,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr) * returns to the serialize function for (VET, 9). * * When the serialize function for (VET, 9) is done with (VET, 8), it - * calls H5C_unprotect(VET, 8), which markes (VET, 8) as dirty and + * calls H5C_unprotect(VET, 8), which marks (VET, 8) as dirty and * unprotected, and places it at the head of the LRU. * * (VET, 0) is the next item on the LRU -- it is clean and is therefore @@ -9846,7 +9846,7 @@ check_flush_cache__flush_op_eviction_test(H5F_t *file_ptr) * Added tests for pinned entries. * * JRM -- 5/17/06 - * Complete reqrite of pinned entry tests to accomodate + * Complete rewrite of pinned entry tests to accommodate * the new H5C_mark_entry_dirty() call. * *------------------------------------------------------------------------- @@ -11842,7 +11842,7 @@ check_expunge_entry(unsigned paged) * Function: check_multiple_read_protect() * * Purpose: Verify that multiple, simultaneous read protects of a - * single entry perform as expectd. + * single entry perform as expected. * * Return: void * @@ -12188,7 +12188,7 @@ check_multiple_read_protect(unsigned paged) if (pass && (entry_ptr->header.is_pinned)) { pass = FALSE; - failure_mssg = "enty (0,4) still pinned.\n"; + failure_mssg = "entry (0,4) still pinned.\n"; } unprotect_entry(file_ptr, 0, 4, /* (0,4) unprotect */ @@ -14391,7 +14391,7 @@ check_destroy_protected_err(unsigned paged) * Function: check_duplicate_insert_err() * * Purpose: Verify that an attempt to insert and entry that is - * alread in the cache will generate an error. + * already in the cache will generate an error. * * Return: void * @@ -16947,7 +16947,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - /* eigth epoch -- force the hit rate to 100% again -- should be steady + /* eight epoch -- force the hit rate to 100% again -- should be steady * state. */ if (pass) { @@ -17043,7 +17043,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - if (pass) { /* twelth epoch -- at minimum size so no more ageouts */ + if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */ rpt_fcn_called = FALSE; i = 0; while (pass && (i < 1000)) { @@ -17324,7 +17324,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - /* eigth epoch -- still 100% hit rate + /* eighth epoch -- still 100% hit rate */ if (pass) { @@ -17427,7 +17427,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - /* twelth epoch -- force the hit rate to 100% again -- should be steady + /* twelfth epoch -- force the hit rate to 100% again -- should be steady * state. */ if (pass) { @@ -17803,7 +17803,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - /* eigth epoch -- force the hit rate to 100% again -- should be steady + /* eighth epoch -- force the hit rate to 100% again -- should be steady * state. */ if (pass) { @@ -17903,7 +17903,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - if (pass) { /* twelth epoch -- at minimum size so no more ageouts */ + if (pass) { /* twelfth epoch -- at minimum size so no more ageouts */ rpt_fcn_called = FALSE; i = 0; @@ -18554,7 +18554,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - /* twelth epoch -- hit rate 1.0 -- decrease as before. + /* twelfth epoch -- hit rate 1.0 -- decrease as before. */ if (pass) { @@ -19567,7 +19567,7 @@ check_auto_cache_resize(hbool_t cork_ageout, unsigned paged) if (show_progress) HDfprintf(stderr, "check point %d\n", checkpoint++); - /* protect and unprotect two more large entries -- shouldnt trigger a + /* protect and unprotect two more large entries -- shouldn't trigger a * flash cache size increase. */ if (pass) { @@ -22127,7 +22127,7 @@ check_auto_cache_resize_disable(unsigned paged) * mode (add space), which is triggered whenever the size of a newly * loaded or inserted entry, or the delta between old and new entry * sizes exceeds some fraction of the current maximum cache size, and - * the cache doesn't have enough free space to accomodate the new/ + * the cache doesn't have enough free space to accommodate the new/ * resize entry without performing evictions. The range of permissible * values for the flash_threshold (0.1 to 1.0 as of this writing), and * for the flash_multiple (0.1 to 10.0) do not permit the facility to @@ -22136,7 +22136,7 @@ check_auto_cache_resize_disable(unsigned paged) * current max_cache_size equal to max_size. * * We have already tested the latter in check_auto_cache_resize(), so - * we need only thest the former here. Do this by disabling flash + * we need only test the former here. Do this by disabling flash * cache size increments via the flash_incr_mode, and then creating * situations that would trigger flash cache size increases were that * code enabled. @@ -25209,7 +25209,7 @@ check_auto_cache_resize_input_errs(unsigned paged) /*------------------------------------------------------------------------- * Function: check_auto_cache_resize_aux_fcns() * - * Purpose: Verify that the auxilary functions associated with + * Purpose: Verify that the auxiliary functions associated with * the automatic cache resize capability are operating * correctly. These functions are: * @@ -25286,13 +25286,13 @@ check_auto_cache_resize_aux_fcns(unsigned paged) }; if (paged) - TESTING("automatic cache resize auxilary functions (paged aggregation)") + TESTING("automatic cache resize auxiliary functions (paged aggregation)") else - TESTING("automatic cache resize auxilary functions") + TESTING("automatic cache resize auxiliary functions") pass = TRUE; - /* allocate a cache, and then test the various auxilary functions. + /* allocate a cache, and then test the various auxiliary functions. */ if (pass) { @@ -25814,7 +25814,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged) if (file_ptr == NULL) { pass = FALSE; - failure_mssg = "bad return from cache intialization.\n"; + failure_mssg = "bad return from cache initialization.\n"; } else cache_ptr = file_ptr->shared->cache; @@ -26033,7 +26033,7 @@ check_metadata_blizzard_absence(hbool_t fill_via_insertion, unsigned paged) H5C__DIRTIED_FLAG); /* unsigned int flags */ } - /* Expected staus is that after each insertion, the entry + /* Expected status is that after each insertion, the entry * inserted 26 insertions ago has been flushed, and the * entry currently getting inserted is now in the cache and * dirty. @@ -28440,7 +28440,7 @@ check_flush_deps_err(unsigned paged) result = H5C_destroy_flush_dependency(&((entries[entry_type])[0]), &((entries[entry_type])[1])); if (result != FAIL) - CACHE_ERROR("Destroying [non-existant] dependency when parent isn't protected") + CACHE_ERROR("Destroying [non-existent] dependency when parent isn't protected") break; /* Verify that parent entry has flush dependency */ @@ -31671,7 +31671,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged) if (file_ptr == NULL) { pass = FALSE; - failure_mssg = "bad return from cache intialization.\n"; + failure_mssg = "bad return from cache initialization.\n"; } else cache_ptr = file_ptr->shared->cache; @@ -32183,7 +32183,7 @@ check_metadata_cork(hbool_t fill_via_insertion, unsigned paged) * an entry that is no longer in the cache, and which may have * been deleted. * - * This function contans tests for correct handling on this + * This function contains tests for correct handling on this * situation. * * Do nothing if pass is FALSE on entry. @@ -32440,7 +32440,7 @@ cedds__expunge_dirty_entry_in_flush_test(H5F_t *file_ptr) * * (HET, 3) Y 16 KB Y N - - * - * Recall that in this test bed, flush operations are excuted the + * Recall that in this test bed, flush operations are executed the * first time the associated entry is flushed, and are then * deleted. */ @@ -32694,7 +32694,7 @@ cedds__H5C_make_space_in_cache(H5F_t *file_ptr) * * (HET, 3) Y 16 KB N N - - * - * Recall that in this test bed, flush operations are excuted the + * Recall that in this test bed, flush operations are executed the * first time the associated entry is flushed, and are then * deleted. */ @@ -33112,7 +33112,7 @@ cedds__H5C__autoadjust__ageout__evict_aged_out_entries(H5F_t *file_ptr) * * (MET, 1-31) Y 64 KB N N - - * - * Recall that in this test bed, flush operations are excuted the + * Recall that in this test bed, flush operations are executed the * first time the associated entry is flushed, and are then * deleted. */ @@ -33479,7 +33479,7 @@ cedds__H5C_flush_invalidate_cache__bucket_scan(H5F_t *file_ptr) if (pass) { - /* load the entry that will have a flush dependencey with (MET, 0), + /* load the entry that will have a flush dependency with (MET, 0), * thus preventing it from being flushed on the first pass through * the skip list. */ diff --git a/test/cache_common.c b/test/cache_common.c index 3be633d..f1fc73b 100644 --- a/test/cache_common.c +++ b/test/cache_common.c @@ -923,7 +923,7 @@ notify_image_len(const void *thing, size_t *image_length) /*------------------------------------------------------------------------- * Function: pre_serialize & friends * - * Purpose: Pre_serialize the supplied entry. For now this consistes of + * Purpose: Pre_serialize the supplied entry. For now this consists of * executing any flush operations and loading the appropriate * values into *new_addr_ptr, *new_len_ptr, and *flags_ptr. * @@ -1096,7 +1096,7 @@ notify_pre_serialize(H5F_t *f, void *thing, haddr_t addr, size_t len, haddr_t *n /*------------------------------------------------------------------------- * Function: serialize & friends * - * Purpose: Serialize the supplied entry. For now this consistes of + * Purpose: Serialize the supplied entry. For now this consists of * loading the type and index of the entry into the first * three bytes of the image (if it is long enough -- if not * just load the low order byte of the index into the first @@ -1538,7 +1538,7 @@ add_flush_op(int target_type, int target_idx, int op_code, int type, int idx, hb * Otherwise, set up a pinned entry dependency so we can * test the pinned entry modifications to the flush routine. * - * Given the types and indicies of the pinned and pinning + * Given the types and indices of the pinned and pinning * entries, add the pinned entry to the list of pinned * entries in the pinning entry, increment the * pinning reference count of the pinned entry, and @@ -1777,7 +1777,7 @@ execute_flush_op(H5F_t *file_ptr, struct test_entry_t *entry_ptr, struct flush_o case FLUSH_OP__EXPUNGE: /* the expunge flush op exists to allow us to simulate the - * case in which an entry is removed from the cashe as the + * case in which an entry is removed from the cache as the * the result of the flush of a second entry. At present, * this can only happen via the take ownership flag, but * we will make this test feature more general to as to make @@ -2322,7 +2322,7 @@ verify_clean(void) * Function: verify_entry_status * * Purpose: Verify that a list of entries have the expected status. - * If any discrepencies are found, set the failure message + * If any discrepancies are found, set the failure message * and set pass to FALSE. * * Do nothing if pass is FALSE on entry. @@ -3833,8 +3833,8 @@ row_major_scan_forward(H5F_t *file_ptr, int32_t max_index, int32_t lag, hbool_t HDassert((!cache_ptr->slist_enabled) || (cache_ptr->slist_size == cache_ptr->dirty_index_size)); - } /* end if */ - } /* end elsef */ + } + } if (verbose) HDfprintf(stdout, "\n"); @@ -5242,7 +5242,7 @@ resize_configs_are_equal(const H5C_auto_size_ctl_t *a, const H5C_auto_size_ctl_t * has both internal and external configuration matching * *config_ptr. * - * Do nothin on success. On failure, set pass to FALSE, and + * Do nothing on success. On failure, set pass to FALSE, and * load an error message into failue_mssg. Note that * failure_msg is assumed to be at least 128 bytes in length. * diff --git a/test/cache_common.h b/test/cache_common.h index 901f5c2..91aedc7 100644 --- a/test/cache_common.h +++ b/test/cache_common.h @@ -251,7 +251,7 @@ typedef struct flush_op { typedef enum test_entry_action_t { TEST_ENTRY_ACTION_NUL = 0, /* No action on entry */ - TEST_ENTRY_ACTION_MOVE /* Entry is beging moved */ + TEST_ENTRY_ACTION_MOVE /* Entry is being moved */ } test_entry_action_t; typedef struct test_entry_t { @@ -370,7 +370,7 @@ typedef struct test_entry_t { * been resized by a resize flush * op, and the flush function has * not yet returned, This field is - * used to turn off overactive santity + * used to turn off overactive sanity * checking code that would otherwise * cause a false test failure. */ diff --git a/test/cache_image.c b/test/cache_image.c index f2f7b51..4e8644f 100644 --- a/test/cache_image.c +++ b/test/cache_image.c @@ -136,7 +136,7 @@ create_datasets(hid_t file_id, int min_dset, int max_dset) } /* set the dataset creation plist to specify that the raw data is - * to be partioned into 10X10 element chunks. + * to be partitioned into 10X10 element chunks. */ if (pass) { @@ -470,7 +470,7 @@ delete_datasets(hid_t file_id, int min_dset, int max_dset) * Set pass to FALSE and issue a suitable failure * message if either the file contains a metadata cache image * superblock extension and mdci_sbem_expected is TRUE, or - * vise versa. + * vice versa. * * If mdci_sbem_expected is TRUE, also verify that the metadata * cache has been advised of this. @@ -544,7 +544,7 @@ open_hdf5_file(hbool_t create_file, hbool_t mdci_sbem_expected, hbool_t read_onl if (show_progress) HDfprintf(stdout, "%s: cp = %d.\n", fcn_name, cp++); - /* create a file access propertly list. */ + /* create a file access property list. */ if (pass) { fapl_id = H5Pcreate(H5P_FILE_ACCESS); @@ -879,7 +879,7 @@ attempt_swmr_open_hdf5_file(const hbool_t create_file, const hbool_t set_mdci_fa H5AC_cache_image_config_t cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE, H5AC__CACHE_IMAGE__ENTRY_AGEOUT__NONE}; - /* create a file access propertly list. */ + /* create a file access property list. */ if (pass) { fapl_id = H5Pcreate(H5P_FILE_ACCESS); @@ -1498,7 +1498,7 @@ check_cache_image_ctl_flow_1(void) * processed as part of the first protect operation after the * superblock is loaded.) * - * In this particular test, we preform the following operations: + * In this particular test, we perform the following operations: * * 1) Create a HDF5 file with the cache image FAPL entry. * @@ -6885,7 +6885,7 @@ cache_image_api_error_check_4(void) * image. * * The objective of this test is to create a test file - * with both non-empty self referential presistant + * with both non-empty self referential persistent * free space managers, and a cache image, and then * verify that this situation is handled correctly if * H5Fget_free_sections() is called before the metadata diff --git a/test/cache_tagging.c b/test/cache_tagging.c index 8761ed5..583c665 100644 --- a/test/cache_tagging.c +++ b/test/cache_tagging.c @@ -441,7 +441,7 @@ check_file_creation_tags(hid_t fcpl_id, int type) /* Variable Declarations */ hid_t fid = -1; /* File Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose test outout */ + int verbose = FALSE; /* verbose test output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag = 0; haddr_t sbe_tag = 0; @@ -537,7 +537,7 @@ check_file_open_tags(hid_t fcpl, int type) /* Variable Declarations */ hid_t fid = -1; /* File Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag; /* Root Group Tag */ haddr_t sbe_tag; /* Sblock Extension Tag */ @@ -659,7 +659,7 @@ check_group_creation_tags(void) hid_t fid = -1; /* File Identifier */ hid_t gid = -1; /* Group Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag = HADDR_UNDEF; /* Root Group Tag */ haddr_t g_tag; /* Group Tag */ @@ -774,7 +774,7 @@ check_multi_group_creation_tags(void) hid_t fid = -1; /* File Identifier */ hid_t gid = -1; /* Group Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ char gname[16]; /* group name buffer */ int i = 0; /* iterator */ @@ -924,7 +924,7 @@ check_link_iteration_tags(void) hid_t sid = -1; /* Group Identifier */ hid_t did = -1; /* Group Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ int i = 0; /* iterator */ haddr_t root_tag = 0; /* Root Group Tag Value */ @@ -1058,7 +1058,7 @@ check_dense_attribute_tags(void) hid_t did = -1; /* Group Identifier */ hid_t dcpl = -1; /* Group Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ int i = 0; /* iterator */ hid_t fapl = -1; /* File access property list */ @@ -1408,7 +1408,7 @@ check_attribute_creation_tags(hid_t fcpl, int type) hid_t gid = -1; /* Group Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag = 0; /* Root group tag */ haddr_t g_tag = 0; @@ -1566,7 +1566,7 @@ check_attribute_open_tags(hid_t fcpl, int type) hid_t gid = -1; /* Group Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag = 0; haddr_t g_tag = 0; @@ -1725,7 +1725,7 @@ check_attribute_rename_tags(hid_t fcpl, int type) hid_t aid = -1; /* Attribute Identifier */ hid_t sid = -1; /* Dataset Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ int *data = NULL; /* data buffer */ int i, j, k = 0; /* iterators */ @@ -1941,7 +1941,7 @@ check_attribute_delete_tags(hid_t fcpl, int type) hid_t aid = -1; /* Attribute Identifier */ hid_t sid = -1; /* Dataset Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ int *data = NULL; /* data buffer */ int i, j, k = 0; /* iterators */ @@ -2121,7 +2121,7 @@ check_dataset_creation_tags(hid_t fcpl, int type) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -2273,7 +2273,7 @@ check_dataset_creation_earlyalloc_tags(hid_t fcpl, int type) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -2431,7 +2431,7 @@ check_dataset_open_tags(void) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -2575,7 +2575,7 @@ check_dataset_write_tags(void) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -2734,7 +2734,7 @@ check_attribute_write_tags(hid_t fcpl, int type) hid_t aid = -1; /* Attribute Identifier */ hid_t sid = -1; /* Dataset Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ int *data = NULL; /* data buffer */ int i, j, k = 0; /* iterators */ @@ -2912,7 +2912,7 @@ check_dataset_read_tags(void) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -3066,7 +3066,7 @@ check_dataset_size_retrieval(void) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -3222,7 +3222,7 @@ check_dataset_extend_tags(void) hid_t did = -1; /* Dataset Identifier */ hid_t sid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -3642,7 +3642,7 @@ check_link_removal_tags(hid_t fcpl, int type) hid_t sid = -1; /* Dataspace Identifier */ hid_t gid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -3823,7 +3823,7 @@ check_link_getname_tags(void) hid_t sid = -1; /* Dataspace Identifier */ hid_t gid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ hid_t dcpl = -1; /* dataset creation pl */ hsize_t cdims[2] = {1, 1}; /* chunk dimensions */ @@ -3992,7 +3992,7 @@ check_external_link_creation_tags(void) hid_t fid2 = -1; /* File Identifier */ hid_t gid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag = 0; @@ -4111,7 +4111,7 @@ check_external_link_open_tags(void) hid_t gid = -1; /* Dataspace Identifier */ hid_t xid = -1; /* Dataspace Identifier */ #ifndef NDEBUG - int verbose = FALSE; /* verbose file outout */ + int verbose = FALSE; /* verbose file output */ #endif /* NDEBUG */ /* end debugging functions */ haddr_t root_tag = 0; haddr_t root2_tag = 0; @@ -4378,7 +4378,7 @@ main(void) /* Create a standard file creation property list */ fcpl_default = H5Pcreate(H5P_FILE_CREATE); - /* Create an fcpl with shared messages and file space managment enabled */ + /* Create an fcpl with shared messages and file space management enabled */ fcpl_shmesg_all = H5Pcreate(H5P_FILE_CREATE); H5Pset_shared_mesg_nindexes(fcpl_shmesg_all, 1); H5Pset_shared_mesg_index(fcpl_shmesg_all, 0, H5O_SHMESG_ALL_FLAG, 20); diff --git a/test/chunk_info.c b/test/chunk_info.c index a0fdac5..10661c6 100644 --- a/test/chunk_info.c +++ b/test/chunk_info.c @@ -138,7 +138,7 @@ reinit_vars(unsigned *read_flt_msk, haddr_t *addr, hsize_t *size) * Return: Success: SUCCEED * Failure: FAIL * - * Note: Note that the dataspace argument in these new functions are + * Note: Note that the dataspace argument in these new functions is * currently not used. The functionality involved the dataspace * will be implemented in the next version. * @@ -648,7 +648,7 @@ test_chunk_info_single_chunk(char *filename, hid_t fapl) if (size != SINGLE_CHUNK_SIZE) TEST_ERROR - /* Release resourse */ + /* Release resource */ if (H5Dclose(dset) < 0) TEST_ERROR if (H5Sclose(dspace) < 0) @@ -906,7 +906,7 @@ test_chunk_info_implicit(char *filename, hid_t fapl) if (read_each_chunk(dset, i * CHUNK_NX, j * CHUNK_NY, (void *)direct_buf[n]) < 0) TEST_ERROR - /* Release resourse */ + /* Release resource */ if (H5Dclose(dset) < 0) TEST_ERROR if (H5Sclose(dspace) < 0) @@ -1153,7 +1153,7 @@ test_chunk_info_fixed_array(char *filename, hid_t fapl) if (read_each_chunk(dset, i * CHUNK_NX, j * CHUNK_NY, (void *)direct_buf[n]) < 0) TEST_ERROR - /* Release resourse */ + /* Release resource */ if (H5Dclose(dset) < 0) TEST_ERROR if (H5Sclose(dspace) < 0) @@ -1397,7 +1397,7 @@ test_chunk_info_extensible_array(char *filename, hid_t fapl) if (read_each_chunk(dset, i * CHUNK_NX, j * CHUNK_NY, (void *)direct_buf[n]) < 0) TEST_ERROR - /* Release resourse */ + /* Release resource */ if (H5Dclose(dset) < 0) TEST_ERROR if (H5Sclose(dspace) < 0) @@ -1641,7 +1641,7 @@ test_chunk_info_version2_btrees(char *filename, hid_t fapl) if (read_each_chunk(dset, i * CHUNK_NX, j * CHUNK_NY, (void *)direct_buf[n]) < 0) TEST_ERROR - /* Release resourse */ + /* Release resource */ if (H5Dclose(dset) < 0) TEST_ERROR if (H5Sclose(dspace) < 0) diff --git a/test/cmpd_dset.c b/test/cmpd_dset.c index d918bcc..f156655 100644 --- a/test/cmpd_dset.c +++ b/test/cmpd_dset.c @@ -713,7 +713,7 @@ test_compound(char *filename, hid_t fapl) /* *###################################################################### * Step 12: Write an array into the middle third of the dataset - * initializeing only members `b' and `d' to -1. + * initializing only members `b' and `d' to -1. */ TESTING("hyperslab part initialized write"); diff --git a/test/cork.c b/test/cork.c index 94f3bb0..6185e12 100644 --- a/test/cork.c +++ b/test/cork.c @@ -97,7 +97,7 @@ verify_old_dset_cork(void) hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */ hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */ hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */ - int **buf = NULL; /* Data bufer (pointers to fake 2D array) */ + int **buf = NULL; /* Data buffer (pointers to fake 2D array) */ int *buf_data = NULL; /* Data buffer (actual data) */ int i = 0, j = 0; /* Local index variables */ H5O_info_t oinfo, oinfo2, oinfo3; /* Object metadata information */ @@ -509,7 +509,7 @@ verify_dset_cork(hbool_t swmr, hbool_t new_format) hsize_t dims[2] = {100, 20}; /* Dataset dimension sizes */ hsize_t max_dims[2] = {100, H5S_UNLIMITED}; /* Dataset maximum dimension sizes */ hsize_t chunk_dims[2] = {2, 5}; /* Dataset chunked dimension sizes */ - int **buf = NULL; /* Data bufer (pointers to fake 2D array) */ + int **buf = NULL; /* Data buffer (pointers to fake 2D array) */ int *buf_data = NULL; /* Data buffer (actual data) */ int i = 0, j = 0; /* Local index variables */ H5O_info_t oinfo, oinfo2, oinfo3; /* Object metadata information */ @@ -1883,7 +1883,7 @@ test_dset_cork(hbool_t swmr, hbool_t new_format) { hid_t fid = H5I_INVALID_HID; /* File ID */ hid_t fapl = H5I_INVALID_HID; /* File access property list */ - hid_t gid = H5I_INVALID_HID; /* Groupd ID */ + hid_t gid = H5I_INVALID_HID; /* Group ID */ hid_t did1 = H5I_INVALID_HID, did2 = H5I_INVALID_HID; /* Dataset IDs */ hid_t tid1 = H5I_INVALID_HID, tid2 = H5I_INVALID_HID; /* Datatype IDs */ hid_t sid = H5I_INVALID_HID; /* Dataspace ID */ diff --git a/test/direct_chunk.c b/test/direct_chunk.c index 1c4d6d7..3d233f6 100644 --- a/test/direct_chunk.c +++ b/test/direct_chunk.c @@ -686,7 +686,7 @@ filter_bogus1(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts, size_t buf_left = *buf_size; /* Amount of data buffer left to process */ if (flags & H5Z_FLAG_REVERSE) { /* read */ - /* Substract the "add on" value to all the data values */ + /* Subtract the "add on" value to all the data values */ while (buf_left > 0) { *int_ptr++ -= (int)ADD_ON; buf_left -= sizeof(int); @@ -722,7 +722,7 @@ filter_bogus2(unsigned int flags, size_t H5_ATTR_UNUSED cd_nelmts, size_t buf_left = *buf_size; /* Amount of data buffer left to process */ if (flags & H5Z_FLAG_REVERSE) { /* read */ - /* Substract the "add on" value to all the data values */ + /* Subtract the "add on" value to all the data values */ while (buf_left > 0) { *int_ptr++ /= (int)FACTOR; buf_left -= sizeof(int); @@ -1427,7 +1427,7 @@ error: * Function: test_direct_chunk_read_no_cache * * Purpose: Test the basic functionality of H5Dread_chunk with the - * chunk cache diabled. + * chunk cache disabled. * * Return: Success: 0 * Failure: 1 @@ -1963,7 +1963,7 @@ error: * Function: test_read_unallocated_chunk * * Purpose: Tests the H5Dread_chunk and H5Dget_chunk_storage_size with valid - * offets to chunks that have not been written to the dataset and are + * offsets to chunks that have not been written to the dataset and are * not allocated in the chunk storage on disk. * * Return: Success: 0 @@ -2014,7 +2014,7 @@ test_read_unallocated_chunk(hid_t file) if ((dxpl = H5Pcreate(H5P_DATASET_XFER)) < 0) FAIL_STACK_ERROR; - /* Write a single chunk to intialize the chunk storage */ + /* Write a single chunk to initialize the chunk storage */ HDmemset(direct_buf, 0, CHUNK_NX * CHUNK_NY * sizeof(int)); offset[0] = 0; offset[1] = 0; @@ -2031,7 +2031,7 @@ test_read_unallocated_chunk(hid_t file) offset[0] = i * CHUNK_NX; offset[1] = j * CHUNK_NY; - /* Read a non-existant chunk using the direct read function. */ + /* Read a non-existent chunk using the direct read function. */ H5E_BEGIN_TRY { status = H5Dread_chunk(dataset, dxpl, offset, &filter_mask, &direct_buf); @@ -2042,7 +2042,7 @@ test_read_unallocated_chunk(hid_t file) if (status != -1) TEST_ERROR - /* Query the size of the non-existant chunk */ + /* Query the size of the non-existent chunk */ direct_chunk_nbytes = ULONG_MAX; H5E_BEGIN_TRY { diff --git a/test/dsets.c b/test/dsets.c index 1569f9a..330c7a0 100644 --- a/test/dsets.c +++ b/test/dsets.c @@ -1821,7 +1821,7 @@ filter_corrupt(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_valu TEST_ERROR HDmemset(data, (int)value, (size_t)length); - if (flags & H5Z_FLAG_REVERSE) { /* Varify data is actually corrupted during read */ + if (flags & H5Z_FLAG_REVERSE) { /* Verify data is actually corrupted during read */ dst += offset; if (HDmemcmp(data, dst, (size_t)length) != 0) TEST_ERROR @@ -2005,7 +2005,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32, if (status >= 0) TEST_ERROR; - /* Callback decides to continue inspite data is corrupted. */ + /* Callback decides to continue in spite data is corrupted. */ if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0) TEST_ERROR; if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0) @@ -2072,7 +2072,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32, if (status >= 0) TEST_ERROR; - /* Callback decides to continue inspite data is corrupted. */ + /* Callback decides to continue in spite data is corrupted. */ if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0) TEST_ERROR; if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0) @@ -2136,7 +2136,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32, if (status >= 0) TEST_ERROR; - /* Callback decides to continue inspite data is corrupted. */ + /* Callback decides to continue in spite data is corrupted. */ if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0) TEST_ERROR; if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0) @@ -2203,7 +2203,7 @@ test_filter_internal(hid_t fid, const char *name, hid_t dcpl, int if_fletcher32, if (status >= 0) TEST_ERROR; - /* Callback decides to continue inspite data is corrupted. */ + /* Callback decides to continue in spite data is corrupted. */ if (H5Pset_filter_callback(dxpl, filter_cb_cont, NULL) < 0) TEST_ERROR; if (H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, check_data) < 0) @@ -3234,7 +3234,7 @@ test_nbit_int(hid_t file) power = HDpow(2.0, (double)(precision - 1)); orig_data[i][j] = (int)(((long long)HDrandom() % (long long)power) << offset); - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((i * size[1] + j + 1) % 2 == 0) orig_data[i][j] = -orig_data[i][j]; } @@ -3812,7 +3812,7 @@ test_nbit_compound(hid_t file) orig_data[i][j].s = (short)(((long long)HDrandom() % (long long)power) << offset[2]); orig_data[i][j].f = float_val[i][j]; - /* some even-numbered integer values are negtive */ + /* some even-numbered integer values are negative */ if ((i * size[1] + j + 1) % 2 == 0) { orig_data[i][j].i = -orig_data[i][j].i; orig_data[i][j].s = (short)-orig_data[i][j].s; @@ -4526,7 +4526,7 @@ test_nbit_int_size(hid_t file) goto error; } - /* Initiliaze data buffer with random data within correct range + /* Initialize data buffer with random data within correct range * corresponding to the memory datatype's precision and offset. */ for (i = 0; i < DSET_DIM1; i++) @@ -4739,7 +4739,7 @@ test_nbit_flt_size(hid_t file) } /* end if */ /* - * Initiliaze data buffer with random data + * Initialize data buffer with random data */ for (i = 0; i < DSET_DIM1; i++) for (j = 0; j < DSET_DIM2; j++) @@ -4898,7 +4898,7 @@ test_scaleoffset_int(hid_t file) for (j = 0; j < (size_t)size[1]; j++) { orig_data[i][j] = HDrandom() % 10000; - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((i * size[1] + j + 1) % 2 == 0) orig_data[i][j] = -orig_data[i][j]; } @@ -5042,7 +5042,7 @@ test_scaleoffset_int_2(hid_t file) for (j = 0; j < (size_t)size[1]; j++) { orig_data[0][j] = (int)HDrandom() % 10000; - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((j + 1) % 2 == 0) orig_data[0][j] = -orig_data[0][j]; } @@ -5164,7 +5164,7 @@ test_scaleoffset_float(hid_t file) for (j = 0; j < (size_t)size[1]; j++) { orig_data[i][j] = (float)(HDrandom() % 100000) / 1000.0F; - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((i * size[1] + j + 1) % 2 == 0) orig_data[i][j] = -orig_data[i][j]; } @@ -5310,7 +5310,7 @@ test_scaleoffset_float_2(hid_t file) for (j = 0; j < (size_t)size[1]; j++) { orig_data[0][j] = (float)(HDrandom() % 100000) / 1000.0F; - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((j + 1) % 2 == 0) orig_data[0][j] = -orig_data[0][j]; } @@ -5431,7 +5431,7 @@ test_scaleoffset_double(hid_t file) for (j = 0; j < (size_t)size[1]; j++) { orig_data[i][j] = (float)(HDrandom() % 10000000) / 10000000.0F; - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((i * size[1] + j + 1) % 2 == 0) orig_data[i][j] = -orig_data[i][j]; } @@ -5577,7 +5577,7 @@ test_scaleoffset_double_2(hid_t file) for (j = 0; j < (size_t)size[1]; j++) { orig_data[0][j] = (float)(HDrandom() % 10000000) / 10000000.0F; - /* even-numbered values are negtive */ + /* even-numbered values are negative */ if ((j + 1) % 2 == 0) orig_data[0][j] = -orig_data[0][j]; } @@ -7712,7 +7712,7 @@ test_missing_chunk(hid_t file) if (H5Dread(did2, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, rdata2_bytes) < 0) TEST_ERROR; - /* Validata values read for the 1-D dataset */ + /* Validate values read for the 1-D dataset */ for (u = 0; u < MISSING_CHUNK_DIM; u++) { if ((u % 10) >= 5) { if (rdata[u] != 911) { @@ -7729,7 +7729,7 @@ test_missing_chunk(hid_t file) } /* end else */ } /* end for */ - /* Validata values read for the 2-D dataset */ + /* Validate values read for the 2-D dataset */ for (i = 0; i < MISSING_CHUNK_DIM; i++) { for (j = 0; j < MISSING_CHUNK_DIM; j++) { @@ -8835,7 +8835,7 @@ test_chunk_cache(hid_t fapl) if ((dsid = H5Dcreate2(fid, "dset", H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, dapl1)) < 0) FAIL_STACK_ERROR - /* Retrieve dapl from dataset, verfiy cache values are the same as on fapl_local */ + /* Retrieve dapl from dataset, verify cache values are the same as on fapl_local */ if ((dapl2 = H5Dget_access_plist(dsid)) < 0) FAIL_STACK_ERROR if (H5Pget_chunk_cache(dapl2, &nslots_4, &nbytes_4, &w0_4) < 0) @@ -8859,7 +8859,7 @@ test_chunk_cache(hid_t fapl) if ((dsid = H5Oopen(fid, "dset", dapl1)) < 0) FAIL_STACK_ERROR - /* Retrieve dapl from dataset, verfiy cache values are the same as on dapl1 */ + /* Retrieve dapl from dataset, verify cache values are the same as on dapl1 */ /* Note we rely on the knowledge that H5Pget_chunk_cache retrieves these * values directly from the dataset structure, and not from a copy of the * dapl used to open the dataset (which is not preserved). @@ -8879,7 +8879,7 @@ test_chunk_cache(hid_t fapl) if ((dsid = H5Dopen2(fid, "dset", H5P_DEFAULT)) < 0) FAIL_STACK_ERROR - /* Retrieve dapl from dataset, verfiy cache values are the same on fapl_local */ + /* Retrieve dapl from dataset, verify cache values are the same on fapl_local */ if ((dapl2 = H5Dget_access_plist(dsid)) < 0) FAIL_STACK_ERROR if (H5Pget_chunk_cache(dapl2, &nslots_4, &nbytes_4, &w0_4) < 0) @@ -8889,7 +8889,7 @@ test_chunk_cache(hid_t fapl) if (H5Pclose(dapl2) < 0) FAIL_STACK_ERROR - /* Similary, test use of H5Dcreate2 with H5P_DEFAULT */ + /* Similarly, test use of H5Dcreate2 with H5P_DEFAULT */ if (H5Dclose(dsid) < 0) FAIL_STACK_ERROR if ((dsid = H5Dcreate2(fid, "dset2", H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT)) < 0) @@ -11974,7 +11974,7 @@ error: * Purpose: Checks that a file created with either: * (a) SWMR-write + non-latest-format * (b) write + latest format - * will generate datset with latest chunk indexing type. + * will generate dataset with latest chunk indexing type. * * Return: Success: 0 * Failure: -1 @@ -14711,7 +14711,7 @@ test_versionbounds(void) hid_t dcpl = -1; /* Dataset creation property list */ hid_t srcspace = -1; /* Source dataspaces */ hid_t vspace = -1; /* Virtual dset dataspaces */ - hid_t srcdset = -1; /* Source datset */ + hid_t srcdset = -1; /* Source dataset */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[1] = {3}; /* Data space current size */ char srcfilename[FILENAME_BUF_SIZE]; @@ -14847,7 +14847,7 @@ error: /*----------------------------------------------------------------------------- * Function: test_object_header_minimization_dcpl * - * Purpose: Test the "datset object header minimization" property as part of + * Purpose: Test the "dataset object header minimization" property as part of * the DCPL. * * Return: Success/pass: 0 @@ -14889,15 +14889,13 @@ test_object_header_minimization_dcpl(void) /* TESTS */ /*********/ - /* default value (not set explicitly) - */ + /* Default value (not set explicitly) */ if (H5Pget_dset_no_attrs_hint(dcpl_id, &minimize) == FAIL) TEST_ERROR if (FALSE != minimize) TEST_ERROR - /* FALSE-set value - */ + /* FALSE-set value */ if (H5Pset_dset_no_attrs_hint(dcpl_id, FALSE) == FAIL) TEST_ERROR if (H5Pget_dset_no_attrs_hint(dcpl_id, &minimize) == FAIL) @@ -14905,8 +14903,7 @@ test_object_header_minimization_dcpl(void) if (FALSE != minimize) TEST_ERROR - /* TRUE-set value - */ + /* TRUE-set value */ if (H5Pset_dset_no_attrs_hint(dcpl_id, TRUE) == FAIL) TEST_ERROR if (H5Pget_dset_no_attrs_hint(dcpl_id, &minimize) == FAIL) @@ -15190,7 +15187,7 @@ main(void) double rdcc_w0; int nerrors = 0; const char *envval; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ int i; /* Don't run this test using certain file drivers */ @@ -15198,7 +15195,7 @@ main(void) if (envval == NULL) envval = "nomatch"; - /* Current VFD that does not support contigous address space */ + /* Current VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0); /* Set the random # seed */ diff --git a/test/dt_arith.c b/test/dt_arith.c index f51ccd6..7f99649 100644 --- a/test/dt_arith.c +++ b/test/dt_arith.c @@ -27,7 +27,7 @@ #define FP_EPSILON 0.000001F /* - * Offset from alinged memory returned by malloc(). This can be used to test + * Offset from aligned memory returned by malloc(). This can be used to test * that type conversions handle non-aligned buffers correctly. */ #define ALIGNMENT 1 @@ -225,7 +225,7 @@ static int without_hardware_g = 0; /*The number of values if multiplied by 10 for each step.*/ \ num_norm = (SRC_MAX_10_EXP - SRC_MIN_10_EXP); \ /*Reduce the number of values by 2^factor. MULTIPLY=10^(2^factor). Using this algorithm \ - *instead of arithmatic operation to avoid any conversion*/ \ + *instead of arithmetic operation to avoid any conversion*/ \ num_norm >>= factor; \ \ /*Total number of values*/ \ @@ -644,7 +644,7 @@ test_particular_fp_integer(void) float src_f = (float)INT_MAX; int dst_i; int fill_value = 13; - int endian; /*endianess */ + int endian; /*endianness */ unsigned int fails_this_test = 0; size_t j; @@ -822,7 +822,7 @@ test_derived_flt(void) size_t src_size, dst_size; unsigned char *buf = NULL, *saved_buf = NULL; int *aligned = NULL; - int endian; /*endianess */ + int endian; /*endianness */ size_t nelmts = NTESTELEM; unsigned int fails_this_test = 0; const size_t max_fails = 40; /*max number of failures*/ @@ -1271,7 +1271,7 @@ test_derived_integer(void) char filename[1024]; size_t src_size, dst_size; unsigned char *buf = NULL, *saved_buf = NULL; - int endian; /*endianess */ + int endian; /*endianness */ size_t nelmts = NTESTELEM; unsigned int fails_this_test = 0; const size_t max_fails = 40; /*max number of failures*/ @@ -1595,7 +1595,7 @@ test_conv_int_1(const char *name, hid_t src, hid_t dst) dtype_t src_type, dst_type; /*data types */ const char *src_type_name = NULL; /*source type name */ const char *dst_type_name = NULL; /*destination type name */ - int endian; /*machine endianess */ + int endian; /*machine endianness */ size_t src_size, dst_size; /*type sizes */ unsigned char *buf = NULL; /*buffer for conversion */ unsigned char *saved = NULL; /*original values */ @@ -2835,8 +2835,8 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst) int overflow = 0; /*overflow occurred */ int uflow = 0; /*underflow debug counters*/ size_t j, k; /*counters */ - int sendian; /* source type endianess */ - int dendian; /* Destination type endianess */ + int sendian; /* source type endianness */ + int dendian; /* Destination type endianness */ size_t dst_ebias; /* Destination type's exponent bias */ size_t src_epos; /* Source type's exponent position */ size_t src_esize; /* Source type's exponent size */ @@ -3153,7 +3153,7 @@ test_conv_flt_1(const char *name, int run_test, hid_t src, hid_t dst) /* * Assume same if both results are NaN. There are many NaN bit - * patterns and the software doesn't attemt to emulate the + * patterns and the software doesn't attempt to emulate the * hardware in this regard. Instead, software uses a single bit * pattern for NaN by setting the significand to all ones. */ @@ -3432,8 +3432,8 @@ test_conv_int_fp(const char *name, int run_test, hid_t src, hid_t dst) dtype_t dst_type; /*data types */ const char *src_type_name = NULL; /*source type name */ const char *dst_type_name = NULL; /*destination type name */ - int sendian; /*source endianess */ - int dendian; /*destination endianess */ + int sendian; /*source endianness */ + int dendian; /*destination endianness */ size_t src_size, dst_size; /*type sizes */ unsigned char *buf = NULL; /*buffer for conversion */ unsigned char *saved = NULL; /*original values */ @@ -3682,8 +3682,8 @@ test_conv_int_fp(const char *name, int run_test, hid_t src, hid_t dst) } /* Allocate and initialize the source buffer through macro INIT_INTEGER if the source is integer, - * INIT_FP_NORM if floating-point. The BUF will be used for the conversion while the SAVED buffer will be - * used for the comparison later. + * INIT_FP_NORM if floating-point. The BUF will be used for the conversion while the SAVED buffer + * will be used for the comparison later. */ if (src_type == INT_SCHAR) { INIT_INTEGER(signed char, SCHAR_MAX, SCHAR_MIN, src_size, dst_size, src_nbits, buf, saved, nelmts); diff --git a/test/dtypes.c b/test/dtypes.c index 17853f0..6f753b4 100644 --- a/test/dtypes.c +++ b/test/dtypes.c @@ -29,7 +29,7 @@ #define ARRAY_DIM 4 /* - * Offset from alinged memory returned by malloc(). This can be used to test + * Offset from aligned memory returned by malloc(). This can be used to test * that type conversions handle non-aligned buffers correctly. */ #define ALIGNMENT 1 @@ -2639,7 +2639,7 @@ test_compound_13(void) data_out.x[u] = (unsigned char)u; data_out.y = 99.99f; - /* Set latest_format in access propertly list to enable the latest + /* Set latest_format in access property list to enable the latest * compound datatype format. */ if ((fapl_id = H5Pcreate(H5P_FILE_ACCESS)) < 0) @@ -3924,7 +3924,7 @@ error: * * Modifications: * Raymond Lu - * Wednesday, Febuary 9, 2005 + * Wednesday, February 9, 2005 * Added test for H5Tenum_valueof, H5Tenum_nameof, and * H5Tget_member_value. *------------------------------------------------------------------------- @@ -6977,7 +6977,7 @@ test_set_order(void) { hid_t dtype; /* Datatype ID */ H5T_order_t order; /* Byte order */ - hsize_t dims[2] = {3, 4}; /* Array dimenstions */ + hsize_t dims[2] = {3, 4}; /* Array dimensions */ herr_t ret; /* Generic return value */ TESTING("H5Tset/get_order"); @@ -7225,7 +7225,7 @@ test_set_order_compound(hid_t fapl) hid_t file = -1; hid_t cmpd = -1, memb_cmpd = -1, memb_array1 = -1, memb_array2 = -1, cmpd_array = -1; hid_t vl_id = -1; - hsize_t dims[2] = {3, 4}; /* Array dimenstions */ + hsize_t dims[2] = {3, 4}; /* Array dimensions */ char filename[1024]; herr_t ret; /* Generic return value */ @@ -8080,7 +8080,7 @@ test_deprec(hid_t fapl) unsigned u; /* Local index variable */ herr_t status; /* Generic routine value */ - TESTING("deprected API routines for datatypes"); + TESTING("deprecated API routines for datatypes"); /* Create an array datatype with an atomic base type */ /* (dimension permutations allowed, but not stored) */ diff --git a/test/earray.c b/test/earray.c index fa9dd0a..40cce1b 100644 --- a/test/earray.c +++ b/test/earray.c @@ -1283,7 +1283,7 @@ typedef struct eiter_fw_t { /*------------------------------------------------------------------------- * Function: eiter_fw_init * - * Purpose: Initialize element interator (forward iteration) + * Purpose: Initialize element iterator (forward iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1427,7 +1427,7 @@ eiter_fw_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa /*------------------------------------------------------------------------- * Function: eiter_fw_term * - * Purpose: Shut down element interator (forward iteration) + * Purpose: Shut down element iterator (forward iteration) * * Return: Success: 0 * Failure: -1 @@ -1471,7 +1471,7 @@ typedef struct eiter_rv_t { /*------------------------------------------------------------------------- * Function: eiter_rv_init * - * Purpose: Initialize element interator (reverse iteration) + * Purpose: Initialize element iterator (reverse iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1652,7 +1652,7 @@ eiter_rv_state(void *in_eiter, const H5EA_create_t *cparam, const earray_test_pa /*------------------------------------------------------------------------- * Function: eiter_rv_term * - * Purpose: Shut down element interator (reverse iteration) + * Purpose: Shut down element iterator (reverse iteration) * * Return: Success: 0 * Failure: -1 @@ -1693,7 +1693,7 @@ typedef struct eiter_rnd_t { /*------------------------------------------------------------------------- * Function: eiter_rnd_init * - * Purpose: Initialize element interator (random iteration) + * Purpose: Initialize element iterator (random iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1802,7 +1802,7 @@ eiter_rnd_max(const void *in_eiter) /*------------------------------------------------------------------------- * Function: eiter_rnd_term * - * Purpose: Shut down element interator (random iteration) + * Purpose: Shut down element iterator (random iteration) * * Return: Success: 0 * Failure: -1 @@ -1842,7 +1842,7 @@ static const earray_iter_t ea_iter_rnd = { /*------------------------------------------------------------------------- * Function: eiter_rnd2_init * - * Purpose: Initialize element interator (random #2 iteration) + * Purpose: Initialize element iterator (random #2 iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1925,7 +1925,7 @@ typedef struct eiter_cyc_t { /*------------------------------------------------------------------------- * Function: eiter_cyc_init * - * Purpose: Initialize element interator (cyclic iteration) + * Purpose: Initialize element iterator (cyclic iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -2018,7 +2018,7 @@ eiter_cyc_max(const void *in_eiter) /*------------------------------------------------------------------------- * Function: eiter_cyc_term * - * Purpose: Shut down element interator (cyclic iteration) + * Purpose: Shut down element iterator (cyclic iteration) * * Return: Success: 0 * Failure: -1 @@ -1,4 +1,5 @@ /* * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * + * Copyright by The HDF Group. * * Copyright by the Board of Trustees of the University of Illinois. * * All rights reserved. * * * diff --git a/test/enc_dec_plist_cross_platform.c b/test/enc_dec_plist_cross_platform.c index 8fb6d0b..d843f98 100644 --- a/test/enc_dec_plist_cross_platform.c +++ b/test/enc_dec_plist_cross_platform.c @@ -28,7 +28,7 @@ int main(void) { if (VERBOSE_MED) - HDprintf("Encode/Decode property list endianess\n"); + HDprintf("Encode/Decode property list endianness\n"); /******* ENCODE/DECODE DCPLS *****/ TESTING("Default DCPL Encoding/Decoding"); diff --git a/test/enum.c b/test/enum.c index 3923892..5695a79 100644 --- a/test/enum.c +++ b/test/enum.c @@ -541,7 +541,7 @@ test_value_dsnt_exist(void) hid_t datatype_id = (-1); /* identifiers */ int val; - char nam[100]; + char name[100]; size_t size = 100; TESTING("for non-existing name and value"); /* Turn off error reporting since we expect failure in this test */ @@ -552,11 +552,11 @@ test_value_dsnt_exist(void) if ((datatype_id = H5Tenum_create(H5T_NATIVE_INT)) < 0) goto error; - /* These calls should fail, since no memebrs exist yet */ + /* These calls should fail, since no members exist yet */ if (H5Tenum_valueof(datatype_id, "SAX", &val) >= 0) goto error; val = 3; - if (H5Tenum_nameof(datatype_id, &val, nam, size) >= 0) + if (H5Tenum_nameof(datatype_id, &val, name, size) >= 0) goto error; val = 2; @@ -571,17 +571,17 @@ test_value_dsnt_exist(void) /* This call should fail since we did not create a member with value = 3*/ val = 3; - if (H5Tenum_nameof(datatype_id, &val, nam, size) >= 0) + if (H5Tenum_nameof(datatype_id, &val, name, size) >= 0) goto error; /* This call should fail since we did not create a member with value = 11*/ val = 11; - if (H5Tenum_nameof(datatype_id, &val, nam, size) >= 0) + if (H5Tenum_nameof(datatype_id, &val, name, size) >= 0) goto error; /* This call should fail since we did not create a member with value = 0*/ val = 0; - if (H5Tenum_nameof(datatype_id, &val, nam, size) >= 0) + if (H5Tenum_nameof(datatype_id, &val, name, size) >= 0) goto error; /* This call should fail since we do not have SAX name in the type */ diff --git a/test/external.c b/test/external.c index b3250c1..41a0eb8 100644 --- a/test/external.c +++ b/test/external.c @@ -1217,7 +1217,7 @@ test_path_relative_cwd(hid_t fapl) } H5E_END_TRY; if (dset3 >= 0) - FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded"); + FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded"); /* Read the entire dataset and compare with the original */ HDmemset(whole, 0, sizeof(whole)); @@ -1233,7 +1233,7 @@ test_path_relative_cwd(hid_t fapl) if (H5Dclose(dset) < 0) FAIL_STACK_ERROR - /* Open dataset (use a differend prefix than for create. + /* Open dataset (use a different prefix than for create. * This works because the dataset was closed. */ if (H5Pset_efile_prefix(dapl2, "${ORIGIN}/.") < 0) @@ -1254,7 +1254,7 @@ test_path_relative_cwd(hid_t fapl) } H5E_END_TRY; if (dset3 >= 0) - FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeded"); + FAIL_PUTS_ERROR("reopening the dataset with a different efile_prefix succeeded"); /* Read the entire dataset and compare with the original */ HDmemset(whole, 0, sizeof(whole)); diff --git a/test/farray.c b/test/farray.c index 9f7e43f..ab04b68 100644 --- a/test/farray.c +++ b/test/farray.c @@ -968,7 +968,7 @@ typedef struct fiter_fw_t { /*------------------------------------------------------------------------- * Function: fiter_fw_init * - * Purpose: Initialize element interator (forward iteration) + * Purpose: Initialize element iterator (forward iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1020,7 +1020,7 @@ fiter_fw_next(void *_fiter) /*------------------------------------------------------------------------- * Function: fiter_term * - * Purpose: Shut down element interator (simple iterators) + * Purpose: Shut down element iterator (simple iterators) * * Return: Success: 0 * Failure: -1 @@ -1054,7 +1054,7 @@ typedef struct fiter_rv_t { /*------------------------------------------------------------------------- * Function: fiter_rv_init * - * Purpose: Initialize element interator (reverse iteration) + * Purpose: Initialize element iterator (reverse iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1119,7 +1119,7 @@ typedef struct fiter_rnd_t { /*------------------------------------------------------------------------- * Function: fiter_rnd_init * - * Purpose: Initialize element interator (random iteration) + * Purpose: Initialize element iterator (random iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL @@ -1192,7 +1192,7 @@ fiter_rnd_next(void *_fiter) /*------------------------------------------------------------------------- * Function: fiter_rnd_term * - * Purpose: Shut down element interator (random iteration) + * Purpose: Shut down element iterator (random iteration) * * Return: Success: 0 * Failure: -1 @@ -1234,7 +1234,7 @@ typedef struct fiter_cyc_t { /*------------------------------------------------------------------------- * Function: fiter_cyc_init * - * Purpose: Initialize element interator (cyclic iteration) + * Purpose: Initialize element iterator (cyclic iteration) * * Return: Success: Pointer to iteration status object * Failure: NULL diff --git a/test/fheap.c b/test/fheap.c index 8969ab4..c584cb5 100644 --- a/test/fheap.c +++ b/test/fheap.c @@ -2469,7 +2469,7 @@ test_id_limits(hid_t fapl, H5HF_create_t *cparam, hid_t fcpl) size_t id_len; /* Size of fractal heap IDs */ size_t tiny_max_len; /* Max. length of tiny objects */ hbool_t tiny_len_extended; /* Do tiny objects use two bytes for the length? */ - hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */ + hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */ /* Set the filename to use for this test (dependent on fapl) */ h5_fixname(FILENAME[0], fapl, filename, sizeof(filename)); @@ -13785,7 +13785,7 @@ test_filtered_huge(hid_t fapl, H5HF_create_t *cparam, fheap_test_param_t *tparam fheap_heap_state_t state; /* State of fractal heap */ unsigned deflate_level; /* Deflation level */ size_t old_actual_id_len = 0; /* Old actual ID length */ - hbool_t huge_ids_direct; /* Are 'huge' objects directly acccessed? */ + hbool_t huge_ids_direct; /* Are 'huge' objects directly accessed? */ /* Test description */ const char *base_desc = "insert 'huge' object into heap with I/O filters, then remove %s"; @@ -16352,7 +16352,7 @@ main(void) unsigned num_pb_fs = 1; /* The number of settings to test for page buffering and file space handling */ int ExpressMode; /* Express testing level */ const char *envval; /* Environment variable */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ hbool_t api_ctx_pushed = FALSE; /* Whether API context pushed */ /* Don't run this test using certain file drivers */ @@ -16360,7 +16360,7 @@ main(void) if (envval == NULL) envval = "nomatch"; - /* Current VFD that does not support contigous address space */ + /* Current VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(envval, "split") != 0 && HDstrcmp(envval, "multi") != 0); /* Reset library */ diff --git a/test/file_image.c b/test/file_image.c index 57f0ac6..5b794c8 100644 --- a/test/file_image.c +++ b/test/file_image.c @@ -145,7 +145,7 @@ test_properties(void) /* Check that sizes are the same, and that the buffers are identical but separate */ VERIFY(temp_size == size, "Sizes of buffers don't match"); - VERIFY(temp2 != NULL, "Recieved buffer not set"); + VERIFY(temp2 != NULL, "Received buffer not set"); VERIFY(temp2 != buffer, "Retrieved buffer is the same as original"); VERIFY(temp2 != temp, "Retrieved buffer is the same as previously retrieved buffer"); VERIFY(0 == HDmemcmp(temp2, buffer, size), "Buffers contain different data"); @@ -260,7 +260,7 @@ free_cb(void *ptr, H5FD_file_image_op_t op, void *udata) * Function: udata_copy_cb * * Purpose: This function allows calls to the udata_copy callback to be tracked. - * No copying actualy takes place; it is easier to deal with only one + * No copying actually takes place; it is easier to deal with only one * instance of the udata. * * Returns: A pointer to the same udata that was passed in. @@ -610,7 +610,7 @@ test_core(void) VERIFY((udata->used_callbacks == MALLOC) || (udata->used_callbacks == (MALLOC | UDATA_COPY | UDATA_FREE)), "opening a core file used the wrong callbacks"); VERIFY(udata->malloc_src == H5FD_FILE_IMAGE_OP_FILE_OPEN, - "Malloc callback came from wrong sourc in core open"); + "Malloc callback came from wrong source in core open"); /* Close file */ reset_udata(udata); @@ -618,7 +618,7 @@ test_core(void) VERIFY(ret >= 0, "H5Fclose failed"); VERIFY(udata->used_callbacks == FREE, "Closing a core file used the wrong callbacks"); VERIFY(udata->free_src == H5FD_FILE_IMAGE_OP_FILE_CLOSE, - "Free callback came from wrong sourc in core close"); + "Free callback came from wrong source in core close"); /* Reopen file */ file = H5Fopen(copied_filename, H5F_ACC_RDWR, fapl); @@ -654,7 +654,7 @@ test_core(void) VERIFY(ret >= 0, "H5Fclose failed"); VERIFY(udata->used_callbacks == (FREE), "Closing a core file used the wrong callbacks"); VERIFY(udata->free_src == H5FD_FILE_IMAGE_OP_FILE_CLOSE, - "Free callback came from wrong sourc in core close"); + "Free callback came from wrong source in core close"); /* Create file image buffer */ fd = HDopen(copied_filename, O_RDONLY); @@ -954,7 +954,7 @@ test_get_file_image(const char *test_banner, const int file_name_num, hid_t fapl err = H5Fclose(core_file_id); VERIFY(err == SUCCEED, "H5Fclose(core_file_id) failed."); - /* dicard core fapl */ + /* discard core fapl */ err = H5Pclose(core_fapl_id); VERIFY(err == SUCCEED, "H5Pclose(core_fapl_id) failed."); @@ -1078,7 +1078,7 @@ test_get_file_image_error_rejection(void) VERIFY(bytes_read < 0, "H5Fget_file_image(2 -- test 1) succeeded."); /* Call H5Fget_file_image() with good buffer and buffer size, - * but non-existant file_id. Should fail. + * but non-existent file_id. Should fail. */ H5E_BEGIN_TRY { diff --git a/test/fillval.c b/test/fillval.c index be3ad14..a0ceac7 100644 --- a/test/fillval.c +++ b/test/fillval.c @@ -1918,7 +1918,7 @@ test_extend_cases(hid_t file, hid_t _dcpl, const char *dset_name, const hsize_t if ((fspace = H5Dget_space(dset)) < 0) TEST_ERROR - /* Create dataspace for single element sized bufer */ + /* Create dataspace for single element sized buffer */ if ((mspace = H5Screate_simple(5, one, NULL)) < 0) TEST_ERROR @@ -2142,7 +2142,7 @@ test_extend(hid_t fapl, const char *base_name, H5D_layout_t layout) #if 1 /* - * Remove this when contiguous datasets can be exended to some + * Remove this when contiguous datasets can be extended to some * predetermined fininte size, even if it's just in the slowest varying * dimension. If it's removed prematurely then you'll get one of the * errors described above or `unable to select fill value region'. diff --git a/test/filter_fail.c b/test/filter_fail.c index cd264ee..64a3df2 100644 --- a/test/filter_fail.c +++ b/test/filter_fail.c @@ -160,7 +160,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled) points[i] = i; /* Write data. If the chunk cache is enabled, H5Dwrite should succeed. If it is - * diabled, H5Dwrite should fail. */ + * disabled, H5Dwrite should fail. */ if (cache_enabled) { if (H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, sid, H5P_DEFAULT, points) < 0) TEST_ERROR @@ -186,7 +186,7 @@ test_filter_write(char *file_name, hid_t my_fapl, hbool_t cache_enabled) TEST_ERROR /* Close dataset. If the chunk cache is enabled, the flushing of chunks should fail - * during H5Dclose. If it is diabled, H5Dwrite should fail but H5Dclose should succeed. */ + * during H5Dclose. If it is disabled, H5Dwrite should fail but H5Dclose should succeed. */ if (cache_enabled) { H5E_BEGIN_TRY { diff --git a/test/filter_plugin.c b/test/filter_plugin.c index ea5c000..50572c5 100644 --- a/test/filter_plugin.c +++ b/test/filter_plugin.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/test/filter_plugin1_dsets.c b/test/filter_plugin1_dsets.c index b22ba24..3a9e714 100644 --- a/test/filter_plugin1_dsets.c +++ b/test/filter_plugin1_dsets.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * @@ -79,7 +79,7 @@ add_sub_value(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_value value = (int)cd_values[0]; if (flags & H5Z_FLAG_REVERSE) { - /* READ - Substract the given value from all the data values */ + /* READ - Subtract the given value from all the data values */ while (buf_left > 0) { *int_ptr++ -= value; buf_left -= sizeof(int); diff --git a/test/filter_plugin2_dsets.c b/test/filter_plugin2_dsets.c index fed8176..6e8c1e1 100644 --- a/test/filter_plugin2_dsets.c +++ b/test/filter_plugin2_dsets.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/test/filter_plugin3_dsets.c b/test/filter_plugin3_dsets.c index e8103a3..c901355 100644 --- a/test/filter_plugin3_dsets.c +++ b/test/filter_plugin3_dsets.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * @@ -108,7 +108,7 @@ add_sub_value_hdf5(unsigned int flags, size_t cd_nelmts, const unsigned int *cd_ value = (int)cd_values[0]; if (flags & H5Z_FLAG_REVERSE) { - /* READ - Substract the given value from all the data values */ + /* READ - Subtract the given value from all the data values */ while (buf_left > 0) { *int_ptr++ -= value; buf_left -= sizeof(int); diff --git a/test/filter_plugin4_groups.c b/test/filter_plugin4_groups.c index 97cb59a..7e576a3 100644 --- a/test/filter_plugin4_groups.c +++ b/test/filter_plugin4_groups.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/test/flushrefresh.c b/test/flushrefresh.c index 0b894ef..08f2e9c 100644 --- a/test/flushrefresh.c +++ b/test/flushrefresh.c @@ -63,10 +63,10 @@ #define NOT_FLUSHED "NOT_FLUSHED" /* Error Handling */ -/* For errors occuring in the main process, use the standard TEST_ERROR macro. +/* For errors occurring in the main process, use the standard TEST_ERROR macro. For errors occurring in the spawned process (from the test script), use the PROCESS_ERROR macro, which will send a signal to the main process so the - main process can propogate errors correctly. */ + main process can propagate errors correctly. */ FILE *errorfile; #define ERRFILE "flushrefresh_ERROR" #define PROCESS_ERROR \ @@ -222,7 +222,7 @@ test_flush(void) * object, a series of verifications will occur on each object in the file. * * Each verification consists of spawning off a new process and determining - * if the object can be opened and its information retreived in said + * if the object can be opened and its information retrieved in said * alternate process. It reports the results, which are compared to an * expected value (either that the object can be found on disk, or that it * cannot). @@ -695,7 +695,7 @@ test_refresh(void) * that an attribute has been added, and is only visible after a * successful call to a H5*refresh function). * - * As with the flush case, the implemention is a bit tricky as it's + * As with the flush case, the implementation is a bit tricky as it's * dealing with signals going back and forth between the two processes * to ensure the timing is correct, but basically, an example: * @@ -1166,7 +1166,7 @@ error: * Purpose: This function is used to communicate with the verification * process spawned by the start_refresh_verification_process * function. It gives it the go-ahead to call H5*refresh - * on an object and conlcude the refresh verification. + * on an object and conclude the refresh verification. * * Return: 0 on Success, 1 on Failure * @@ -1360,7 +1360,7 @@ error: * * Purpose: This function checks the status of external verification * processes to see if they've succeeded. It checks for the - * existance of flushrefresh_ERROR file. If present, that indicates + * existence of flushrefresh_ERROR file. If present, that indicates * an external verification process has failed, and this function * thus fails as well. If not present, then nothing else has * failed, and this function succeeds. diff --git a/test/gen_new_array.c b/test/gen_new_array.c index 5811106..8161c4e 100644 --- a/test/gen_new_array.c +++ b/test/gen_new_array.c @@ -89,7 +89,7 @@ main(void) if (ret < 0) printf("field 1 insert<0!\n"); - /* Creat the array datatype */ + /* Create the array datatype */ arr_type = H5Tarray_create2(H5T_NATIVE_FLOAT, ARRAY1_RANK, tdims1); if (arr_type < 0) printf("arr_type<0!\n"); @@ -104,7 +104,7 @@ main(void) if (ret < 0) printf("field 3 array close<0!\n"); - /* Creat the array datatype */ + /* Create the array datatype */ arr_type = H5Tarray_create2(H5T_NATIVE_LONG, ARRAY1_RANK, tdims1); if (arr_type < 0) printf("arr_type<0!\n"); diff --git a/test/gen_old_group.c b/test/gen_old_group.c index 1b4e0b2..15e67a4 100644 --- a/test/gen_old_group.c +++ b/test/gen_old_group.c @@ -19,7 +19,7 @@ * empty "symbol table" group. * * This file is used in the v1.7 branch (after the "compact group" - * checkin) to test compatability. Compile and run this + * checkin) to test compatibility. Compile and run this * program (with the 1.6.x branch), it will generate a file* * called "group_old.h5". You need to move it to the test * directory in the HDF5 v1.7 source tree. The test/stab.c diff --git a/test/gen_plist.c b/test/gen_plist.c index 8e5e0f1..4840f52 100644 --- a/test/gen_plist.c +++ b/test/gen_plist.c @@ -81,7 +81,7 @@ main(void) H5AC_cache_image_config_t my_cache_image_config = {H5AC__CURR_CACHE_IMAGE_CONFIG_VERSION, TRUE, FALSE, -1}; - /* check endianess */ + /* check endianness */ { short int word = 0x0001; char *byte = (char *)&word; diff --git a/test/genall5.c b/test/genall5.c index 948e3f9..46d6bb0 100644 --- a/test/genall5.c +++ b/test/genall5.c @@ -14,7 +14,7 @@ /* Programmer: John Mainzer * 9/23/15 * - * This file contains a heavily edited and functionaly reduce + * This file contains a heavily edited and functionally reduce * version of the test code first written by Quincey in a file * of the same name. */ diff --git a/test/getname.c b/test/getname.c index 30ab98d..b8b88c1 100644 --- a/test/getname.c +++ b/test/getname.c @@ -1526,7 +1526,7 @@ test_main(hid_t file_id, hid_t fapl) { char name[NAME_BUF_SIZE]; /* Buffer to hold name and its size */ - /* Get name for non commited datatype, it should fail */ + /* Get name for non-commited datatype, it should fail */ H5E_BEGIN_TRY { if (H5Iget_name(type_id, name, NAME_BUF_SIZE) > 0) diff --git a/test/gheap.c b/test/gheap.c index 71b3bd1..7eb99d2 100644 --- a/test/gheap.c +++ b/test/gheap.c @@ -352,7 +352,7 @@ error: * Function: test_4 * * Purpose: Tests the H5HG_remove() feature by writing lots of objects - * and occassionally removing some. When we're done they're all + * and occasionally removing some. When we're done they're all * removed. * * Return: Success: 0 diff --git a/test/h5test.c b/test/h5test.c index cdd9da8..6f89f5b 100644 --- a/test/h5test.c +++ b/test/h5test.c @@ -1406,7 +1406,7 @@ h5_szip_can_encode(void) * further calls to getenv_all take place. * * Return: pointer to a string containing the value of the environment variable - * NULL if the varialbe doesn't exist in task 'root's environment. + * NULL if the variable doesn't exist in task 'root's environment. * * Programmer: Leon Arber * 4/4/05 @@ -1701,7 +1701,7 @@ h5_send_message(const char *send, const char *arg1, const char *arg2) * signal file from disk, and only continues once it has * successfully done so (i.e., only after another process has * called the "h5_send_message" function to write the signal file). - * This functon will then immediately remove the file (i.e., + * This function will then immediately remove the file (i.e., * to indicate that it has been received and can be reused), * and then exits, allowing the calling function to continue. * diff --git a/test/hdfs.c b/test/hdfs.c index f1c33c4..0bb1aa0 100644 --- a/test/hdfs.c +++ b/test/hdfs.c @@ -52,7 +52,7 @@ * 4) Configurable expected-actual order in generated comparison strings. * Some prefer `VERIFY(expected, actual)`, others * `VERIFY(actual, expected)`. Provide preprocessor ifdef switch - * to satifsy both parties, assuming one paradigm per test file. + * to satisfy both parties, assuming one paradigm per test file. * (One could #undef and redefine the flag through the file as desired, * but _why_.) * diff --git a/test/links.c b/test/links.c index e6e385e..7f52ecf 100644 --- a/test/links.c +++ b/test/links.c @@ -3206,7 +3206,7 @@ external_link_dangling(hid_t fapl, hbool_t new_format) if (status >= 0) { H5_FAILED(); HDputs( - " Retreiving name of object by index through dangling file external link should have failed."); + " Retrieving name of object by index through dangling file external link should have failed."); } /* end if */ /* Close root group */ @@ -3679,7 +3679,7 @@ external_link_abstar(hid_t fapl, hbool_t new_format) } H5E_END_TRY; - /* should be able to find the target file with abolute path */ + /* should be able to find the target file with absolute path */ if (gid < 0) { H5_FAILED(); HDputs(" Should have found the file in tmp_links directory."); @@ -11572,7 +11572,7 @@ link_filters(hid_t fapl, hbool_t new_format) TEST_ERROR filesize_unfiltered = h5_get_file_size(filename, fapl); - /* Set deflate fitler */ + /* Set deflate filter */ if (H5Pset_deflate(fcpl, 6) < 0) TEST_ERROR @@ -17256,7 +17256,7 @@ main(void) for (minimize_dset_oh = 0; minimize_dset_oh <= 1; minimize_dset_oh++) { if (minimize_dset_oh) { - HDprintf("\n-Testing with minimzed dataset object headers-\n"); + HDprintf("\n-Testing with minimized dataset object headers-\n"); dcpl_g = H5Pcreate(H5P_DATASET_CREATE); if (0 > dcpl_g) TEST_ERROR @@ -222,7 +222,7 @@ test_mf_eoa(const char *env_h5_drvr, hid_t fapl) haddr_t addr1, addr2; haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF; hsize_t ma_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MM_alloc() of file allocation"); @@ -370,7 +370,7 @@ test_mf_eoa_shrink(const char *env_h5_drvr, hid_t fapl) haddr_t addr = 0; haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_try_shrink() of file allocation: test 1"); @@ -657,7 +657,7 @@ test_mf_eoa_extend(const char *env_h5_drvr, hid_t fapl) htri_t was_extended; haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_try_extend() of file allocation: test 1"); @@ -1933,7 +1933,7 @@ test_mf_fs_extend(hid_t fapl) if (check_stats(f, f->shared->fs_man[H5FD_MEM_SUPER], &state)) TEST_ERROR - /* Remove section A from free-space manger */ + /* Remove section A from free-space manager */ if (H5MF__find_sect(f, H5FD_MEM_SUPER, (hsize_t)(TBLOCK_SIZE30 - 10), f->shared->fs_man[H5FD_MEM_SUPER], &tmp) != TRUE) TEST_ERROR @@ -2011,7 +2011,7 @@ test_mf_fs_absorb(const char *env_h5_drvr, hid_t fapl) haddr_t ma_addr = HADDR_UNDEF; hsize_t ma_size = 0; H5MF_free_section_t *sect_node = NULL; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("A free-space section absorbs an aggregator: test 1"); @@ -2187,7 +2187,7 @@ test_mf_aggr_alloc1(const char *env_h5_drvr, hid_t fapl) haddr_t addr1, addr2; haddr_t ma_addr = HADDR_UNDEF; hsize_t ma_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator:test 1"); @@ -2338,7 +2338,7 @@ test_mf_aggr_alloc2(const char *env_h5_drvr, hid_t fapl) haddr_t addr1, addr2, addr3; haddr_t ma_addr = HADDR_UNDEF; hsize_t ma_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator:test 2"); @@ -2496,7 +2496,7 @@ test_mf_aggr_alloc3(const char *env_h5_drvr, hid_t fapl) hsize_t ma_size = 0, new_ma_size = 0; haddr_t sdata_addr = HADDR_UNDEF; hsize_t sdata_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator: test 3"); @@ -2664,7 +2664,7 @@ test_mf_aggr_alloc4(const char *env_h5_drvr, hid_t fapl) haddr_t addr1, addr2, saddr1, saddr2, saddr3; haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator:test 4"); @@ -2813,7 +2813,7 @@ test_mf_aggr_alloc5(const char *env_h5_drvr, hid_t fapl) haddr_t addr1, addr2, addr3; haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator:test 5"); @@ -2951,7 +2951,7 @@ test_mf_aggr_alloc6(const char *env_h5_drvr, hid_t fapl) haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0; H5FS_stat_t state; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator:test 6"); @@ -3125,7 +3125,7 @@ test_mf_aggr_alloc7(const char *env_h5_drvr, hid_t fapl) haddr_t ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF; hsize_t ma_size = 0, sdata_size = 0; H5FS_stat_t state; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_alloc() of meta/sdata aggregator:test 7"); @@ -3296,7 +3296,7 @@ test_mf_aggr_extend(const char *env_h5_drvr, hid_t fapl) haddr_t ma_addr = HADDR_UNDEF, new_ma_addr = HADDR_UNDEF, sdata_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0, sdata_size = 0; htri_t was_extended; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_try_extend() of meta/sdata aggregator: test 1"); @@ -3594,7 +3594,7 @@ test_mf_aggr_absorb(const char *env_h5_drvr, hid_t fapl) haddr_t new_sdata_addr = HADDR_UNDEF; hsize_t ma_size = 0, new_ma_size = 0; hsize_t sdata_size = 0, new_sdata_size = 0; - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ TESTING("H5MF_try_shrink() of meta/sdata aggregator: test 1"); @@ -5837,7 +5837,7 @@ error: * * Alignment = 4096 aggr->alloc_size = 2048 * - * Allocate first block (30) from meta_aggr: (meta_aggr is emtpy) + * Allocate first block (30) from meta_aggr: (meta_aggr is empty) * request-size is > meta_aggr->size and < meta_aggr->alloc_size * Result: * A block of meta_aggr->alloc_size is allocated from file allocation @@ -7113,7 +7113,7 @@ test_mf_fs_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format) else TESTING("File's free-space is going away with old library format") - /* Current VFD that does not support contigous address space */ + /* Current VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0); if (contig_addr_vfd) { @@ -7314,7 +7314,7 @@ test_mf_strat_thres_persist(const char *env_h5_drvr, hid_t fapl, hbool_t new_for else TESTING("File space strategy/persisting/threshold with old library format") - /* Current VFD that does not support contigous address space */ + /* Current VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0); /* Set the filename to use for this test (dependent on fapl) */ @@ -7484,7 +7484,7 @@ test_mf_strat_thres_gone(const char *env_h5_drvr, hid_t fapl, hbool_t new_format else TESTING("File space merge/shrink for section size < threshold with old library format") - /* Current VFD that does not support contigous address space */ + /* Current VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0); /* Set the filename to use for this test (dependent on fapl) */ @@ -8045,7 +8045,7 @@ test_page_try_shrink(const char *env_h5_drvr, hid_t fapl) haddr_t addr1; /* Address for small metadata block */ haddr_t saddr1; /* Address for small raw data block */ haddr_t gaddr1; /* Address for large data block */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ htri_t status; /* status from shrinking */ h5_stat_size_t file_size; /* File size */ char filename[FILENAME_LEN]; /* Filename to use */ @@ -8175,7 +8175,7 @@ test_page_small_try_extend(const char *env_h5_drvr, hid_t fapl) H5F_t *f = NULL; /* Internal file object pointer */ haddr_t addr1, addr2, addr3; /* Addresses for small metadata blocks */ haddr_t saddr1; /* Address for small raw data block */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ htri_t was_extended; /* Whether the block can be extended or not */ char filename[FILENAME_LEN]; /* Filename to use */ @@ -8354,7 +8354,7 @@ test_page_large_try_extend(const char *env_h5_drvr, hid_t fapl) hid_t fcpl = -1; /* File creation property list */ H5F_t *f = NULL; /* Internal file object pointer */ haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for large data blocks */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ htri_t was_extended; /* Whether the block can be extended or not */ char filename[FILENAME_LEN]; /* Filename to use */ @@ -8517,7 +8517,7 @@ test_page_large(const char *env_h5_drvr, hid_t fapl) hid_t fcpl = -1; /* File creation property list ID */ H5F_t *f = NULL; /* Internal file object pointer */ haddr_t gaddr1, gaddr2, gaddr3, gaddr4; /* Addresses for blocks */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ H5FS_stat_t fs_stat; /* Information for free-space manager */ h5_stat_size_t file_size; /* File size */ char filename[FILENAME_LEN]; /* Filename to use */ diff --git a/test/mirror_vfd.c b/test/mirror_vfd.c index 7d838c6..c6e9ef0 100644 --- a/test/mirror_vfd.c +++ b/test/mirror_vfd.c @@ -1496,7 +1496,7 @@ create_datasets(hid_t file_id, unsigned min_dset, unsigned max_dset) } } - /* select on disk hyperslab */ + /* Select on disk hyperslab */ offset[0] = (hsize_t)i; offset[1] = (hsize_t)j; LOGPRINT(3, " H5Sselect_hyperslab()\n"); diff --git a/test/mount.c b/test/mount.c index 1841889..53379b6 100644 --- a/test/mount.c +++ b/test/mount.c @@ -726,7 +726,7 @@ test_move(hid_t fapl) H5E_END_TRY; if (status >= 0) { H5_FAILED(); - HDputs(" Moving an object across files should't have been possible"); + HDputs(" Moving an object across files shouldn't have been possible"); TEST_ERROR } /* end if */ @@ -1808,7 +1808,7 @@ test_mount_after_unmount(hid_t fapl) if (HDstrcmp(objname, "/X/M/Y") != 0) TEST_ERROR - /* Rename object in file #3 that is "disconnected" from name hiearchy */ + /* Rename object in file #3 that is "disconnected" from name hierarchy */ /* (It is "disconnected" because it's parent file has been unmounted) */ if (H5Lmove(gidAMX, "M/Y", gidAMX, "M/Z", H5P_DEFAULT, H5P_DEFAULT) < 0) TEST_ERROR diff --git a/test/mtime.c b/test/mtime.c index 255ab5d..3e51185 100644 --- a/test/mtime.c +++ b/test/mtime.c @@ -113,7 +113,7 @@ main(void) if (0 == oi1.ctime) { SKIPPED(); HDputs(" The modification time could not be decoded on this OS."); - HDputs(" Modification times will be mantained in the file but"); + HDputs(" Modification times will be maintained in the file but"); HDputs(" cannot be queried on this system. See H5O_mtime_decode()."); return 0; } diff --git a/test/objcopy.c b/test/objcopy.c index 9b055b6..aacc4a0 100644 --- a/test/objcopy.c +++ b/test/objcopy.c @@ -2750,7 +2750,7 @@ test_copy_dataset_simple_empty(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h char dst_filename[NAME_BUF_SIZE]; if (test_open) { - TESTING("H5Ocopy(): empty and openend contiguous dataset"); + TESTING("H5Ocopy(): empty and opened contiguous dataset"); } else { TESTING("H5Ocopy(): empty contiguous dataset"); @@ -3953,7 +3953,7 @@ test_copy_dataset_chunked_empty(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, if ((did = H5Dopen2(fid_src, NAME_DATASET_CHUNKED3, H5P_DEFAULT)) < 0) TEST_ERROR - /* open the copied dataset NAME_DATASET_CHUNKED3 at destinaion */ + /* open the copied dataset NAME_DATASET_CHUNKED3 at destination */ if ((did2 = H5Dopen2(fid_dst, NAME_DATASET_CHUNKED3, H5P_DEFAULT)) < 0) TEST_ERROR @@ -8633,7 +8633,7 @@ test_copy_soft_link(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t dst_fa char dst_filename[NAME_BUF_SIZE]; if (test_open) { - TESTING("H5Ocopy(): openend object through soft link"); + TESTING("H5Ocopy(): opened object through soft link"); } else { TESTING("H5Ocopy(): object through soft link"); @@ -9620,7 +9620,7 @@ test_copy_dataset_contig_named_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl char src_filename[NAME_BUF_SIZE]; char dst_filename[NAME_BUF_SIZE]; - TESTING("H5Ocopy(): contigous dataset with named VLEN datatype"); + TESTING("H5Ocopy(): contiguous dataset with named VLEN datatype"); /* set initial data values */ for (i = 0; i < DIM_SIZE_1; i++) { @@ -10380,10 +10380,10 @@ test_copy_dataset_contig_vl_vl(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, h char dst_filename[NAME_BUF_SIZE]; if (test_open) { - TESTING("H5Ocopy(): contigous and opened dataset with nested VLEN datatype"); + TESTING("H5Ocopy(): contiguous and opened dataset with nested VLEN datatype"); } else { - TESTING("H5Ocopy(): contigous dataset with nested VLEN datatype"); + TESTING("H5Ocopy(): contiguous dataset with nested VLEN datatype"); } /* set initial data values */ @@ -11009,7 +11009,7 @@ typedef struct cmpd_vl_t { * Failure: number of errors * * Programmer: Neil Fortner - * Tuseday, September 29, 2009 + * Tuesday, September 29, 2009 * *------------------------------------------------------------------------- */ @@ -13658,7 +13658,7 @@ test_copy_cdt_hier_merge(hid_t fcpl_src, hid_t fcpl_dst, hid_t src_fapl, hid_t d TEST_ERROR /* - * Test 3: copy datsets in /g0 at SRC to DST group /uncopied + * Test 3: copy datasets in /g0 at SRC to DST group /uncopied */ if (H5Ocopy(fid_src, NAME_GROUP_TOP "/" SRC_DSET1, fid_dst, NAME_GROUP_UNCOPIED "/" SRC_DSET1, ocpypl_id, H5P_DEFAULT) < 0) diff --git a/test/ohdr.c b/test/ohdr.c index bf24298..1b41d0d 100644 --- a/test/ohdr.c +++ b/test/ohdr.c @@ -583,7 +583,7 @@ test_unknown(unsigned bogus_id, char *filename, hid_t fapl) TESTING("object with unknown header message & 'shareable' flag set"); - /* Open the dataset with the unknown header message, adn "shareable" flag */ + /* Open the dataset with the unknown header message, and "shareable" flag */ if ((did = H5Dopen2(loc_bogus, "Dataset5", H5P_DEFAULT)) < 0) FAIL_STACK_ERROR if (H5Dclose(did) < 0) @@ -1088,13 +1088,13 @@ test_minimized_dset_ohdr_size_comparisons(hid_t fapl_id) /* IDs for non-minimized file open */ hid_t file_f_id = -1; /* lower 'f' for standard file setting */ hid_t dset_f_x_id = -1; /* 'x' for default */ - hid_t dset_f_N_id = -1; /* 'N' for explcit non-minimized dset */ + hid_t dset_f_N_id = -1; /* 'N' for explicit non-minimized dset */ hid_t dset_f_Y_id = -1; /* 'Y' for minimized dset */ /* IDs for minimized file open */ hid_t file_F_id = -1; /* upper 'F' for minimized file setting */ hid_t dset_F_x_id = -1; /* 'x' for default */ - hid_t dset_F_N_id = -1; /* 'N' for explcit non-minimized dset */ + hid_t dset_F_N_id = -1; /* 'N' for explicit non-minimized dset */ hid_t dset_F_Y_id = -1; /* 'Y' for minimized dset */ char filename_a[512] = ""; diff --git a/test/page_buffer.c b/test/page_buffer.c index 874aea7..f31580f 100644 --- a/test/page_buffer.c +++ b/test/page_buffer.c @@ -74,7 +74,7 @@ const char *FILENAME[] = {"filepaged", NULL}; * Any data mis-matches or failures reported by the HDF5 * library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -214,7 +214,7 @@ error: * Any data mis-matches or failures reported by the HDF5 * library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -369,12 +369,12 @@ error: * 1) verifying that API errors are caught. * * 2) verifying that the page buffer behaves more or less - * as advertized. + * as advertised. * * Any data mis-matches or unexpected failures or successes * reported by the HDF5 library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -548,7 +548,7 @@ error: * Any data mis-matches or failures reported by the HDF5 * library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -826,7 +826,7 @@ error: * Any data mis-matches or failures reported by the HDF5 * library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -1080,7 +1080,7 @@ error: * Any data mis-matches or failures reported by the HDF5 * library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -1501,7 +1501,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr) if (f->shared->page_buf->raw_count != 2) TEST_ERROR; - /* adding more meta entires should replace meta entries since raw data + /* adding more meta entries should replace meta entries since raw data * is at its minimum */ if (H5F_block_write(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 600), sizeof(int) * 100, data) < 0) @@ -1516,7 +1516,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr) if (f->shared->page_buf->raw_count != 2) TEST_ERROR; - /* bring existing raw entires up the LRU */ + /* bring existing raw entries up the LRU */ if (H5F_block_read(f, H5FD_MEM_DRAW, raw_addr + (sizeof(int) * 750), sizeof(int) * 100, data) < 0) FAIL_STACK_ERROR; @@ -1532,7 +1532,7 @@ test_min_threshold(hid_t orig_fapl, const char *env_h5_drvr) if (f->shared->page_buf->raw_count != 3) TEST_ERROR; - /* adding 2 meta entries should replace 2 entires at the bottom of the LRU */ + /* adding 2 meta entries should replace 2 entries at the bottom of the LRU */ if (H5F_block_read(f, H5FD_MEM_SUPER, meta_addr + (sizeof(int) * 98), sizeof(int) * 100, data) < 0) FAIL_STACK_ERROR; @@ -1715,7 +1715,7 @@ error: * Any data mis-matches or failures reported by the HDF5 * library result in test failure. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown @@ -2018,7 +2018,7 @@ error: * At present, page buffering should be disabled in parallel * builds. Verify this. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: John Mainzer @@ -2133,7 +2133,7 @@ error: * * Purpose: Main function for the page buffer tests. * - * Return: 0 if test is sucessful + * Return: 0 if test is successful * 1 if test fails * * Programmer: unknown diff --git a/test/reserved.c b/test/reserved.c index 6e41a16..c5c76e8 100644 --- a/test/reserved.c +++ b/test/reserved.c @@ -479,7 +479,7 @@ error: int main(void) { - /* This test is currently not working properly; it should be revisted + /* This test is currently not working properly; it should be revisited * when we have time. * * (Also, we should try to make this test work with all the VFDs) @@ -509,8 +509,7 @@ main(void) return num_errs; } else { - HDputs( - "All address space reservation tests skippped - Incompatible with current Virtual File Driver"); + HDputs("All address space reservation tests skipped - Incompatible with current Virtual File Driver"); } #endif /* BROKEN */ diff --git a/test/ros3.c b/test/ros3.c index c852016..03d2ae9 100644 --- a/test/ros3.c +++ b/test/ros3.c @@ -56,7 +56,7 @@ * 4) Configurable expected-actual order in generated comparison strings. * Some prefer `VERIFY(expected, actual)`, others * `VERIFY(actual, expected)`. Provide preprocessor ifdef switch - * to satifsy both parties, assuming one paradigm per test file. + * to satisfy both parties, assuming one paradigm per test file. * (One could #undef and redefine the flag through the file as desired, * but _why_.) * @@ -1444,7 +1444,7 @@ test_noops_and_autofails(void) JSVERIFY(SUCCEED, H5FDlock(file, FALSE), NULL) JSVERIFY(SUCCEED, H5FDunlock(file), NULL) /* Lock/unlock with null file or similar error crashes tests. - * HDassert in calling heirarchy, `H5FD[un]lock()` and `H5FD_[un]lock()` + * HDassert in calling hierarchy, `H5FD[un]lock()` and `H5FD_[un]lock()` */ /************ @@ -1760,7 +1760,7 @@ main(void) bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL"); if (bucket_url_env == NULL || bucket_url_env[0] == '\0') { - HDprintf("WARNING: S3 bucket url is not defined in enviornment " + HDprintf("WARNING: S3 bucket url is not defined in environment " "variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n"); } else { diff --git a/test/s3comms.c b/test/s3comms.c index 1cc47b7..ac768d6 100644 --- a/test/s3comms.c +++ b/test/s3comms.c @@ -47,7 +47,7 @@ * 4) Configurable expected-actual order in generated comparison strings. * Some prefer `VERIFY(expected, actual)`, others * `VERIFY(actual, expected)`. Provide preprocessor ifdef switch - * to satifsy both parties, assuming one paradigm per test file. + * to satisfy both parties, assuming one paradigm per test file. * (One could #undef and redefine the flag through the file as desired, * but _why_.) * Provided as courtesy, per consideration for inclusion in the library @@ -232,7 +232,7 @@ } #ifdef JSVERIFY_EXP_ACT -/* VERIFY rountines with paramter order (<expected>, <actual> [, <msg> ]) +/* VERIFY routines with parameter order (<expected>, <actual> [, <msg> ]) */ /*---------------------------------------------------------------------------- @@ -423,7 +423,7 @@ error: * * Purpose: * - * Demonstrate the construction of a Canoncial Request (and Signed Headers) + * Demonstrate the construction of a Canonical Request (and Signed Headers) * * Elided / not yet implemented: * Query strings @@ -718,7 +718,7 @@ test_hrb_init_request(void) FALSE, }, { - "null verb substitues to GET", + "null verb substitutes to GET", NULL, "/MYPATH/MYFILE.tiff", "/MYPATH/MYFILE.tiff", @@ -1131,10 +1131,10 @@ test_hrb_node_set(void) *********/ for (mock_i = 0; test->given[mock_i] != NULL; mock_i += 2) { - const char *name = test->given[mock_i]; - const char *valu = test->given[mock_i + 1]; + const char *name = test->given[mock_i]; + const char *value = test->given[mock_i + 1]; - FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, valu)) + FAIL_IF(SUCCEED != H5FD_s3comms_hrb_node_set(&list, name, value)) } /******** * TEST * @@ -1150,11 +1150,11 @@ test_hrb_node_set(void) node = list; mock_i = 0; while (test->expected[mock_i] != NULL && node != NULL) { - const char *name = test->expected[mock_i]; - const char *valu = test->expected[mock_i + 1]; + const char *name = test->expected[mock_i]; + const char *value = test->expected[mock_i + 1]; JSVERIFY_STR(name, node->name, NULL) - JSVERIFY_STR(valu, node->value, NULL) + JSVERIFY_STR(value, node->value, NULL) mock_i += 2; node = node->next; @@ -1362,7 +1362,7 @@ test_nlowercase(void) HDfree(dest); } /* end for each testcase */ - JSVERIFY(FAIL, H5FD_s3comms_nlowercase(NULL, cases[0].in, cases[0].len), "null distination should fail") + JSVERIFY(FAIL, H5FD_s3comms_nlowercase(NULL, cases[0].in, cases[0].len), "null destination should fail") PASSED(); return 0; @@ -1547,7 +1547,7 @@ test_parse_url(void) NULL, "a=b&d=b", }, - "QUERY with implict PATH", + "QUERY with implicit PATH", }, { "http://[5]/path?a=b&d=b", @@ -2002,7 +2002,7 @@ error: * H5FD_s3comms_s3r_read << called by getsize(), multiple times working * H5FD_s3comms_s3r_close * - * Shows most basic curl interation. + * Shows most basic curl iteration. * * Programmer: Jacob Smith * 2017-10-06 @@ -2598,7 +2598,7 @@ main(void) bucket_url_env = HDgetenv("HDF5_ROS3_TEST_BUCKET_URL"); if (bucket_url_env == NULL || bucket_url_env[0] == '\0') { - HDprintf("WARNING: S3 bucket url is not defined in enviornment " + HDprintf("WARNING: S3 bucket url is not defined in environment " "variable 'HDF5_ROS3_TEST_BUCKET_URL'!\n"); } else { @@ -2606,7 +2606,7 @@ main(void) s3_test_bucket_defined = TRUE; } - /* tests ordered rougly by dependence */ + /* tests ordered roughly by dependence */ nerrors += test_macro_format_credential() < 0 ? 1 : 0; nerrors += test_trim() < 0 ? 1 : 0; nerrors += test_nlowercase() < 0 ? 1 : 0; diff --git a/test/set_extent.c b/test/set_extent.c index dca41c1..daa9c18 100644 --- a/test/set_extent.c +++ b/test/set_extent.c @@ -115,12 +115,12 @@ main(void) unsigned chunk_cache; /* Whether to enable chunk caching */ int nerrors = 0; const char *env_h5_drvr; /* File Driver value from environment */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ env_h5_drvr = HDgetenv("HDF5_DRIVER"); if (env_h5_drvr == NULL) env_h5_drvr = "nomatch"; - /* Current VFD that does not support contigous address space */ + /* Current VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0); /* Initialize random number seed */ @@ -649,7 +649,7 @@ test_rank1(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f HDprintf("\n"); #endif - /* compare the read array with the shrinked array */ + /* compare the read array with the shrunk array */ for (i = 0; i < (int)dims_r[0]; i++) if (buf_s[i] != buf_o[i]) { HDprintf("buf_s[%d] = %d\n", i, buf_s[i]); @@ -1055,7 +1055,7 @@ test_rank2(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f } #endif - /* compare the read array with the shrinked array */ + /* compare the read array with the shrunk array */ for (i = 0; i < (int)dims_r[0]; i++) { for (j = 0; j < (int)dims_r[1]; j++) { if (buf_s[i][j] != buf_o[i][j]) { @@ -1536,7 +1536,7 @@ test_rank3(hid_t fapl, hid_t dcpl, hbool_t do_fill_value, hbool_t disable_edge_f HDprintf("\n"); #endif - /* compare the read array with the shrinked array */ + /* compare the read array with the shrunk array */ for (i = 0; i < (int)dims_r[0]; i++) { for (j = 0; j < (int)dims_r[1]; j++) { for (k = 0; k < (int)dims_r[2]; k++) { @@ -1734,7 +1734,7 @@ test_external(hid_t fapl) hsize_t dims_r[RANK2]; /* read dimensions */ hsize_t maxdims[RANK2] = {DIME0, DIM1}; /* only the first dimension can be extendible */ int buf_o[DIM0][DIM1]; /* original buffer, for writing */ - int buf_s[DIMS0][DIMS1]; /* shrinked buffer, for reading */ + int buf_s[DIMS0][DIMS1]; /* shrunk buffer, for reading */ int buf_e[DIME0][DIM1]; /* extended buffer, for writing, dimension 1 is the original */ int buf_ro[DIM0][DIM1]; /* original buffer for reading */ int i, j; @@ -1946,7 +1946,7 @@ test_external(hid_t fapl) } #endif - /* compare the read array with the shrinked array */ + /* compare the read array with the shrunk array */ for (i = 0; i < (int)dims_r[0]; i++) { for (j = 0; j < (int)dims_r[1]; j++) { if (buf_s[i][j] != buf_o[i][j]) { diff --git a/test/stab.c b/test/stab.c index 50c3a3f..9bdde1a 100644 --- a/test/stab.c +++ b/test/stab.c @@ -125,7 +125,7 @@ test_misc(hid_t fcpl, hid_t fapl, hbool_t new_format) if (H5Gclose(g3) < 0) TEST_ERROR - /* Open all groups with absolute names to check for exsistence */ + /* Open all groups with absolute names to check for existence */ if ((g1 = H5Gopen2(fid, "/test_1a", H5P_DEFAULT)) < 0) TEST_ERROR if ((g2 = H5Gopen2(fid, "/test_1a/sub_1", H5P_DEFAULT)) < 0) @@ -1416,7 +1416,7 @@ main(void) hid_t fcpl, fcpl2; /* File creation property list ID */ unsigned new_format; /* Whether to use the new format or not */ const char *env_h5_drvr; /* File Driver value from environment */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ int nerrors = 0; /* Get the VFD to use */ @@ -1424,7 +1424,7 @@ main(void) if (env_h5_drvr == NULL) env_h5_drvr = "nomatch"; - /* VFD that does not support contigous address space */ + /* VFD that does not support contiguous address space */ contig_addr_vfd = (hbool_t)(HDstrcmp(env_h5_drvr, "split") != 0 && HDstrcmp(env_h5_drvr, "multi") != 0); /* Reset library */ diff --git a/test/swmr.c b/test/swmr.c index 7dd5e75..f290ffc 100644 --- a/test/swmr.c +++ b/test/swmr.c @@ -4037,7 +4037,7 @@ test_append_flush_dataset_chunked(hid_t in_fapl) { hid_t fid = -1; /* file ID */ hid_t fapl = -1; /* A copy of file access property */ - hid_t did1 = -1, did2 = -1; /* The datset ID */ + hid_t did1 = -1, did2 = -1; /* The dataset ID */ hid_t sid = -1; /* The dataspace ID */ hid_t dcpl = -1; /* A copy of dataset creation property */ hid_t dapl = -1; /* A copy of dataset access property */ @@ -4260,7 +4260,7 @@ test_append_flush_dataset_fixed(hid_t in_fapl) { hid_t fid = -1; /* file ID */ hid_t fapl = -1; /* A copy of file access property */ - hid_t did1 = -1, did2 = -1; /* The datset ID */ + hid_t did1 = -1, did2 = -1; /* The dataset ID */ hid_t sid = -1; /* The dataspace ID */ hid_t dapl = -1; /* A copy of dataset access property */ hid_t ddapl = -1; /* The dataset access property of the opened dataset */ @@ -4478,7 +4478,7 @@ test_append_flush_dataset_multiple(hid_t in_fapl) { hid_t fid = -1; /* file ID */ hid_t fapl = -1; /* A copy of file access property */ - hid_t did1 = -1, did2 = -1; /* The datset ID */ + hid_t did1 = -1, did2 = -1; /* The dataset ID */ hid_t sid = -1; /* The dataspace ID */ hid_t dcpl = -1; /* A copy of dataset creation property */ hid_t dapl1 = -1; /* A copy of dataset access property */ @@ -7785,7 +7785,7 @@ main(void) /* * Modify the following routines to test for files: * H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format) - * --both result in v3 superblock and latest version suppport + * --both result in v3 superblock and latest version support */ nerrors += test_start_swmr_write(fapl, TRUE); nerrors += test_start_swmr_write(fapl, FALSE); @@ -7820,7 +7820,7 @@ main(void) * * Modify the following 2 routines to test for files: * H5Fcreate(write, latest format) or H5Fcreate(SWMR write, non-latest-format) - * --both result in v3 superblock and latest version suppport + * --both result in v3 superblock and latest version support */ nerrors += test_file_lock_swmr_same(fapl); nerrors += test_file_lock_swmr_concur(fapl); diff --git a/test/swmr_common.c b/test/swmr_common.c index 2201427..2665969 100644 --- a/test/swmr_common.c +++ b/test/swmr_common.c @@ -187,7 +187,7 @@ generate_name(char *name_buf, unsigned level, unsigned count) /*------------------------------------------------------------------------- * Function: generate_symbols * - * Purpose: Initializes the global dataset infomration arrays. + * Purpose: Initializes the global dataset information arrays. * * Parameters: N/A * diff --git a/test/swmr_generator.c b/test/swmr_generator.c index a455f3f..93cfa0f 100644 --- a/test/swmr_generator.c +++ b/test/swmr_generator.c @@ -221,7 +221,7 @@ gen_skeleton(const char *filename, hbool_t verbose, hbool_t swmr_write, int comp if (verbose) HDfprintf(stderr, "Closing objects\n"); - /* Close everythign */ + /* Close everything */ if (H5Pclose(dcpl) < 0) return -1; if (H5Sclose(sid) < 0) diff --git a/test/swmr_reader.c b/test/swmr_reader.c index 0566613..3369562 100644 --- a/test/swmr_reader.c +++ b/test/swmr_reader.c @@ -275,7 +275,7 @@ read_records(const char *filename, hbool_t verbose, FILE *verbose_file, unsigned if ((fapl = h5_fileaccess()) < 0) return -1; - /* Log I/O when verbose output it enbabled */ + /* Log I/O when verbose output it enabled */ if (use_log_vfd) { char verbose_name[1024]; diff --git a/test/swmr_sparse_writer.c b/test/swmr_sparse_writer.c index 96eff6c..0c8e391 100644 --- a/test/swmr_sparse_writer.c +++ b/test/swmr_sparse_writer.c @@ -149,7 +149,7 @@ add_records(hid_t fid, unsigned verbose, unsigned long nrecords, unsigned long f symbol_t record; /* The record to add to the dataset */ unsigned long rec_to_flush; /* # of records left to write before flush */ #ifdef OUT - volatile int dummy; /* Dummy varialbe for busy sleep */ + volatile int dummy; /* Dummy variable for busy sleep */ #endif /* OUT */ hsize_t dim[2] = {1, 0}; /* Dataspace dimensions */ unsigned long u, v; /* Local index variables */ diff --git a/test/swmr_start_write.c b/test/swmr_start_write.c index 38d6aa4..54ea363 100644 --- a/test/swmr_start_write.c +++ b/test/swmr_start_write.c @@ -125,7 +125,7 @@ create_file(const char *filename, hbool_t verbose, FILE *verbose_file, unsigned * * Parameters: * fid: file ID for the SWMR test file - * comp_level: the compresssion level + * comp_level: the compression level * index_type: The chunk index type (b1 | b2 | ea | fa) * verbose: whether verbose console output is desired. * verbose_file: file pointer for verbose output diff --git a/test/tattr.c b/test/tattr.c index ab33881..8ca959b 100644 --- a/test/tattr.c +++ b/test/tattr.c @@ -5847,7 +5847,7 @@ test_attr_info_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl) is_dense = H5O__is_attr_dense_test(my_dataset); VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test"); - /* Check for query on non-existant attribute */ + /* Check for query on non-existent attribute */ ret = H5Aget_info_by_idx(my_dataset, ".", H5_INDEX_CRT_ORDER, H5_ITER_INC, (hsize_t)0, &ainfo, H5P_DEFAULT); VERIFY(ret, FAIL, "H5Aget_info_by_idx"); @@ -6109,7 +6109,7 @@ test_attr_delete_by_idx(hbool_t new_format, hid_t fcpl, hid_t fapl) is_dense = H5O__is_attr_dense_test(my_dataset); VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test"); - /* Check for deleting non-existant attribute */ + /* Check for deleting non-existent attribute */ ret = H5Adelete_by_idx(my_dataset, ".", idx_type, order, (hsize_t)0, H5P_DEFAULT); VERIFY(ret, FAIL, "H5Adelete_by_idx"); @@ -7761,7 +7761,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl) is_dense = H5O__is_attr_dense_test(my_dataset); VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test"); - /* Check for opening a non-existant attribute on an object with no attributes */ + /* Check for opening a non-existent attribute on an object with no attributes */ ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT); VERIFY(ret_id, FAIL, "H5Aopen"); @@ -7800,7 +7800,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl) is_dense = H5O__is_attr_dense_test(my_dataset); VERIFY(is_dense, FALSE, "H5O__is_attr_dense_test"); - /* Check for opening a non-existant attribute on an object with compact attribute storage */ + /* Check for opening a non-existent attribute on an object with compact attribute storage */ ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT); VERIFY(ret_id, FAIL, "H5Aopen"); @@ -7881,7 +7881,7 @@ test_attr_open_by_name(hbool_t new_format, hid_t fcpl, hid_t fapl) VERIFY(name_count, (max_compact * 2), "H5O__attr_dense_info_test"); } /* end if */ - /* Check for opening a non-existant attribute on an object with dense attribute storage */ + /* Check for opening a non-existent attribute on an object with dense attribute storage */ ret_id = H5Aopen(my_dataset, "foo", H5P_DEFAULT); VERIFY(ret_id, FAIL, "H5Aopen"); @@ -10743,7 +10743,7 @@ test_attr_bug8(hid_t fcpl, hid_t fapl) ** chunk. Do this twice: once with only attributes in the ** object header chunk and once with a (small) soft link in ** the chunk as well. In both cases, the shrunk chunk will -** initally be too small and a new NULL message must be +** initially be too small and a new NULL message must be ** created. ** ****************************************************************/ diff --git a/test/tcoords.c b/test/tcoords.c index 7273677..37c1c1a 100644 --- a/test/tcoords.c +++ b/test/tcoords.c @@ -17,7 +17,7 @@ * * Test the element coordinates for dataspace selection. For * chunked dataset, when the hyperslab selection of some - * dimensions is full, the library optimize it by "flattenning" + * dimensions is full, the library optimize it by "flattening" * the fully selected dimensions. This program tests if the * coordinates of selected elements are correctly calculated. * diff --git a/test/testmeta.c b/test/testmeta.c index c59c6cb..6c67f06 100644 --- a/test/testmeta.c +++ b/test/testmeta.c @@ -15,7 +15,7 @@ * This program illustrates assertion errors when linked * to HDF5 1.4.1-post2 or 1.4.2-pre3 debug library. * - * If the assertion errors are ignored, the program eventially causes + * If the assertion errors are ignored, the program eventually causes * an error in H5Gcreate2 when writing object 83381. * * When writing in single file mode, the assertion errors still occur diff --git a/test/tfile.c b/test/tfile.c index ee8ca02..329d9a5 100644 --- a/test/tfile.c +++ b/test/tfile.c @@ -224,7 +224,7 @@ test_file_create(void) /* First ensure the file does not exist */ HDremove(FILE1); - /* Try opening a non-existant file */ + /* Try opening a non-existent file */ fid1 = H5Fopen(FILE1, H5F_ACC_RDWR, H5P_DEFAULT); VERIFY(fid1, FAIL, "H5Fopen"); @@ -1141,7 +1141,7 @@ test_get_obj_ids(void) CHECK(oid_count, FAIL, "H5Fget_obj_count"); VERIFY(oid_count, NGROUPS + 2, "H5Fget_obj_count"); - /* Get the IDs of the left opend objects */ + /* Get the IDs of the left opened objects */ ret_count = H5Fget_obj_ids(fid, H5F_OBJ_ALL, (size_t)oid_list_size, oid_list); CHECK(ret_count, FAIL, "H5Fget_obj_ids"); VERIFY(ret_count, oid_list_size, "H5Fget_obj_count"); @@ -1610,7 +1610,7 @@ test_file_perm2(void) ** ** test_file_ishdf5(): low-level file test routine. ** This test checks whether the H5Fis_hdf5() routine is working -** correctly in variuous situations. +** correctly in various situations. ** *****************************************************************/ static void @@ -1651,7 +1651,7 @@ test_file_ishdf5(void) file = H5Fcreate(FILE1, H5F_ACC_TRUNC, fcpl, H5P_DEFAULT); CHECK(file, FAIL, "H5Fcreate"); - /* Release file-creation property list */ + /* Release file creation property list */ ret = H5Pclose(fcpl); CHECK(ret, FAIL, "H5Pclose"); @@ -1689,7 +1689,7 @@ test_file_ishdf5(void) ** ** test_file_open_dot(): low-level file test routine. ** This test checks whether opening objects with "." for a name -** works correctly in variuous situations. +** works correctly in various situations. ** *****************************************************************/ static void @@ -2057,7 +2057,7 @@ test_file_double_root_open(void) grp2_id = H5Gopen2(file2_id, "/", H5P_DEFAULT); CHECK(grp2_id, FAIL, "H5Gopen2"); - /* Note "assymetric" close order */ + /* Note "asymmetric" close order */ ret = H5Gclose(grp1_id); CHECK(ret, FAIL, "H5Gclose"); ret = H5Gclose(grp2_id); @@ -2096,7 +2096,7 @@ test_file_double_group_open(void) grp2_id = H5Gopen2(file2_id, GRP_NAME, H5P_DEFAULT); CHECK(grp2_id, FAIL, "H5Gopen2"); - /* Note "assymetric" close order */ + /* Note "asymmetric" close order */ ret = H5Gclose(grp1_id); CHECK(ret, FAIL, "H5Gclose"); ret = H5Gclose(grp2_id); @@ -2145,7 +2145,7 @@ test_file_double_dataset_open(void) ret = H5Sclose(space_id); CHECK(ret, FAIL, "H5Sclose"); - /* Note "assymetric" close order */ + /* Note "asymmetric" close order */ ret = H5Dclose(dset1_id); CHECK(ret, FAIL, "H5Dclose"); ret = H5Dclose(dset2_id); @@ -2546,7 +2546,7 @@ test_file_double_datatype_open(void) type2_id = H5Topen2(file2_id, TYPE_NAME, H5P_DEFAULT); CHECK(type2_id, FAIL, "H5Topen2"); - /* Note "assymetric" close order */ + /* Note "asymmetric" close order */ ret = H5Tclose(type1_id); CHECK(ret, FAIL, "H5Tclose"); ret = H5Tclose(type2_id); @@ -2590,7 +2590,7 @@ test_userblock_file_size(void) ret = H5Pset_userblock(fcpl2_id, USERBLOCK_SIZE); CHECK(ret, FAIL, "H5Pset_userblock"); - /* Create files. Onyl file2 with have a userblock. */ + /* Create files. Only file2 with have a userblock. */ file1_id = H5Fcreate(FILE1, H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); CHECK(file1_id, FAIL, "H5Fcreate"); file2_id = H5Fcreate(FILE2, H5F_ACC_TRUNC, fcpl2_id, H5P_DEFAULT); @@ -3502,7 +3502,7 @@ test_userblock_alignment_paged(void) * Strategy is H5F_FILE_SPACE_NONE; fsp_size = 1024 * H5Pset_alignment() is 16 * Outcome: - * Should succed: + * Should succeed: * userblock (512) is integral multiple of alignment (16) */ /* Create file creation property list with user block */ @@ -3541,7 +3541,7 @@ test_userblock_alignment_paged(void) * H5Pset_alignment() is 3 * Reopen the file; H5Pset_alignment() is 1024 * Outcome: - * Should succed: + * Should succeed: * Userblock (512) is the same as alignment (512); * The H5Pset_alignment() calls have no effect */ @@ -3610,7 +3610,7 @@ test_filespace_info(const char *env_h5_drvr) hsize_t fs_threshold; /* Free-space section threshold--iteration variable */ hsize_t fsp_size; /* File space page size */ char filename[FILENAME_LEN]; /* Filename to use */ - hbool_t contig_addr_vfd; /* Whether VFD used has a contigous address space */ + hbool_t contig_addr_vfd; /* Whether VFD used has a contiguous address space */ herr_t ret; /* Return value */ /* Output message about test being performed */ @@ -4852,7 +4852,7 @@ test_filespace_1_10_0_compatible(void) ** Verify that the trunk can open, read and modify these files-- ** 1) They are initially created (via gen_filespace.c) in the trunk ** with combinations of file space strategies, default/non-default -** threshold, and file spacing paging enabled/disbled. +** threshold, and file spacing paging enabled/disabled. ** The library creates the file space info message with ** "mark if unknown" in these files. ** 2) They are copied to the 1.8 branch, and are opened/read/modified @@ -6000,7 +6000,7 @@ test_libver_bounds_dataset(hid_t fapl) VERIFY(dset->shared->dcpl_cache.fill.version, H5O_fill_ver_bounds[low], "H5O_fill_ver_bounds"); } - /* Verify filter pipleline message version */ + /* Verify filter pipeline message version */ VERIFY(dset->shared->dcpl_cache.pline.version, H5O_pline_ver_bounds[low], "H5O_pline_ver_bounds"); /* Close the dataset */ @@ -6053,7 +6053,7 @@ test_libver_bounds_dataset(hid_t fapl) ret = H5Sclose(sid); CHECK(ret, FAIL, "H5Sclose"); - /* Close the datset creation property list */ + /* Close the dataset creation property list */ ret = H5Pclose(dcpl); CHECK(ret, FAIL, "H5Pclose"); @@ -6073,7 +6073,7 @@ test_libver_bounds_dataset(hid_t fapl) /* Loop through all the combinations of low/high bounds in new_fapl */ /* Open the file with the fapl and create the chunked dataset */ - /* Verify the dataset's layout, fill value and filter pipleline message versions */ + /* Verify the dataset's layout, fill value and filter pipeline message versions */ for (low = H5F_LIBVER_EARLIEST; low < H5F_LIBVER_NBOUNDS; low++) { for (high = H5F_LIBVER_EARLIEST; high < H5F_LIBVER_NBOUNDS; high++) { H5E_BEGIN_TRY @@ -6178,7 +6178,7 @@ test_libver_bounds_dataset(hid_t fapl) ** new fapl: ** --Open the same file in (a) with the fapl ** --Create a chunked dataset, a compact dataset and -** a contigous dataset +** a contiguous dataset ** --Verify the dataspace message version for these ** three datasets ** --Delete the three datasets and the dataspaces @@ -6527,7 +6527,7 @@ static void test_libver_bounds_datatype_check(hid_t fapl, hid_t tid) { hid_t fid = H5I_INVALID_HID; /* File ID */ - hid_t new_fapl = H5I_INVALID_HID; /* File acess property list */ + hid_t new_fapl = H5I_INVALID_HID; /* File access property list */ hid_t dcpl = H5I_INVALID_HID; /* Dataset creation property list */ hid_t dtid = H5I_INVALID_HID; /* Datatype ID for the dataset */ hid_t str_tid = H5I_INVALID_HID; /* String datatype ID */ @@ -6787,7 +6787,7 @@ test_libver_bounds_attributes(hid_t fapl) fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, H5P_DEFAULT, fapl); CHECK(fid, H5I_INVALID_HID, "H5Fcreate"); - /* Integer datatpye */ + /* Integer datatype */ tid = H5Tcopy(H5T_NATIVE_INT); CHECK(tid, H5I_INVALID_HID, "H5Tcopy"); @@ -6896,7 +6896,7 @@ test_libver_bounds_attributes(hid_t fapl) fid = H5Fcreate(FILE8, H5F_ACC_TRUNC, fcpl, fapl); CHECK(fid, H5I_INVALID_HID, "H5Fcreate"); - /* Create an integer datatye */ + /* Create an integer datatype */ tid = H5Tcopy(H5T_NATIVE_INT); CHECK(tid, H5I_INVALID_HID, "H5Tcopy"); @@ -7691,7 +7691,7 @@ test_file(void) test_libver_macros(); /* Test the macros for library version comparison */ test_libver_macros2(); /* Test the macros for library version comparison */ test_incr_filesize(); /* Test H5Fincrement_filesize() and H5Fget_eoa() */ - test_min_dset_ohdr(); /* Test datset object header minimization */ + test_min_dset_ohdr(); /* Test dataset object header minimization */ #ifndef H5_NO_DEPRECATED_SYMBOLS test_deprec(); /* Test deprecated routines */ #endif /* H5_NO_DEPRECATED_SYMBOLS */ diff --git a/test/tgenprop.c b/test/tgenprop.c index 719b928..0a5e569 100644 --- a/test/tgenprop.c +++ b/test/tgenprop.c @@ -107,7 +107,7 @@ test_genprop_basic_class(void) ret = H5Pequal(cid2, H5P_ROOT); VERIFY(ret, 1, "H5Pequal"); - /* Make certain false postives aren't being returned */ + /* Make certain false positives aren't being returned */ ret = H5Pequal(cid2, H5P_FILE_CREATE); VERIFY(ret, 0, "H5Pequal"); @@ -185,7 +185,7 @@ test_genprop_basic_class_prop(void) CHECK_I(ret, "H5Pget_nprops"); VERIFY(nprops, 0, "H5Pget_nprops"); - /* Check the existance of the first property (should fail) */ + /* Check the existence of the first property (should fail) */ ret = H5Pexist(cid1, PROP1_NAME); VERIFY(ret, 0, "H5Pexist"); @@ -199,7 +199,7 @@ test_genprop_basic_class_prop(void) H5Pregister2(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL); VERIFY(ret, FAIL, "H5Pregister2"); - /* Check the existance of the first property */ + /* Check the existence of the first property */ ret = H5Pexist(cid1, PROP1_NAME); VERIFY(ret, 1, "H5Pexist"); @@ -223,7 +223,7 @@ test_genprop_basic_class_prop(void) H5Pregister2(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL); VERIFY(ret, FAIL, "H5Pregister2"); - /* Check the existance of the second property */ + /* Check the existence of the second property */ ret = H5Pexist(cid1, PROP2_NAME); VERIFY(ret, 1, "H5Pexist"); @@ -242,7 +242,7 @@ test_genprop_basic_class_prop(void) H5Pregister2(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL, NULL); CHECK_I(ret, "H5Pregister2"); - /* Check the existance of the third property */ + /* Check the existence of the third property */ ret = H5Pexist(cid1, PROP3_NAME); VERIFY(ret, 1, "H5Pexist"); @@ -1942,7 +1942,7 @@ test_genprop_deprec_class(void) CHECK_I(ret, "H5Pget_nprops"); VERIFY(nprops, 0, "H5Pget_nprops"); - /* Check the existance of the first property (should fail) */ + /* Check the existence of the first property (should fail) */ ret = H5Pexist(cid1, PROP1_NAME); VERIFY(ret, 0, "H5Pexist"); @@ -1954,7 +1954,7 @@ test_genprop_deprec_class(void) ret = H5Pregister1(cid1, PROP1_NAME, PROP1_SIZE, PROP1_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL); VERIFY(ret, FAIL, "H5Pregister1"); - /* Check the existance of the first property */ + /* Check the existence of the first property */ ret = H5Pexist(cid1, PROP1_NAME); VERIFY(ret, 1, "H5Pexist"); @@ -1976,7 +1976,7 @@ test_genprop_deprec_class(void) ret = H5Pregister1(cid1, PROP2_NAME, PROP2_SIZE, PROP2_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL); VERIFY(ret, FAIL, "H5Pregister1"); - /* Check the existance of the second property */ + /* Check the existence of the second property */ ret = H5Pexist(cid1, PROP2_NAME); VERIFY(ret, 1, "H5Pexist"); @@ -1994,7 +1994,7 @@ test_genprop_deprec_class(void) ret = H5Pregister1(cid1, PROP3_NAME, PROP3_SIZE, PROP3_DEF_VALUE, NULL, NULL, NULL, NULL, NULL, NULL); CHECK_I(ret, "H5Pregister1"); - /* Check the existance of the third property */ + /* Check the existence of the third property */ ret = H5Pexist(cid1, PROP3_NAME); VERIFY(ret, 1, "H5Pexist"); diff --git a/test/th5o.c b/test/th5o.c index f1790d0..31b6782 100644 --- a/test/th5o.c +++ b/test/th5o.c @@ -466,7 +466,7 @@ test_h5o_refcount(void) CHECK(ret, FAIL, "H5Oget_info_by_name"); VERIFY(oinfo.rc, 1, "reference count in H5Oget_info_by_name"); - /* Increment the reference counts and then close the file to make sure the increment is permanant */ + /* Increment the reference counts and then close the file to make sure the increment is permanent */ ret = H5Oincr_refcount(grp); CHECK(ret, FAIL, "H5Oincr_refcount"); ret = H5Oincr_refcount(dtype); diff --git a/test/th5s.c b/test/th5s.c index 8144504..8d414c9 100644 --- a/test/th5s.c +++ b/test/th5s.c @@ -1104,7 +1104,7 @@ test_h5s_zero_dim(void) dset1 = H5Dopen2(fid1, BASICDATASET1, H5P_DEFAULT); CHECK(dset1, FAIL, "H5Dopen2"); - /* Get the space of the dataset and querry it */ + /* Get the space of the dataset and query it */ sid1 = H5Dget_space(dset1); CHECK(sid1, FAIL, "H5Dget_space"); @@ -41,7 +41,7 @@ basic_id_test(void) int num_ref; hsize_t num_members; - /* Try to register an ID with ficticious types */ + /* Try to register an ID with fictitious types */ H5E_BEGIN_TRY arrayID = H5Iregister((H5I_type_t)420, testObj); H5E_END_TRY @@ -58,7 +58,7 @@ basic_id_test(void) if (arrayID != H5I_INVALID_HID) goto out; - /* Try to access IDs with ficticious types */ + /* Try to access IDs with fictitious types */ H5E_BEGIN_TRY testPtr = H5Iobject_verify((hid_t)100, (H5I_type_t)0); H5E_END_TRY diff --git a/test/timer.c b/test/timer.c index ea9787a..2de7980 100644 --- a/test/timer.c +++ b/test/timer.c @@ -5,12 +5,10 @@ * * * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * - * the files COPYING and Copyright.html. COPYING can be found at the root * - * of the source code distribution tree; Copyright.html can be found at the * - * root level of an installed copy of the electronic HDF5 document set and * - * is linked from the top-level documents page. It can also be found at * - * http://hdfgroup.org/HDF5/doc/Copyright.html. If you do not have * - * access to either file, you may request a copy from help@hdfgroup.org. * + * the COPYING file, which can be found at the root of the source code * + * distribution tree, or in https://www.hdfgroup.org/licenses. * + * If you do not have access to either file, you may request a copy from * + * help@hdfgroup.org. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ /* diff --git a/test/tmisc.c b/test/tmisc.c index 6d813dd..da39318 100644 --- a/test/tmisc.c +++ b/test/tmisc.c @@ -1280,7 +1280,7 @@ test_misc8(void) for (v = 0; v < MISC8_DIM1; v++) *tdata++ = (int)(((u * MISC8_DIM1) + v) % 13); - /* Create a file acccess property list */ + /* Create a file access property list */ fapl = H5Pcreate(H5P_FILE_ACCESS); CHECK(fapl, FAIL, "H5Pcreate"); @@ -3802,7 +3802,7 @@ test_misc21(void) /**************************************************************** ** -** test_misc22(): Test SZIP bits-per-pixel paramter. +** test_misc22(): Test SZIP bits-per-pixel parameter. ** This should be set according to the datatype. ** Tests for precision and offset combo's. ** diff --git a/test/trefer.c b/test/trefer.c index 5b7a3cf..2a13bbc 100644 --- a/test/trefer.c +++ b/test/trefer.c @@ -1588,14 +1588,14 @@ test_reference_sel_none(void) hid_t sid = H5I_INVALID_HID; /* Dataspace ID */ hid_t ref_sid = H5I_INVALID_HID; /* Dataspace ID for referenced */ hid_t did = H5I_INVALID_HID; /* Dataset ID */ - hsize_t dims[1] = {SPACE1_DIM1}; /* Dimension siez ofr dataset */ + hsize_t dims[1] = {SPACE1_DIM1}; /* Dimension size of dataset */ int buf[SPACE1_DIM1]; /* Buffer for data */ hdset_reg_ref_t ref[1]; /* Buffer for reference */ hssize_t npoints; /* Number of points in selection */ herr_t status; /* Return status */ int i; /* Local index variable */ - /* Iniitialize buffer */ + /* Initialize buffer */ for (i = 0; i < SPACE1_DIM1; i++) buf[i] = i; diff --git a/test/tselect.c b/test/tselect.c index 6a6a6cf..2fd7082 100644 --- a/test/tselect.c +++ b/test/tselect.c @@ -1879,7 +1879,7 @@ verify_select_hyper_contig_dr__run_test(const uint16_t *cube_buf, size_t H5_ATTR uint16_t expected_value; /* Expected value in dataset */ unsigned i, j, k, l, m; /* Local index variables */ size_t s; /* Local index variable */ - hbool_t mis_match; /* Flag to indicate mis-match in expected value */ + hbool_t mis_match; /* Flag to indicate mismatch in expected value */ HDassert(cube_buf); HDassert(cube_size > 0); @@ -2357,7 +2357,7 @@ test_select_hyper_contig_dr__run_test(int test_num, const uint16_t *cube_buf, co /* Now write the contents of the in memory small cube to slices of * the on disk cube. After each write, read the on disk cube - * into memeory, and verify that it contains the expected + * into memory, and verify that it contains the expected * data. Verify that H5Sselect_shape_same() returns true on * the memory and file selections. */ @@ -2742,7 +2742,7 @@ test_select_hyper_checker_board_dr__select_checker_board(hid_t tgt_n_cube_sid, u i++; } while ((i <= 1) && (0 >= sel_offset)); - /* Wierdness alert: + /* Weirdness alert: * * Some how, it seems that selections can extend beyond the * boundaries of the target dataspace -- hence the following @@ -3401,7 +3401,7 @@ test_select_hyper_checker_board_dr__run_test(int test_num, const uint16_t *cube_ /* Now write checker board selections of the entries in memory * small cube to slices of the on disk cube. After each write, - * read the on disk large cube * into memeory, and verify that + * read the on disk large cube * into memory, and verify that * it contains the expected * data. Verify that * H5Sselect_shape_same() returns true on the memory and file * selections. @@ -5256,7 +5256,7 @@ test_select_hyper_union_stagger(void) error = H5Fclose(file_id); CHECK(error, FAIL, "H5Fclose"); - /* Initialize intput buffer */ + /* Initialize input buffer */ HDmemset(data_out, 0, 7 * 7 * sizeof(int)); /* Open file */ @@ -6785,7 +6785,7 @@ test_select_point_chunk(void) /* Verify data (later) */ - /* Close everything (inclusing selections) */ + /* Close everything (including selections) */ ret = H5Sclose(pnt1_space); CHECK(ret, FAIL, "H5Sclose"); ret = H5Sclose(pnt2_space); @@ -10955,7 +10955,7 @@ test_shape_same_dr__full_space_vs_slice(int test_num, int small_rank, int large_ ** m-cube (m >= n). ** ** 2) The dimensions selected in the slice through the m-cube -** are the dimesnions with the most quickly changing +** are the dimensions with the most quickly changing ** indices. ** ****************************************************************/ @@ -11251,7 +11251,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i CHECK(ret, FAIL, "H5Sselect_hyperslab"); } /* end if */ - /* Wierdness alert: + /* Weirdness alert: * * Some how, it seems that selections can extend beyond the * boundaries of the target dataspace -- hence the following @@ -11346,7 +11346,7 @@ test_shape_same_dr__checkerboard(int test_num, int small_rank, int large_rank, i CHECK(ret, FAIL, "H5Sselect_hyperslab"); } /* end if */ - /* Wierdness alert: + /* Weirdness alert: * * Again, it seems that selections can extend beyond the * boundaries of the target dataspace -- hence the following @@ -12126,7 +12126,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ MESSAGE(7, ("Testing functionality to rebuild 2-D hyperslab selection\n")); /* Create 2-D dataspace */ @@ -12218,7 +12218,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ MESSAGE(7, ("Testing functionality to rebuild 3-D hyperslab selection\n")); @@ -12320,7 +12320,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ MESSAGE(7, ("Testing functionality to rebuild 4-D hyperslab selection\n")); @@ -12439,7 +12439,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ MESSAGE(7, ("Testing functionality to rebuild 5-D hyperslab selection\n")); @@ -12568,7 +12568,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ /* We use 5-D to test a special case with rebuilding routine TRUE, FALSE and TRUE */ @@ -12615,7 +12615,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ /* Adding some selections to make it real irregular */ start5[3] = 1; @@ -12643,7 +12643,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ /* Add more selections to make it regular again */ start5[3] = 5; @@ -12671,7 +12671,7 @@ test_space_rebuild(void) ret = FAIL; CHECK(ret, FAIL, "H5S_hyper_rebuild"); } - /* No need to do shape comparision */ + /* No need to do shape comparison */ H5Sclose(sid_reg1); CHECK(ret, FAIL, "H5Sclose"); @@ -12977,7 +12977,7 @@ test_space_update_diminfo(void) CHECK(ret, FAIL, "H5S_hyper_update_diminfo"); } /* end if */ - /* Add block parially overlapping first, with OR */ + /* Add block partially overlapping first, with OR */ start1[0] = 4; count1[0] = 1; block1[0] = 2; @@ -13022,7 +13022,7 @@ test_space_update_diminfo(void) CHECK(ret, FAIL, "H5S_hyper_update_diminfo"); } /* end if */ - /* Add equally sized block parially overlapping current, with XOR */ + /* Add equally sized block partially overlapping current, with XOR */ start1[0] = 3; count1[0] = 1; block1[0] = 5; @@ -13056,7 +13056,7 @@ test_space_update_diminfo(void) CHECK(ret, FAIL, "H5S_hyper_rebuild"); } /* end if */ - /* Add differently sized block parially overlapping current, with XOR */ + /* Add differently sized block partially overlapping current, with XOR */ start1[0] = 4; count1[0] = 1; block1[0] = 5; @@ -14232,7 +14232,7 @@ test_hyper_regular(void) ** ****************************************************************/ static void -test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enblocks, hsize_t *eblock1, +test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t endpoints, hssize_t enblocks, hsize_t *eblock1, hsize_t *eblock2) { hid_t lim_sid; @@ -14257,7 +14257,7 @@ test_hyper_unlim_check(hid_t sid, hsize_t *dims, hssize_t enpoints, hssize_t enb /* Check number of elements */ npoints = H5Sget_select_npoints(lim_sid); CHECK(npoints, FAIL, "H5Sget_select_npoints"); - VERIFY(npoints, enpoints, "H5Sget_select_npoints"); + VERIFY(npoints, endpoints, "H5Sget_select_npoints"); /* Get selection type */ sel_type = H5Sget_select_type(lim_sid); diff --git a/test/tsohm.c b/test/tsohm.c index 7ef283b..fdc3375 100644 --- a/test/tsohm.c +++ b/test/tsohm.c @@ -636,7 +636,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_ if (GetTestExpress() > 1) test_file_closing = FALSE; - /* Intialize wdata */ + /* Initialize wdata */ HDmemset(&wdata, 0, sizeof(wdata)); wdata.i1 = 11; HDstrcpy(wdata.str, "string"); @@ -649,7 +649,7 @@ size1_helper(hid_t file, const char *filename, hid_t fapl_id, hbool_t test_file_ wdata.i8 = 88; wdata.f1 = 0.0F; - /* Intialize rdata */ + /* Initialize rdata */ HDmemset(&rdata, 0, sizeof(rdata)); dtype1_id = make_dtype_1(); @@ -762,7 +762,7 @@ error: * Function: getsize_testsize1 * * Purpose: Creates a test file, populates it, and returns its file size. - * Oject header information from the "first" dataset in the file + * Object header information from the "first" dataset in the file * is stored in pointer `oinfo`. * * Programmer: Jacob Smith @@ -1551,7 +1551,7 @@ size2_dump_struct(const char *name, size2_helper_struct *sizes) /*------------------------------------------------------------------------- * Function: size2_helper * - * Purpose: A helper functon for test_sohm_size2. + * Purpose: A helper function for test_sohm_size2. * * Creates a file using the given fcpl, then creates lots * of different kinds of messages within the file and @@ -1931,7 +1931,7 @@ size2_helper(hid_t fcpl_id, int test_file_closing, size2_helper_struct *ret_size /*------------------------------------------------------------------------- * Function: size2_verify * - * Purpose: A helper functon to verify the file created by size2_helper. + * Purpose: A helper function to verify the file created by size2_helper. * * Runs various tests (not exhaustive) to ensure that the * file FILENAME actually has the structure that size2_helper diff --git a/test/ttsafe_attr_vlen.c b/test/ttsafe_attr_vlen.c index 2fe140c..d105ffe 100644 --- a/test/ttsafe_attr_vlen.c +++ b/test/ttsafe_attr_vlen.c @@ -136,7 +136,7 @@ tts_attr_vlen_thread(void H5_ATTR_UNUSED *client_data) gid = H5Gopen2(fid, "/", H5P_DEFAULT); CHECK(gid, H5I_INVALID_HID, "H5Gopen"); - /* Open the attribte */ + /* Open the attribute */ aid = H5Aopen(gid, "root_attr", H5P_DEFAULT); CHECK(aid, H5I_INVALID_HID, "H5Aopen"); diff --git a/test/ttsafe_error.c b/test/ttsafe_error.c index 7c92d65..ceaca25 100644 --- a/test/ttsafe_error.c +++ b/test/ttsafe_error.c @@ -114,7 +114,7 @@ tts_error(void) H5TS_wait_for_thread(threads[i]); if (error_flag_g) { - TestErrPrintf("At least one thread reported a value that was different from the exected value\n"); + TestErrPrintf("At least one thread reported a value that was different from the expected value\n"); HDprintf("(Update this test if the error stack changed!)\n"); } diff --git a/test/tvltypes.c b/test/tvltypes.c index 1b9bfad..9e4f685 100644 --- a/test/tvltypes.c +++ b/test/tvltypes.c @@ -131,7 +131,7 @@ test_vltypes_dataset_create(void) dcpl = H5Pcreate(H5P_DATASET_CREATE); CHECK(dcpl, FAIL, "H5Pcreate"); - /* Set fill value writting time to be NEVER */ + /* Set fill value writing time to be NEVER */ ret = H5Pset_fill_time(dcpl, H5D_FILL_TIME_NEVER); CHECK(ret, FAIL, "H5Pset_fill_time"); diff --git a/test/twriteorder.c b/test/twriteorder.c index 9cff00a..c7984e7 100644 --- a/test/twriteorder.c +++ b/test/twriteorder.c @@ -54,7 +54,7 @@ * * If the system, in which the writer and reader processes run, the readers * will always get all chain-linked blocks correctly. If the order of write - * is not maintained, some reader processes may found unexpect block data. + * is not maintained, some reader processes may found unexpected block data. * *************************************************************/ @@ -294,7 +294,7 @@ write_wo_file(void) return -1; } - /* all writes done. return succeess. */ + /* all writes done. return success. */ #ifdef DEBUG HDprintf("wrote %d blocks\n", nlinkedblock_g); #endif diff --git a/test/unlink.c b/test/unlink.c index a154671..92cfab7 100644 --- a/test/unlink.c +++ b/test/unlink.c @@ -614,7 +614,7 @@ test_filespace(hid_t fapl) /* Create common objects for datasets */ - /* Create dataset creation property list for contigous storage */ + /* Create dataset creation property list for contiguous storage */ if ((contig_dcpl = H5Pcreate(H5P_DATASET_CREATE)) < 0) TEST_ERROR @@ -1096,7 +1096,7 @@ test_filespace(hid_t fapl) PASSED(); - /* Create simple group hiearchy, remove it & verify file size */ + /* Create simple group hierarchy, remove it & verify file size */ TESTING(" simple group hierarchy"); /* Create file */ @@ -1135,7 +1135,7 @@ test_filespace(hid_t fapl) PASSED(); - /* Create complex group hiearchy, remove it & verify file size */ + /* Create complex group hierarchy, remove it & verify file size */ TESTING(" complex group hierarchy"); /* Create file */ diff --git a/test/unregister.c b/test/unregister.c index b2e1a57..9b72f2c 100644 --- a/test/unregister.c +++ b/test/unregister.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/test/use_common.c b/test/use_common.c index 8f78f3b..7b5df53 100644 --- a/test/use_common.c +++ b/test/use_common.c @@ -298,7 +298,7 @@ write_uc_file(hbool_t tosend, hid_t file_id, options_t *opts) return -1; } - /* verify chunk_dims against set paramenters */ + /* verify chunk_dims against set parameters */ if (chunk_dims[0] != opts->chunkdims[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) { HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n", (unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1], @@ -533,7 +533,7 @@ read_uc_file(hbool_t towait, options_t *opts) /* quit when all nplanes have been read */ loops_waiting_for_plane = 0; while (nplanes_seen < opts->nplanes) { - /* print progress message according to if new planes are availalbe */ + /* print progress message according to if new planes are available */ if (nplanes_seen < dims[0]) { if (loops_waiting_for_plane) { /* end the previous message */ diff --git a/test/use_disable_mdc_flushes.c b/test/use_disable_mdc_flushes.c index d1b693b..ea674ec 100644 --- a/test/use_disable_mdc_flushes.c +++ b/test/use_disable_mdc_flushes.c @@ -360,7 +360,7 @@ write_file(void) return -1; } - /* verify chunk_dims against set paramenters */ + /* verify chunk_dims against set parameters */ if (chunk_dims[0] != chunkdims_g[0] || chunk_dims[1] != cz || chunk_dims[2] != cz) { HDfprintf(stderr, "chunk size is not as expected. Got dims=(%llu,%llu,%llu)\n", (unsigned long long)chunk_dims[0], (unsigned long long)chunk_dims[1], @@ -196,7 +196,7 @@ vds_select_equal(hid_t space1, hid_t space2) if (nblocks1 != nblocks2) return FALSE; - /* Allocate block lists. Do not return directly afer + /* Allocate block lists. Do not return directly after * allocating, to make sure buffers are freed. */ if (NULL == (buf1 = (hsize_t *)HDmalloc((size_t)2 * (size_t)rank1 * (size_t)nblocks1 * sizeof(*buf1)))) @@ -479,7 +479,7 @@ test_api_get_ex_dcpl(test_api_config_t config, hid_t fapl, hid_t dcpl, hid_t *ex /* Verify examination DCPL is equal to original DCPL. Do not compare the * plist to itself, and do not do the comparison if we reopened the file, - * because in that case the extent of the source dset will not be corrent. + * because in that case the extent of the source dset will not be current. */ if ((*ex_dcpl != dcpl) && (config != TEST_API_REOPEN_FILE)) { if ((tri_ret = H5Pequal(dcpl, *ex_dcpl)) < 0) @@ -1044,7 +1044,7 @@ test_api(test_api_config_t config, hid_t fapl) if ((vspace[i] = H5Screate_simple(2, dims, NULL)) < 0) TEST_ERROR - /* Select row in virual dataspace */ + /* Select row in virtual dataspace */ start[0] = (hsize_t)i; if (H5Sselect_hyperslab(vspace[i], H5S_SELECT_SET, start, NULL, count, block) < 0) TEST_ERROR @@ -1181,7 +1181,7 @@ test_vds_prefix_first(unsigned config, hid_t vds_fapl, hid_t src_fapl) hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */ hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */ hid_t memspace = -1; /* Memory dataspace */ - hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */ + hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */ int buf[10][26]; /* Write and expected read buffer */ @@ -1460,7 +1460,7 @@ test_basic_io(unsigned config, hid_t vds_fapl, hid_t src_fapl) hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */ hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */ hid_t memspace = -1; /* Memory dataspace */ - hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */ + hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */ hsize_t start[4]; /* Hyperslab start */ @@ -3659,7 +3659,7 @@ test_unlim(unsigned config, hid_t vds_fapl, hid_t src_fapl) hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */ hid_t memspace = -1; /* Memory dataspace */ hid_t filespace = -1; /* File dataspace */ - hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */ + hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[2] = {10, 10}; /* Data space current size */ hsize_t mdims[2] = {10, 20}; /* Data space maximum size */ @@ -6673,7 +6673,7 @@ test_printf(unsigned config, hid_t vds_fapl, hid_t src_fapl) hid_t vspace[2] = {-1, -1}; /* Virtual dset dataspaces */ hid_t memspace = -1; /* Memory dataspace */ hid_t filespace = -1; /* File dataspace */ - hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datsets */ + hid_t srcdset[6] = {-1, -1, -1, -1, -1, -1}; /* Source datasets */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[2] = {10, 0}; /* Data space current size */ hsize_t mdims[2] = {10, 20}; /* Data space maximum size */ @@ -10381,7 +10381,7 @@ test_all(unsigned config, hid_t vds_fapl, hid_t src_fapl) hid_t vspace[3] = {-1, -1, -1}; /* Virtual dset dataspaces */ hid_t memspace = -1; /* Memory dataspace */ hid_t filespace = -1; /* File dataspace */ - hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datsets */ + hid_t srcdset[5] = {-1, -1, -1, -1, -1}; /* Source datasets */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[2] = {6, 6}; /* Data space current size */ hsize_t mdims[2] = {10, 10}; /* Data space maximum size */ diff --git a/test/vds_env.c b/test/vds_env.c index b8d8bbd..88b649e 100644 --- a/test/vds_env.c +++ b/test/vds_env.c @@ -58,7 +58,7 @@ test_vds_prefix_second(unsigned config, hid_t fapl) hid_t srcspace[4] = {-1, -1, -1, -1}; /* Source dataspaces */ hid_t vspace[4] = {-1, -1, -1, -1}; /* Virtual dset dataspaces */ hid_t memspace = -1; /* Memory dataspace */ - hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datsets */ + hid_t srcdset[4] = {-1, -1, -1, -1}; /* Source datasets */ hid_t vdset = -1; /* Virtual dataset */ hsize_t dims[4] = {10, 26, 0, 0}; /* Data space current size */ int buf[10][26]; /* Write and expected read buffer */ diff --git a/testpar/t_bigio.c b/testpar/t_bigio.c index 692f6e9..50aecf2 100644 --- a/testpar/t_bigio.c +++ b/testpar/t_bigio.c @@ -1170,7 +1170,7 @@ create_faccess_plist(MPI_Comm comm, MPI_Info info, int l_facc_type) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: One big singluar selection inside one chunk + * Descriptions for the selection: One big singular selection inside one chunk * Two dimensions, * * dim1 = space_dim1(5760)*mpi_size @@ -1274,7 +1274,7 @@ coll_chunk2(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = space_dim1*mpi_size @@ -1828,7 +1828,7 @@ main(int argc, char **argv) /* Set the bigio processing limit to be 'newsize' bytes */ hsize_t oldsize = H5_mpi_set_bigio_count(newsize); - /* Having set the bigio handling to a size that is managable, + /* Having set the bigio handling to a size that is manageable, * we'll set our 'bigcount' variable to be 2X that limit so * that we try to ensure that our bigio handling is actually * envoked and tested. diff --git a/testpar/t_cache.c b/testpar/t_cache.c index e119f6a..97c7a74 100644 --- a/testpar/t_cache.c +++ b/testpar/t_cache.c @@ -126,7 +126,7 @@ int total_writes = 0; * * local_pinned: Boolean flag that is set to true iff the entry * has been pinned in the local cache, but probably not all - * caches. Such pins will typically not be consistant across + * caches. Such pins will typically not be consistent across * processes, and thus cannot be marked as dirty unless they * happen to overlap some collective operation. * @@ -205,7 +205,7 @@ struct datum *data = NULL; * even divisor of NUM_DATA_ENTRIES. So far, all tests have been with * powers of 10 that meet these criteria. * - * Further, this value must be consistant across all processes. + * Further, this value must be consistent across all processes. */ #define STD_VIRT_NUM_DATA_ENTRIES NUM_DATA_ENTRIES @@ -248,11 +248,11 @@ int *data_index = NULL; * in which no acks are sent after writes. Instead, the metadata cache is * provided with a callback function to call after each sequence of writes. * This callback simply causes the client to send the server process a - * "sync" message and and await an ack in reply. + * "sync" message and await an ack in reply. * * Strangely, at least on Phoenix, the first solution runs faster by a * rather large margin. However, I can imagine this changing with - * different OS's and MPI implementatins. + * different OS's and MPI implementations. * * Thus I have left code supporting the second solution in place. * @@ -1282,7 +1282,7 @@ reset_server_counters(void) * Function: server_main() * * Purpose: Main function for the server process. This process exists - * to provide an independant view of the data array. + * to provide an independent view of the data array. * * The function handles request from the other processes in * the test until the count of done messages received equals @@ -3961,7 +3961,7 @@ setup_cache_for_test(hid_t *fid_ptr, H5F_t **file_ptr_ptr, H5C_t **cache_ptr_ptr * Purpose: Verify that the indicated entries have been written exactly * once each, and that the indicated total number of writes * has been processed by the server process. Flag an error if - * discrepency is noted. Finally reset the counters maintained + * discrepancy is noted. Finally reset the counters maintained * by the server process. * * This function should only be called by the metadata cache @@ -4049,7 +4049,7 @@ verify_writes(unsigned num_writes, haddr_t *written_entries_tbl) /* final barrier to ensure that all processes think that the server * counters have been reset before we leave the sync point. This - * barrier is probaby not necessary at this point in time (5/9/10), + * barrier is probably not necessary at this point in time (5/9/10), * but I can think of at least one likely change to the metadata write * strategies that will require it -- hence its insertion now. */ @@ -4793,7 +4793,7 @@ server_smoke_check(void) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -5140,7 +5140,7 @@ smoke_check_1(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -5304,7 +5304,7 @@ smoke_check_2(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -5452,7 +5452,7 @@ smoke_check_2(int metadata_write_strategy) * * Purpose: Third smoke check for the parallel cache. * - * Use random reads to vary the loads on the diffferent + * Use random reads to vary the loads on the different * processors. Also force different cache size adjustments. * * In this test, load process 0 heavily, and the other @@ -5509,7 +5509,7 @@ smoke_check_3(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -5737,7 +5737,7 @@ smoke_check_3(int metadata_write_strategy) * * Purpose: Fourth smoke check for the parallel cache. * - * Use random reads to vary the loads on the diffferent + * Use random reads to vary the loads on the different * processors. Also force different cache size adjustments. * * In this test, load process 0 lightly, and the other @@ -5794,7 +5794,7 @@ smoke_check_4(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -6069,7 +6069,7 @@ smoke_check_5(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -6354,7 +6354,7 @@ trace_file_check(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -6661,7 +6661,7 @@ smoke_check_6(int metadata_write_strategy) if (!server_main()) { - /* some error occured in the server -- report failure */ + /* some error occurred in the server -- report failure */ nerrors++; if (verbose) { HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, FUNC); @@ -6723,7 +6723,7 @@ smoke_check_6(int metadata_write_strategy) if (FALSE != entry_ptr->header.coll_access) { nerrors++; if (verbose) { - HDfprintf(stdout, "%d:%s: Entry inserted indepedently marked as collective.\n", + HDfprintf(stdout, "%d:%s: Entry inserted independently marked as collective.\n", world_mpi_rank, FUNC); } } @@ -6779,7 +6779,7 @@ smoke_check_6(int metadata_write_strategy) if (FALSE != entry_ptr->header.coll_access) { nerrors++; if (verbose) { - HDfprintf(stdout, "%d:%s: Entry inserted indepedently marked as collective.\n", + HDfprintf(stdout, "%d:%s: Entry inserted independently marked as collective.\n", world_mpi_rank, FUNC); } } @@ -6903,17 +6903,10 @@ main(int argc, char **argv) H5open(); express_test = do_express_test(); -#if 0 /* JRM */ - express_test = 0; -#endif /* JRM */ - if (express_test) { - + if (express_test) virt_num_data_entries = EXPRESS_VIRT_NUM_DATA_ENTRIES; - } - else { - + else virt_num_data_entries = STD_VIRT_NUM_DATA_ENTRIES; - } #ifdef H5_HAVE_MPE if (MAINPROCESS) diff --git a/testpar/t_cache_image.c b/testpar/t_cache_image.c index d8a48bb..e9a8e6c 100644 --- a/testpar/t_cache_image.c +++ b/testpar/t_cache_image.c @@ -452,7 +452,7 @@ create_data_sets(hid_t file_id, int min_dset, int max_dset) } /* set the dataset creation plist to specify that the raw data is - * to be partioned into 10X10 element chunks. + * to be partitioned into 10X10 element chunks. */ if (pass) { @@ -798,7 +798,7 @@ delete_data_sets(hid_t file_id, int min_dset, int max_dset) * Set pass to FALSE and issue a suitable failure * message if either the file contains a metadata cache image * superblock extension and mdci_sbem_expected is TRUE, or - * vise versa. + * vice versa. * * If mdci_sbem_expected is TRUE, also verify that the metadata * cache has been advised of this. @@ -810,7 +810,7 @@ delete_data_sets(hid_t file_id, int min_dset, int max_dset) * FAPL entry when opening the file, and verify that the * metadata cache is notified. * - * If config_fsm is TRUE, setup the persistant free space + * If config_fsm is TRUE, setup the persistent free space * manager. Note that this flag may only be set if * create_file is also TRUE. * @@ -887,7 +887,7 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons if (show_progress) HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass); - /* create a file access propertly list. */ + /* create a file access property list. */ if (pass) { fapl_id = H5Pcreate(H5P_FILE_ACCESS); @@ -959,7 +959,7 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons if (show_progress) HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass); - /* setup the persistant free space manager if indicated */ + /* setup the persistent free space manager if indicated */ if ((pass) && (config_fsm)) { fcpl_id = H5Pcreate(H5P_FILE_CREATE); @@ -1141,13 +1141,13 @@ open_hdf5_file(const hbool_t create_file, const hbool_t mdci_sbem_expected, cons if ((file_ptr->shared->page_buf) && ((!enable_page_buffer) || (l_facc_type == FACC_MPIO))) { pass = FALSE; - failure_mssg = "page buffer unexepectedly enabled."; + failure_mssg = "page buffer unexpectedly enabled."; } else if ((file_ptr->shared->page_buf != NULL) && ((enable_page_buffer) || (l_facc_type != FACC_MPIO))) { pass = FALSE; - failure_mssg = "page buffer unexepectedly disabled."; + failure_mssg = "page buffer unexpectedly disabled."; } } @@ -1363,7 +1363,7 @@ par_create_dataset(int dset_num, hid_t file_id, int mpi_rank, int mpi_size) HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass); /* set the dataset creation plist to specify that the raw data is - * to be partioned into 1X10X10 element chunks. + * to be partitioned into 1X10X10 element chunks. */ if (pass) { @@ -3016,7 +3016,7 @@ verify_cache_image_RO(int file_name_id, int md_write_strat, int mpi_rank) /* Verify that all other processes receive the cache image block * from process 0. * - * Since we have alread verified that only process 0 has read the + * Since we have already verified that only process 0 has read the * image, it is sufficient to verify that the image was loaded on * all processes. */ @@ -3287,7 +3287,7 @@ verify_cache_image_RW(int file_name_id, int md_write_strat, int mpi_rank) /* Verify that all other processes receive the cache image block * from process 0. * - * Since we have alread verified that only process 0 has read the + * Since we have already verified that only process 0 has read the * image, it is sufficient to verify that the image was loaded on * all processes. */ @@ -3613,7 +3613,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size) /* Verify that all other processes receive the cache image block * from process 0. * - * Since we have alread verified that only process 0 has read the + * Since we have already verified that only process 0 has read the * image, it is sufficient to verify that the image was loaded on * all processes. */ @@ -3708,7 +3708,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size) /* Verify that all other processes receive the cache image block * from process 0. * - * Since we have alread verified that only process 0 has read the + * Since we have already verified that only process 0 has read the * image, it is sufficient to verify that the image was loaded on * all processes. */ @@ -3753,9 +3753,9 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size) HDfprintf(stdout, "%s: cp = %d, pass = %d.\n", fcn_name, cp++, pass); /* 13) Get the size of the file. Verify that it is less - * than 20 KB. Without deletions and persistant free + * than 20 KB. Without deletions and persistent free * space managers, size size is about 30 MB, so this - * is sufficient to verify that the persistant free + * is sufficient to verify that the persistent free * space managers are more or less doing their job. * * Note that this test will have to change if we use @@ -3827,7 +3827,7 @@ smoke_check_1(MPI_Comm mpi_comm, MPI_Info mpi_info, int mpi_rank, int mpi_size) * Purpose: Run parallel tests on the cache image feature. * * At present, cache image is disabled in parallel, and - * thus these tests are restructed to verifying that a + * thus these tests are restricted to verifying that a * file with a cache image can be opened in the parallel * case, and verifying that instructions to create a * cache image are ignored in the parallel case. diff --git a/testpar/t_coll_chunk.c b/testpar/t_coll_chunk.c index 0801f5c..8f699ea 100644 --- a/testpar/t_coll_chunk.c +++ b/testpar/t_coll_chunk.c @@ -53,7 +53,7 @@ static void coll_chunktest(const char *filename, int chunk_factor, int select_fa */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: One big singluar selection inside one chunk + * Descriptions for the selection: One big singular selection inside one chunk * Two dimensions, * * dim1 = SPACE_DIM1(5760)*mpi_size @@ -155,7 +155,7 @@ coll_chunk2(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -209,7 +209,7 @@ coll_chunk3(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -261,7 +261,7 @@ coll_chunk4(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -315,7 +315,7 @@ coll_chunk5(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -367,7 +367,7 @@ coll_chunk6(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -419,7 +419,7 @@ coll_chunk7(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -471,7 +471,7 @@ coll_chunk8(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size @@ -523,7 +523,7 @@ coll_chunk9(void) */ /* ------------------------------------------------------------------------ - * Descriptions for the selection: one singular selection accross many chunks + * Descriptions for the selection: one singular selection across many chunks * Two dimensions, Num of chunks = 2* mpi_size * * dim1 = SPACE_DIM1*mpi_size diff --git a/testpar/t_dset.c b/testpar/t_dset.c index 2d4c70b..4fc8057 100644 --- a/testpar/t_dset.c +++ b/testpar/t_dset.c @@ -2880,7 +2880,7 @@ none_selection_chunk(void) * Simple independent I/O. This tests that the defaults are properly set. * * TEST_ACTUAL_IO_RESET: - * Perfroms collective and then independent I/O wit hthe same dxpl to + * Performs collective and then independent I/O with hthe same dxpl to * make sure the peroperty is correctly reset to the default on each use. * Specifically, this test runs TEST_ACTUAL_IO_MULTI_CHUNK_NO_OPT_MIX_DISAGREE * (The most complex case that works on all builds) and then performs @@ -2889,7 +2889,7 @@ none_selection_chunk(void) * Note: DIRECT_MULTI_CHUNK_MIX and DIRECT_MULTI_CHUNK_MIX_DISAGREE * is not needed as they are covered by DIRECT_CHUNK_MIX and * MULTI_CHUNK_MIX_DISAGREE cases. _DIRECT_ cases are only for testing - * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO insted of num-threshold. + * path way to multi-chunk-io by H5FD_MPIO_CHUNK_MULTI_IO instead of num-threshold. * * Modification: * - Refctore to remove multi-chunk-without-opimization test and update for @@ -3050,7 +3050,7 @@ test_actual_io_mode(int selection_mode) * process. To get mixed I/O, have the root select all chunks and each * subsequent process select the first and nth chunk. The first chunk, * accessed by all, will be assigned collective I/O while each other chunk - * will be accessed only by the root and the nth procecess and will be + * will be accessed only by the root and the nth process and will be * assigned independent I/O. Each process will access one chunk collectively * and at least one chunk independently, reporting mixed I/O. */ @@ -3077,7 +3077,7 @@ test_actual_io_mode(int selection_mode) break; /* RESET tests that the properties are properly reset to defaults each time I/O is - * performed. To acheive this, we have RESET perform collective I/O (which would change + * performed. To achieve this, we have RESET perform collective I/O (which would change * the values from the defaults) followed by independent I/O (which should report the * default values). RESET doesn't need to have a unique selection, so we reuse * MULTI_CHUMK_MIX_DISAGREE, which was chosen because it is a complex case that works @@ -3093,7 +3093,7 @@ test_actual_io_mode(int selection_mode) * assgigned collective I/O, while each other process gets independent I/O. * Since the root process with only access the first chunk, it will report * collective I/O. The subsequent processes will access the first chunk - * collectively, and their other chunk indpendently, reporting mixed I/O. + * collectively, and their other chunk independently, reporting mixed I/O. */ if (mpi_rank == 0) { @@ -3201,14 +3201,14 @@ test_actual_io_mode(int selection_mode) /* Set the threshold number of processes per chunk to twice mpi_size. * This will prevent the threshold from ever being met, thus forcing * multi chunk io instead of link chunk io. - * This is via deault. + * This is via default. */ if (multi_chunk_io) { /* force multi-chunk-io by threshold */ ret = H5Pset_dxpl_mpio_chunk_opt_num(dxpl_write, (unsigned)mpi_size * 2); VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_num succeeded"); - /* set this to manipulate testing senario about allocating processes + /* set this to manipulate testing scenario about allocating processes * to chunks */ ret = H5Pset_dxpl_mpio_chunk_opt_ratio(dxpl_write, (unsigned)99); VRFY((ret >= 0), "H5Pset_dxpl_mpio_chunk_opt_ratio succeeded"); @@ -3232,12 +3232,12 @@ test_actual_io_mode(int selection_mode) H5Eprint2(H5E_DEFAULT, stdout); VRFY((ret >= 0), "H5Dwrite() dataset multichunk write succeeded"); - /* Retreive Actual io valuess */ + /* Retrieve Actual io values */ ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write); - VRFY((ret >= 0), "retriving actual io mode suceeded"); + VRFY((ret >= 0), "retrieving actual io mode succeeded"); ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write); - VRFY((ret >= 0), "retriving actual chunk opt mode succeeded"); + VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded"); /* Read */ ret = H5Dread(dataset, data_type, mem_space, file_space, dxpl_read, buffer); @@ -3245,12 +3245,12 @@ test_actual_io_mode(int selection_mode) H5Eprint2(H5E_DEFAULT, stdout); VRFY((ret >= 0), "H5Dread() dataset multichunk read succeeded"); - /* Retreive Actual io values */ + /* Retrieve Actual io values */ ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read); - VRFY((ret >= 0), "retriving actual io mode succeeded"); + VRFY((ret >= 0), "retrieving actual io mode succeeded"); ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read); - VRFY((ret >= 0), "retriving actual chunk opt mode succeeded"); + VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded"); /* Check write vs read */ VRFY((actual_io_mode_read == actual_io_mode_write), @@ -3271,7 +3271,7 @@ test_actual_io_mode(int selection_mode) actual_io_mode_write); } - /* To test that the property is succesfully reset to the default, we perform some + /* To test that the property is successfully reset to the default, we perform some * independent I/O after the collective I/O */ if (selection_mode == TEST_ACTUAL_IO_RESET) { @@ -3288,9 +3288,9 @@ test_actual_io_mode(int selection_mode) /* Check Properties */ ret = H5Pget_mpio_actual_io_mode(dxpl_write, &actual_io_mode_write); - VRFY((ret >= 0), "retriving actual io mode succeeded"); + VRFY((ret >= 0), "retrieving actual io mode succeeded"); ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_write, &actual_chunk_opt_mode_write); - VRFY((ret >= 0), "retriving actual chunk opt mode succeeded"); + VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded"); VRFY(actual_chunk_opt_mode_write == H5D_MPIO_NO_CHUNK_OPTIMIZATION, "actual_chunk_opt_mode has correct value for reset write (independent)"); @@ -3303,9 +3303,9 @@ test_actual_io_mode(int selection_mode) /* Check Properties */ ret = H5Pget_mpio_actual_io_mode(dxpl_read, &actual_io_mode_read); - VRFY((ret >= 0), "retriving actual io mode succeeded"); + VRFY((ret >= 0), "retrieving actual io mode succeeded"); ret = H5Pget_mpio_actual_chunk_opt_mode(dxpl_read, &actual_chunk_opt_mode_read); - VRFY((ret >= 0), "retriving actual chunk opt mode succeeded"); + VRFY((ret >= 0), "retrieving actual chunk opt mode succeeded"); VRFY(actual_chunk_opt_mode_read == H5D_MPIO_NO_CHUNK_OPTIMIZATION, "actual_chunk_opt_mode has correct value for reset read (independent)"); @@ -3358,11 +3358,11 @@ actual_io_mode_tests(void) test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_IND); test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_COL); - /* The Multi Chunk Mixed test requires atleast three processes. */ + /* The Multi Chunk Mixed test requires at least three processes. */ if (mpi_size > 2) test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX); else - HDfprintf(stdout, "Multi Chunk Mixed test requires 3 proceses minimum\n"); + HDfprintf(stdout, "Multi Chunk Mixed test requires 3 processes minimum\n"); test_actual_io_mode(TEST_ACTUAL_IO_MULTI_CHUNK_MIX_DISAGREE); @@ -3406,7 +3406,7 @@ actual_io_mode_tests(void) * Test for Data Type Conversion as the cause of breaking collective I/O. * * TEST_DATA_TRANSFORMS: - * Test for Data Transfrom feature as the cause of breaking collective I/O. + * Test for Data Transform feature as the cause of breaking collective I/O. * * TEST_NOT_SIMPLE_OR_SCALAR_DATASPACES: * Test for NULL dataspace as the cause of breaking collective I/O. @@ -3545,7 +3545,7 @@ test_no_collective_cause_mode(int selection_mode) } if (selection_mode & TEST_DATA_TRANSFORMS) { - test_name = "Broken Collective I/O - DATA Transfroms"; + test_name = "Broken Collective I/O - DATA Transforms"; no_collective_cause_local_expected |= H5D_MPIO_DATA_TRANSFORMS; no_collective_cause_global_expected |= H5D_MPIO_DATA_TRANSFORMS; } @@ -3635,7 +3635,7 @@ test_no_collective_cause_mode(int selection_mode) /* Get the cause of broken collective I/O */ ret = H5Pget_mpio_no_collective_cause(dxpl_write, &no_collective_cause_local_write, &no_collective_cause_global_write); - VRFY((ret >= 0), "retriving no collective cause succeeded"); + VRFY((ret >= 0), "retrieving no collective cause succeeded"); /*--------------------- * Test Read access @@ -3655,7 +3655,7 @@ test_no_collective_cause_mode(int selection_mode) /* Get the cause of broken collective I/O */ ret = H5Pget_mpio_no_collective_cause(dxpl_read, &no_collective_cause_local_read, &no_collective_cause_global_read); - VRFY((ret >= 0), "retriving no collective cause succeeded"); + VRFY((ret >= 0), "retrieving no collective cause succeeded"); /* Check write vs read */ VRFY((no_collective_cause_local_read == no_collective_cause_local_write), @@ -3839,7 +3839,7 @@ dataset_atomicity(void) /* file locking allows only one file open (serial) for writing */ if (MAINPROCESS) { fid = H5Fopen(filename, H5F_ACC_RDWR, H5P_DEFAULT); - VRFY((fid >= 0), "H5Fopen succeeed"); + VRFY((fid >= 0), "H5Fopen succeeded"); /* should fail */ H5E_BEGIN_TRY diff --git a/testpar/t_file.c b/testpar/t_file.c index 4e8835c..bfd9578 100644 --- a/testpar/t_file.c +++ b/testpar/t_file.c @@ -182,7 +182,7 @@ test_page_buffer_access(void) data = (int *)HDmalloc(sizeof(int) * (size_t)num_elements); - /* intialize all the elements to have a value of -1 */ + /* initialize all the elements to have a value of -1 */ for (i = 0; i < num_elements; i++) data[i] = -1; diff --git a/testpar/t_filters_parallel.c b/testpar/t_filters_parallel.c index 305fb56..2099120 100644 --- a/testpar/t_filters_parallel.c +++ b/testpar/t_filters_parallel.c @@ -5,12 +5,10 @@ * * * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * - * the files COPYING and Copyright.html. COPYING can be found at the root * - * of the source code distribution tree; Copyright.html can be found at the * - * root level of an installed copy of the electronic HDF5 document set and * - * is linked from the top-level documents page. It can also be found at * - * http://hdfgroup.org/HDF5/doc/Copyright.html. If you do not have * - * access to either file, you may request a copy from help@hdfgroup.org. * + * the COPYING file, which can be found at the root of the source code * + * distribution tree, or in https://www.hdfgroup.org/licenses. * + * If you do not have access to either file, you may request a copy from * + * help@hdfgroup.org. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ /* diff --git a/testpar/t_filters_parallel.h b/testpar/t_filters_parallel.h index 800604c..ea50c81 100644 --- a/testpar/t_filters_parallel.h +++ b/testpar/t_filters_parallel.h @@ -5,12 +5,10 @@ * * * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * - * the files COPYING and Copyright.html. COPYING can be found at the root * - * of the source code distribution tree; Copyright.html can be found at the * - * root level of an installed copy of the electronic HDF5 document set and * - * is linked from the top-level documents page. It can also be found at * - * http://hdfgroup.org/HDF5/doc/Copyright.html. If you do not have * - * access to either file, you may request a copy from help@hdfgroup.org. * + * the COPYING file, which can be found at the root of the source code * + * distribution tree, or in https://www.hdfgroup.org/licenses. * + * If you do not have access to either file, you may request a copy from * + * help@hdfgroup.org. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ /* diff --git a/testpar/t_init_term.c b/testpar/t_init_term.c index 1533765..5bb3df3 100644 --- a/testpar/t_init_term.c +++ b/testpar/t_init_term.c @@ -64,7 +64,7 @@ main(int argc, char **argv) if (0 == nerrors) PASSED(); else - H5_FAILED() + H5_FAILED(); } return (nerrors != 0); diff --git a/testpar/t_mdset.c b/testpar/t_mdset.c index 5d6f8d8..a107dad 100644 --- a/testpar/t_mdset.c +++ b/testpar/t_mdset.c @@ -1710,7 +1710,7 @@ io_mode_confusion(void) * Open the each of the data sets in turn. If all opens are successful, * the test passes. Otherwise the test fails. * - * Note that this test will probably become irrelevent shortly, when we + * Note that this test will probably become irrelevant shortly, when we * land the journaling modifications on the trunk -- at which point all * cache clients will have to construct on disk images on demand. * @@ -1771,7 +1771,7 @@ rr_obj_hdr_flush_confusion(void) mrc = MPI_Comm_split(MPI_COMM_WORLD, is_reader, mpi_rank, &comm); VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split"); - /* The reader proocesses branches off to do reading + /* The reader processes branches off to do reading * while the writer processes continues to do writing * Whenever writers finish one writing step, including a H5Fflush, * they inform the readers, via MPI_COMM_WORLD, to verify. @@ -2054,8 +2054,8 @@ rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) /* * flush the metadata cache yet again to clean the object headers. * - * This is an attempt to crate a situation where we have dirty - * object header continuation chunks, but clean opject headers + * This is an attempt to create a situation where we have dirty + * object header continuation chunks, but clean object headers * to verify a speculative bug fix -- it doesn't seem to work, * but I will leave the code in anyway, as the object header * code is going to change a lot in the near future. diff --git a/testpar/t_mpi.c b/testpar/t_mpi.c index 3d95450..37e38dd 100644 --- a/testpar/t_mpi.c +++ b/testpar/t_mpi.c @@ -225,7 +225,7 @@ test_mpio_gb_file(char *filename) * sizes. */ if (MAINPROCESS) { /* only process 0 needs to check it*/ - HDprintf("MPI_Offset is %s %d bytes integeral type\n", is_signed ? "signed" : "unsigned", + HDprintf("MPI_Offset is %s %d bytes integral type\n", is_signed ? "signed" : "unsigned", (int)sizeof(MPI_Offset)); if (sizeof_mpi_offset <= 4 && is_signed) { HDprintf("Skipped 2GB range test " @@ -410,7 +410,7 @@ finish: * MPI-IO Test: One writes, Many reads. * Verify if only one process writes some data and then all other * processes can read them back correctly. This tests if the - * underlaying parallel I/O and file system supports parallel I/O + * underlying parallel I/O and file system supports parallel I/O * correctly. * * Algorithm: Only one process (e.g., process 0) writes some data. diff --git a/testpar/t_pflush1.c b/testpar/t_pflush1.c index e242257..4fcc69c 100644 --- a/testpar/t_pflush1.c +++ b/testpar/t_pflush1.c @@ -215,6 +215,7 @@ error: HDfflush(stderr); HDprintf("*** ERROR ***\n"); HDprintf("THERE WAS A REAL ERROR IN t_pflush1.\n"); + HDfflush(stdout); if (data_g) HDfree(data_g); diff --git a/testpar/t_pflush2.c b/testpar/t_pflush2.c index a3df64c..edcb89f 100644 --- a/testpar/t_pflush2.c +++ b/testpar/t_pflush2.c @@ -208,7 +208,7 @@ main(int argc, char *argv[]) PASSED(); } else { - H5_FAILED() + H5_FAILED(); goto error; } diff --git a/testpar/t_ph5basic.c b/testpar/t_ph5basic.c index 05f2c77..f8d4129 100644 --- a/testpar/t_ph5basic.c +++ b/testpar/t_ph5basic.c @@ -56,7 +56,7 @@ test_fapl_mpio_dup(void) HDprintf("rank/size of MPI_COMM_WORLD are %d/%d\n", mpi_rank, mpi_size); /* Create a new communicator that has the same processes as MPI_COMM_WORLD. - * Use MPI_Comm_split because it is simplier than MPI_Comm_create + * Use MPI_Comm_split because it is simpler than MPI_Comm_create */ mrc = MPI_Comm_split(MPI_COMM_WORLD, 0, 0, &comm); VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split"); @@ -141,7 +141,7 @@ test_fapl_mpio_dup(void) VRFY((ret >= 0), "H5Pget_fapl_mpio neither"); /* now get both and check validity too. */ - /* Donot free the returned objects which are used in the next case. */ + /* Do not free the returned objects which are used in the next case. */ ret = H5Pget_fapl_mpio(acc_pl, &comm_tmp, &info_tmp); VRFY((ret >= 0), "H5Pget_fapl_mpio"); MPI_Comm_size(comm_tmp, &mpi_size_tmp); diff --git a/testpar/t_pread.c b/testpar/t_pread.c index 9783284..ba01f42 100644 --- a/testpar/t_pread.c +++ b/testpar/t_pread.c @@ -68,7 +68,7 @@ static char *test_argv0 = NULL; * more or less in half. Each sub group will operate * collectively on their assigned file. This split into * subgroups validates that parallel groups can successfully - * open and read data independantly from the other parallel + * open and read data independently from the other parallel * operations taking place. * * Return: Success: 0 @@ -132,7 +132,7 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) } /* Setup the file names - * The test specfic filenames are stored as consecutive + * The test specific filenames are stored as consecutive * array entries in the global 'FILENAMES' array above. * Here, we simply decide on the starting index for * file construction. The reading portion of the test @@ -1043,11 +1043,11 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) * Purpose: To implement a parallel test which validates whether the * new superblock lookup functionality is working correctly. * - * The test consists of creating two seperate HDF datasets + * The test consists of creating two separate HDF datasets * in which random text is inserted at the start of each * file using the 'j5jam' application. This forces the * HDF5 file superblock to a non-zero offset. - * Having created the two independant files, we create two + * Having created the two independent files, we create two * non-overlapping MPI groups, each of which is then tasked * with the opening and validation of the data contained * therein. @@ -1152,7 +1152,7 @@ main(int argc, char **argv) goto finish; } - /* Now read the generated test file (stil using MPI_COMM_WORLD) */ + /* Now read the generated test file (still using MPI_COMM_WORLD) */ nerrs += test_parallel_read(MPI_COMM_WORLD, mpi_rank, mpi_size, which_group); if (nerrs > 0) { diff --git a/testpar/t_prestart.c b/testpar/t_prestart.c index 384fb1b..e6c40ac 100644 --- a/testpar/t_prestart.c +++ b/testpar/t_prestart.c @@ -125,7 +125,7 @@ main(int argc, char **argv) if (0 == nerrors) PASSED(); else - H5_FAILED() + H5_FAILED(); } return (nerrors != 0); diff --git a/testpar/t_prop.c b/testpar/t_prop.c index a0de987..67b4b50 100644 --- a/testpar/t_prop.c +++ b/testpar/t_prop.c @@ -76,7 +76,7 @@ test_encode_decode(hid_t orig_pl, int mpi_rank, int recv_proc) HDfree(sbuf); MPI_Barrier(MPI_COMM_WORLD); - return (0); + return 0; } void diff --git a/testpar/t_pshutdown.c b/testpar/t_pshutdown.c index b8028a4..3bf0e04e 100644 --- a/testpar/t_pshutdown.c +++ b/testpar/t_pshutdown.c @@ -114,7 +114,7 @@ main(int argc, char **argv) if (0 == nerrors) PASSED(); else - H5_FAILED() + H5_FAILED(); } return (nerrors != 0); diff --git a/testpar/t_shapesame.c b/testpar/t_shapesame.c index 42e960e..cbae5e1 100644 --- a/testpar/t_shapesame.c +++ b/testpar/t_shapesame.c @@ -11,7 +11,7 @@ * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ /* - This program will test independant and collective reads and writes between + This program will test independent and collective reads and writes between selections of different rank that non-the-less are deemed as having the same shape by H5Sselect_shape_same(). */ @@ -415,11 +415,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker /* setup selections for writing initial data to the small data set */ ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded"); if (MAINPROCESS) { /* add an additional slice to the selections */ @@ -427,11 +427,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, or) succeeded"); } /* write the initial value of the small data set to file */ @@ -475,21 +475,21 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) succeeded"); /* In passing, setup the process slice dataspaces as well */ ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_process_slice_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_process_slice_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_process_slice_sid, set) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_process_slice_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_process_slice_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_process_slice_sid, set) succeeded"); if (MAINPROCESS) { /* add an additional slice to the selections */ @@ -497,11 +497,11 @@ hs_dr_pio_test__setup(const int test_num, const int edge_size, const int checker ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_OR, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, or) succeeded"); } /* write the initial value of the large data set to file */ @@ -736,7 +736,7 @@ contig_hs_dr_pio_test__d2m_l2s(struct hs_dr_pio_test_vars_t *tv_ptr) * of the large data set. However, in the parallel version, each * process only works with that slice of the large cube indicated * by its rank -- hence we set the most slowly changing index to - * mpi_rank, and don't itterate over it. + * mpi_rank, and don't iterate over it. */ if (PAR_SS_DR_MAX_RANK - tv_ptr->large_rank == 0) { @@ -934,7 +934,7 @@ contig_hs_dr_pio_test__d2m_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded"); #if CONTIG_HS_DR_PIO_TEST__D2M_S2L__DEBUG HDfprintf(stdout, "%s reading slices of on disk small data set into slices of big data set.\n", fcnName); @@ -966,7 +966,7 @@ contig_hs_dr_pio_test__d2m_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) * * However, in the parallel version, each process only works with that * slice of the large (and small) data set indicated by its rank -- hence - * we set the most slowly changing index to mpi_rank, and don't itterate + * we set the most slowly changing index to mpi_rank, and don't iterate * over it. */ @@ -1187,11 +1187,11 @@ contig_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr) ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded"); /* set up start, stride, count, and block -- note that we will * change start[] so as to read slices of the large cube. @@ -1223,7 +1223,7 @@ contig_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr) * * However, in the parallel version, each process only works with that * slice of the large (and small) data set indicated by its rank -- hence - * we set the most slowly changing index to mpi_rank, and don't itterate + * we set the most slowly changing index to mpi_rank, and don't iterate * over it. */ @@ -1465,7 +1465,7 @@ contig_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded"); /* set up start, stride, count, and block -- note that we will * change start[] so as to write slices of the small data set to @@ -1575,7 +1575,7 @@ contig_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) ret = H5Dwrite(tv_ptr->large_dataset, H5T_NATIVE_UINT32, tv_ptr->large_ds_slice_sid, tv_ptr->file_large_ds_process_slice_sid, tv_ptr->xfer_plist, tv_ptr->large_ds_buf_2); - VRFY((ret != FAIL), "H5Dwrite() to zero large ds suceeded"); + VRFY((ret != FAIL), "H5Dwrite() to zero large ds succeeded"); /* select the portion of the in memory large cube to which we * are going to write data. @@ -2012,8 +2012,8 @@ contig_hs_dr_pio_test(ShapeSameTestMethods sstest_type) ** edge_size, and a checker_edge_size, select a checker ** board selection of a sel_rank (sel_rank < tgt_rank) ** dimensional slice through the dataspace parallel to the -** sel_rank fastest changing indicies, with origin (in the -** higher indicies) as indicated by the start array. +** sel_rank fastest changing indices, with origin (in the +** higher indices) as indicated by the start array. ** ** Note that this function, like all its relatives, is ** hard coded to presume a maximum dataspace rank of 5. @@ -2322,7 +2322,7 @@ ckrbrd_hs_dr_pio_test__slct_ckrbrd(const int mpi_rank, const hid_t tgt_sid, cons ** or writing a checker board selection of an m (1 <= m < ** rank) dimensional slice through this processes slice ** of the target data set. Also, this slice must be parallel -** to the fastest changing indicies. +** to the fastest changing indices. ** ** It is further presumed that the buffer was zeroed before ** the read/write, and that the full target data set (i.e. @@ -2357,7 +2357,7 @@ ckrbrd_hs_dr_pio_test__slct_ckrbrd(const int mpi_rank, const hid_t tgt_sid, cons ** of the buffer resides either at the origin of either ** a selected or an unselected checker. (Translation: ** if partial checkers appear in the buffer, they will -** intersect the edges of the n-cube oposite the origin.) +** intersect the edges of the n-cube opposite the origin.) ** ****************************************************************/ @@ -2607,7 +2607,7 @@ ckrbrd_hs_dr_pio_test__d2m_l2s(struct hs_dr_pio_test_vars_t *tv_ptr) * of the large data set. However, in the parallel version, each * process only works with that slice of the large cube indicated * by its rank -- hence we set the most slowly changing index to - * mpi_rank, and don't itterate over it. + * mpi_rank, and don't iterate over it. */ if (PAR_SS_DR_MAX_RANK - tv_ptr->large_rank == 0) { @@ -2830,7 +2830,7 @@ ckrbrd_hs_dr_pio_test__d2m_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) * * However, in the parallel version, each process only works with that * slice of the large (and small) data set indicated by its rank -- hence - * we set the most slowly changing index to mpi_rank, and don't itterate + * we set the most slowly changing index to mpi_rank, and don't iterate * over it. */ @@ -3097,11 +3097,11 @@ ckrbrd_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr) ret = H5Sselect_hyperslab(tv_ptr->file_small_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid_0, set) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->mem_small_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded"); sel_start[0] = sel_start[1] = sel_start[2] = sel_start[3] = sel_start[4] = 0; sel_start[tv_ptr->small_ds_offset] = (hsize_t)(tv_ptr->mpi_rank); @@ -3142,7 +3142,7 @@ ckrbrd_hs_dr_pio_test__m2d_l2s(struct hs_dr_pio_test_vars_t *tv_ptr) * * However, in the parallel version, each process only works with that * slice of the large (and small) data set indicated by its rank -- hence - * we set the most slowly changing index to mpi_rank, and don't itterate + * we set the most slowly changing index to mpi_rank, and don't iterate * over it. */ @@ -3391,11 +3391,11 @@ ckrbrd_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) ret = H5Sselect_hyperslab(tv_ptr->file_large_ds_sid_0, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid_0, set) succeeded"); ret = H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, H5S_SELECT_SET, tv_ptr->start, tv_ptr->stride, tv_ptr->count, tv_ptr->block); - VRFY((ret >= 0), "H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(tv_ptr->mem_large_ds_sid, set) succeeded"); /* setup a checkerboard selection of the slice of the in memory small * data set associated with the process's mpi rank. @@ -3503,7 +3503,7 @@ ckrbrd_hs_dr_pio_test__m2d_s2l(struct hs_dr_pio_test_vars_t *tv_ptr) */ ret = H5Dwrite(tv_ptr->large_dataset, H5T_NATIVE_UINT32, tv_ptr->mem_large_ds_sid, tv_ptr->file_large_ds_sid_0, tv_ptr->xfer_plist, tv_ptr->large_ds_buf_2); - VRFY((ret != FAIL), "H5Dwrite() to zero large ds suceeded"); + VRFY((ret != FAIL), "H5Dwrite() to zero large ds succeeded"); /* select the portion of the in memory large cube to which we * are going to write data. @@ -4218,42 +4218,42 @@ create_faccess_plist(MPI_Comm comm, MPI_Info info, int l_facc_type) return (ret_pl); } -/* Shape Same test using contigous hyperslab using independent IO on contigous datasets */ +/* Shape Same test using contiguous hyperslab using independent IO on contiguous datasets */ static void sscontig1(void) { contig_hs_dr_pio_test(IND_CONTIG); } -/* Shape Same test using contigous hyperslab using collective IO on contigous datasets */ +/* Shape Same test using contiguous hyperslab using collective IO on contiguous datasets */ static void sscontig2(void) { contig_hs_dr_pio_test(COL_CONTIG); } -/* Shape Same test using contigous hyperslab using independent IO on chunked datasets */ +/* Shape Same test using contiguous hyperslab using independent IO on chunked datasets */ static void sscontig3(void) { contig_hs_dr_pio_test(IND_CHUNKED); } -/* Shape Same test using contigous hyperslab using collective IO on chunked datasets */ +/* Shape Same test using contiguous hyperslab using collective IO on chunked datasets */ static void sscontig4(void) { contig_hs_dr_pio_test(COL_CHUNKED); } -/* Shape Same test using checker hyperslab using independent IO on contigous datasets */ +/* Shape Same test using checker hyperslab using independent IO on contiguous datasets */ static void sschecker1(void) { ckrbrd_hs_dr_pio_test(IND_CONTIG); } -/* Shape Same test using checker hyperslab using collective IO on contigous datasets */ +/* Shape Same test using checker hyperslab using collective IO on contiguous datasets */ static void sschecker2(void) { @@ -4321,7 +4321,7 @@ main(int argc, char **argv) /* Initialize testing framework */ TestInit(argv[0], usage, parse_options); - /* Shape Same tests using contigous hyperslab */ + /* Shape Same tests using contiguous hyperslab */ AddTest("sscontig1", sscontig1, NULL, "Cntg hslab, ind IO, cntg dsets", PARATESTFILE); AddTest("sscontig2", sscontig2, NULL, "Cntg hslab, col IO, cntg dsets", PARATESTFILE); AddTest("sscontig3", sscontig3, NULL, "Cntg hslab, ind IO, chnk dsets", PARATESTFILE); diff --git a/testpar/t_span_tree.c b/testpar/t_span_tree.c index 6f0efa3..7b9c6f7 100644 --- a/testpar/t_span_tree.c +++ b/testpar/t_span_tree.c @@ -431,7 +431,7 @@ coll_write_test(int chunk_factor) the correctedness of collective write compared with independent write, - In order to throughly test this feature, we choose + In order to thoroughly test this feature, we choose a different selection set for reading the data out. @@ -891,8 +891,8 @@ coll_read_test(void) ** edge_size, and a checker_edge_size, select a checker ** board selection of a sel_rank (sel_rank < tgt_rank) ** dimensional slice through the dataspace parallel to the -** sel_rank fastest changing indicies, with origin (in the -** higher indicies) as indicated by the start array. +** sel_rank fastest changing indices, with origin (in the +** higher indices) as indicated by the start array. ** ** Note that this function, is hard coded to presume a ** maximum dataspace rank of 5. @@ -1229,7 +1229,7 @@ lower_dim_size_comp_test__select_checker_board(const int mpi_rank, const hid_t t ** or writing a checker board selection of an m (1 <= m < ** rank) dimensional slice through this processes slice ** of the target data set. Also, this slice must be parallel -** to the fastest changing indicies. +** to the fastest changing indices. ** ** It is further presumed that the buffer was zeroed before ** the read/write, and that the full target data set (i.e. @@ -1264,7 +1264,7 @@ lower_dim_size_comp_test__select_checker_board(const int mpi_rank, const hid_t t ** of the buffer resides either at the origin of either ** a selected or an unselected checker. (Translation: ** if partial checkers appear in the buffer, they will -** intersect the edges of the n-cube oposite the origin.) +** intersect the edges of the n-cube opposite the origin.) ** ****************************************************************/ @@ -1725,7 +1725,7 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_ if (!use_collective_io) { ret = H5Pset_dxpl_mpio_collective_opt(xfer_plist, H5FD_MPIO_INDIVIDUAL_IO); - VRFY((ret >= 0), "H5Pset_dxpl_mpio_collective_opt() suceeded"); + VRFY((ret >= 0), "H5Pset_dxpl_mpio_collective_opt() succeeded"); } /* setup selection to write initial data to the small data sets */ @@ -1757,10 +1757,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_ /* setup selections for writing initial data to the small data set */ ret = H5Sselect_hyperslab(mem_small_ds_sid, H5S_SELECT_SET, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, set) succeeded"); ret = H5Sselect_hyperslab(file_small_ds_sid, H5S_SELECT_SET, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, set) succeeded"); if (MAINPROCESS) { /* add an additional slice to the selections */ @@ -1781,10 +1781,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_ #endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */ ret = H5Sselect_hyperslab(mem_small_ds_sid, H5S_SELECT_OR, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_small_ds_sid, or) succeeded"); ret = H5Sselect_hyperslab(file_small_ds_sid, H5S_SELECT_OR, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_small_ds_sid, or) succeeded"); } check = H5Sselect_valid(mem_small_ds_sid); @@ -1863,10 +1863,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_ #endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */ ret = H5Sselect_hyperslab(mem_large_ds_sid, H5S_SELECT_SET, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, set) succeeded"); ret = H5Sselect_hyperslab(file_large_ds_sid, H5S_SELECT_SET, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, set) succeeded"); #if LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG if (mpi_rank == LOWER_DIM_SIZE_COMP_TEST_DEBUG_TARGET_RANK) { @@ -1896,10 +1896,10 @@ lower_dim_size_comp_test__run_test(const int chunk_edge_size, const hbool_t use_ #endif /* LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG */ ret = H5Sselect_hyperslab(mem_large_ds_sid, H5S_SELECT_OR, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(mem_large_ds_sid, or) succeeded"); ret = H5Sselect_hyperslab(file_large_ds_sid, H5S_SELECT_OR, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, or) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_large_ds_sid, or) succeeded"); #if LOWER_DIM_SIZE_COMP_TEST__RUN_TEST__DEBUG if (mpi_rank == LOWER_DIM_SIZE_COMP_TEST_DEBUG_TARGET_RANK) { @@ -2426,7 +2426,7 @@ link_chunk_collective_io_test(void) /* select the file and mem spaces */ start[0] = (hsize_t)(mpi_rank * LINK_CHUNK_COLLECTIVE_IO_TEST_CHUNK_SIZE); ret = H5Sselect_hyperslab(file_ds_sid, H5S_SELECT_SET, start, stride, count, block); - VRFY((ret >= 0), "H5Sselect_hyperslab(file_ds_sid, set) suceeded"); + VRFY((ret >= 0), "H5Sselect_hyperslab(file_ds_sid, set) succeeded"); ret = H5Sselect_all(write_mem_ds_sid); VRFY((ret != FAIL), "H5Sselect_all(mem_ds_sid) succeeded"); diff --git a/testpar/testpflush.sh.in b/testpar/testpflush.sh.in index be0dcb3..b400447 100644 --- a/testpar/testpflush.sh.in +++ b/testpar/testpflush.sh.in @@ -50,7 +50,7 @@ RUNPARALLELSCRIPT=`echo "@RUNPARALLEL@" | sed "s/$$/\$/g"` echo "*** NOTE ***********************************************************" echo "You may see complaints from mpiexec et al. that not all processes" echo "called MPI_Finalize(). This is an intended characteristic of the" -echo "test and should not be considered an error." +echo "test and should not be considered an error." echo "********************************************************************" eval ${RUNPARALLELSCRIPT} ./t_pflush1 diff --git a/testpar/testphdf5.h b/testpar/testphdf5.h index 0085fd5..ac99651 100644 --- a/testpar/testphdf5.h +++ b/testpar/testphdf5.h @@ -214,8 +214,8 @@ typedef int DATATYPE; /* Shape Same Tests Definitions */ typedef enum { - IND_CONTIG, /* Independent IO on contigous datasets */ - COL_CONTIG, /* Collective IO on contigous datasets */ + IND_CONTIG, /* Independent IO on contiguous datasets */ + COL_CONTIG, /* Collective IO on contiguous datasets */ IND_CHUNKED, /* Independent IO on chunked datasets */ COL_CHUNKED /* Collective IO on chunked datasets */ } ShapeSameTestMethods; diff --git a/tools/lib/h5diff.c b/tools/lib/h5diff.c index 01a3563..e9a9fee 100644 --- a/tools/lib/h5diff.c +++ b/tools/lib/h5diff.c @@ -182,7 +182,7 @@ is_exclude_path(char *path, h5trav_type_t type, diff_opt_t *opts) /* search objects in exclude list */ while (NULL != exclude_path_ptr) { - /* if exclude path is is group, exclude its members as well */ + /* if exclude path is in group, exclude its members as well */ if (exclude_path_ptr->obj_type == H5TRAV_TYPE_GROUP) { ret_cmp = HDstrncmp(exclude_path_ptr->obj_path, path, HDstrlen(exclude_path_ptr->obj_path)); if (ret_cmp == 0) { /* found matching members */ @@ -246,7 +246,7 @@ is_exclude_attr(const char *path, h5trav_type_t type, diff_opt_t *opts) /* search objects in exclude list */ while (NULL != exclude_ptr) { - /* if exclude path is is group, exclude its members as well */ + /* if exclude path is in group, exclude its members as well */ if (exclude_ptr->obj_type == H5TRAV_TYPE_GROUP) { ret_cmp = HDstrncmp(exclude_ptr->obj_path, path, HDstrlen(exclude_ptr->obj_path)); if (ret_cmp == 0) { /* found matching members */ diff --git a/tools/lib/h5tools_ref.c b/tools/lib/h5tools_ref.c index c29c9c3..66d69ba 100644 --- a/tools/lib/h5tools_ref.c +++ b/tools/lib/h5tools_ref.c @@ -152,7 +152,7 @@ term_ref_path_table(void) * Purpose: Looks up a table entry given a path name. * Used during construction of the table. * - * Return: The table entre (pte) or NULL if not in the + * Return: The table entry or NULL if not in the * table. * * Programmer: REMcG diff --git a/tools/lib/h5tools_utils.c b/tools/lib/h5tools_utils.c index 16f0fea..02b1a8b 100644 --- a/tools/lib/h5tools_utils.c +++ b/tools/lib/h5tools_utils.c @@ -850,7 +850,7 @@ tmpfile(void) * Purpose: Get symbolic link (soft, external) info and its target object type (dataset, group, named datatype) and path, if exist * - * Patameters: + * Parameters: * - [IN] fileid : link file id * - [IN] linkpath : link path * - [OUT] link_info: returning target object info (h5tool_link_info_t) diff --git a/tools/lib/h5trav.c b/tools/lib/h5trav.c index afc9bc2..e88db8d 100644 --- a/tools/lib/h5trav.c +++ b/tools/lib/h5trav.c @@ -528,7 +528,7 @@ trav_info_free(trav_info_t *info) /*------------------------------------------------------------------------- * Function: trav_table_visit_obj * - * Purpose: Callback for visiting object, with 'table' sructure + * Purpose: Callback for visiting object, with 'table' structure * * Return: 0 on success, * -1 on failure @@ -553,7 +553,7 @@ trav_table_visit_obj(const char *path, const H5O_info_t *oinfo, const char *alre /*------------------------------------------------------------------------- * Function: trav_table_visit_lnk * - * Purpose: Callback for visiting link, with 'table' sructure + * Purpose: Callback for visiting link, with 'table' structure * * Return: 0 on success, * -1 on failure @@ -1109,7 +1109,7 @@ symlink_is_visited(symlink_trav_t *visited, H5L_type_t type, const char *file, c /* Look for symlink */ for (u = 0; u < visited->nused; u++) { /* Check for symlink values already in array */ - /* check type and path pair to distingush between symbolic links */ + /* check type and path pair to distinguish between symbolic links */ if ((visited->objs[u].type == type) && !HDstrcmp(visited->objs[u].path, path)) { /* if external link, file need to be matched as well */ if (visited->objs[u].type == H5L_TYPE_EXTERNAL) diff --git a/tools/libtest/h5tools_test_utils.c b/tools/libtest/h5tools_test_utils.c index 2f3cffb..0b2a4c5 100644 --- a/tools/libtest/h5tools_test_utils.c +++ b/tools/libtest/h5tools_test_utils.c @@ -240,7 +240,7 @@ tuple_testcase_t tuple_cases_g[] = { {"stuff keeps on going"}, }, { - "4-ple, escaped seperator", + "4-ple, escaped separator", "(elem0,elem1,el\\,em2,elem3)", /* "el\,em" */ ',', SUCCEED, diff --git a/tools/src/h5perf/perf.c b/tools/src/h5perf/perf.c index 54d83e1..cac7266 100644 --- a/tools/src/h5perf/perf.c +++ b/tools/src/h5perf/perf.c @@ -480,7 +480,7 @@ parse_args(int argc, char **argv) * further calls to getenv_all take place. * * Return: pointer to a string containing the value of the environment variable - * NULL if the varialbe doesn't exist in task 'root's environment. + * NULL if the variable doesn't exist in task 'root's environment. * * Programmer: Leon Arber * 4/4/05 diff --git a/tools/src/h5perf/pio_perf.c b/tools/src/h5perf/pio_perf.c index 6aff995..f997166 100644 --- a/tools/src/h5perf/pio_perf.c +++ b/tools/src/h5perf/pio_perf.c @@ -1694,7 +1694,7 @@ usage(const char *prog) HDprintf(" Environment variables:\n"); HDprintf(" HDF5_NOCLEANUP Do not remove data files if set [default remove]\n"); HDprintf(" HDF5_MPI_INFO MPI INFO object key=value separated by ;\n"); - HDprintf(" HDF5_PARAPREFIX Paralllel data files prefix\n"); + HDprintf(" HDF5_PARAPREFIX Parallel data files prefix\n"); fflush(stdout); } /* end if */ } /* end usage() */ diff --git a/tools/test/h5copy/dynlib_copy.c b/tools/test/h5copy/dynlib_copy.c index f474a55..e71ca47 100644 --- a/tools/test/h5copy/dynlib_copy.c +++ b/tools/test/h5copy/dynlib_copy.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/tools/test/h5copy/h5copygentest.c b/tools/test/h5copy/h5copygentest.c index d62639c..118286d 100644 --- a/tools/test/h5copy/h5copygentest.c +++ b/tools/test/h5copy/h5copygentest.c @@ -923,7 +923,7 @@ out: /*------------------------------------------------------------------------- * Function: Test_Extlink_Copy * - * Purpose: gerenate external link files + * Purpose: generate external link files * *------------------------------------------------------------------------*/ static void diff --git a/tools/test/h5diff/dynlib_diff.c b/tools/test/h5diff/dynlib_diff.c index 9d277e0..fd675b0 100644 --- a/tools/test/h5diff/dynlib_diff.c +++ b/tools/test/h5diff/dynlib_diff.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/tools/test/h5diff/h5diffgentest.c b/tools/test/h5diff/h5diffgentest.c index 4537a19..ae3d49e 100644 --- a/tools/test/h5diff/h5diffgentest.c +++ b/tools/test/h5diff/h5diffgentest.c @@ -2071,7 +2071,7 @@ gen_dataset_idx(const char *file, int format) int buf[10]; /* data buffer */ int i; /* local index variable */ - /* Get a copy of the file aaccess property */ + /* Get a copy of the file access property */ fapl = H5Pcreate(H5P_FILE_ACCESS); /* Set the "use the latest format" bounds for creating objects in the file */ diff --git a/tools/test/h5dump/dynlib_dump.c b/tools/test/h5dump/dynlib_dump.c index 9d277e0..fd675b0 100644 --- a/tools/test/h5dump/dynlib_dump.c +++ b/tools/test/h5dump/dynlib_dump.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/tools/test/h5dump/h5dumpgentest.c b/tools/test/h5dump/h5dumpgentest.c index 3589462..77345c4 100644 --- a/tools/test/h5dump/h5dumpgentest.c +++ b/tools/test/h5dump/h5dumpgentest.c @@ -5699,7 +5699,7 @@ gent_filters(void) ret = H5Tcommit2(fid, "mytype", tid, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); HDassert(ret >= 0); - ret = H5Oset_comment_by_name(fid, "mytype", "This is a commited datatype", H5P_DEFAULT); + ret = H5Oset_comment_by_name(fid, "mytype", "This is a committed datatype", H5P_DEFAULT); HDassert(ret >= 0); ret = H5Tclose(tid); @@ -7146,7 +7146,7 @@ gent_dataset_idx(void) int i, j; int H5_ATTR_NDEBUG_UNUSED ret; - /* Get a copy of the file aaccess property */ + /* Get a copy of the file access property */ fapl = H5Pcreate(H5P_FILE_ACCESS); /* Set the "use the latest version of the format" bounds for creating objects in the file */ diff --git a/tools/test/h5format_convert/h5fc_gentest.c b/tools/test/h5format_convert/h5fc_gentest.c index ccb64b5..5060432 100644 --- a/tools/test/h5format_convert/h5fc_gentest.c +++ b/tools/test/h5format_convert/h5fc_gentest.c @@ -63,7 +63,7 @@ const char *FILENAME[] = {"h5fc_ext1_i.h5", /* 0 */ /* * Function: gen_non() * - * Create a file with SWMR write+non-latest-format--this will result in v3 superbock+latest version support: + * Create a file with SWMR write+non-latest-format--this will result in v3 superblock+latest version support: * 1) 1 chunked dataset with extensible array chunk indexing type (without data) * 2) 1 chunked dataset with version 2 B-tree chunk indexing type (with data) * Re-open the file with write+non-latest-format and create: diff --git a/tools/test/h5ls/dynlib_ls.c b/tools/test/h5ls/dynlib_ls.c index 9d277e0..fd675b0 100644 --- a/tools/test/h5ls/dynlib_ls.c +++ b/tools/test/h5ls/dynlib_ls.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/tools/test/h5repack/dynlib_rpk.c b/tools/test/h5repack/dynlib_rpk.c index b9e863b..225cea8 100644 --- a/tools/test/h5repack/dynlib_rpk.c +++ b/tools/test/h5repack/dynlib_rpk.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/tools/test/h5repack/dynlib_vrpk.c b/tools/test/h5repack/dynlib_vrpk.c index 981b848..f9f9299 100644 --- a/tools/test/h5repack/dynlib_vrpk.c +++ b/tools/test/h5repack/dynlib_vrpk.c @@ -2,7 +2,7 @@ * Copyright by The HDF Group. * * All rights reserved. * * * - * This file is part of HDF5. The full HDF5 copyright notice, including * + * This file is part of HDF5. The full HDF5 copyright notice, including * * terms governing use, modification, and redistribution, is contained in * * the COPYING file, which can be found at the root of the source code * * distribution tree, or in https://www.hdfgroup.org/licenses. * diff --git a/tools/test/perform/gen_report.pl b/tools/test/perform/gen_report.pl index a1fbeb0..c853b70 100644 --- a/tools/test/perform/gen_report.pl +++ b/tools/test/perform/gen_report.pl @@ -391,7 +391,7 @@ sub plot_default_graph1 { print GNUPLOT_PIPE "set term x11 1\n"; print GNUPLOT_PIPE "set xlabel \"Number of Processors\"\n"; print GNUPLOT_PIPE "set title \"" . $data_type . " Performance (Speed vs. Num. Procs)\"\n"; - print GNUPLOT_PIPE "set ylabel \"Bandwdith (MB/s)\"\n"; + print GNUPLOT_PIPE "set ylabel \"Bandwidth (MB/s)\"\n"; print GNUPLOT_PIPE "set label 1 \"Transfer buffer size: " . $transfer_buffer_size . "K\" at graph 0.7, graph 0.7 left \n"; #the next line attempts to hack gnuplot to get around it's inability to linearly scale, but logarithmically label an axis @@ -435,7 +435,7 @@ sub plot_default_graph2 { print GNUPLOT_PIPE "set term x11 2\n"; print GNUPLOT_PIPE "set xlabel \"Transfer Buffer Size (in bytes)\"\n"; print GNUPLOT_PIPE "set title \"" . $data_type . " Performance (Speed vs. Transfer Buffer Size)\"\n"; - print GNUPLOT_PIPE "set ylabel \"Bandwdith (MB/s)\"\n"; + print GNUPLOT_PIPE "set ylabel \"Bandwidth (MB/s)\"\n"; print GNUPLOT_PIPE "set label 1 \"Procs: " . $num_procs_graph . "\" at graph 0.7, graph 0.7 left \n"; #the next line attempts to hack gnuplot to get around it's inability to linearly scale, but logarithmically label an axis diff --git a/tools/test/perform/overhead.c b/tools/test/perform/overhead.c index 41eef0c..e1036e5 100644 --- a/tools/test/perform/overhead.c +++ b/tools/test/perform/overhead.c @@ -182,7 +182,7 @@ test(fill_t fill_style, const double splits[], hbool_t verbose, hbool_t use_rdcc hsize_t hs_count[1] = {1}; /*hyperslab nelmts */ int fd = (-1); /*h5 file direct */ int *had = NULL; /*for random filling */ - const char *sname = NULL; /*fill style nam */ + const char *sname = NULL; /*fill style name */ int mdc_nelmts; /*num meta objs to cache*/ hsize_t i, k; int j; diff --git a/utils/mirror_vfd/mirror_writer.c b/utils/mirror_vfd/mirror_writer.c index f8d7544..a69ea73 100644 --- a/utils/mirror_vfd/mirror_writer.c +++ b/utils/mirror_vfd/mirror_writer.c @@ -291,7 +291,7 @@ session_start(int socketfd, const H5FD_mirror_xmit_open_t *xmit_open) { struct mirror_session *session = NULL; struct mirror_writer_opts opts; -#if 0 /* TODO: behaviro option */ +#if 0 /* TODO: behavior option */ char logpath[H5FD_MIRROR_XMIT_FILEPATH_MAX] = ""; #endif |