diff options
author | Dana Robinson <43805+derobins@users.noreply.github.com> | 2022-04-07 16:36:51 (GMT) |
---|---|---|
committer | GitHub <noreply@github.com> | 2022-04-07 16:36:51 (GMT) |
commit | 7df729bbf3e0f0b53dda73829f05fdf709993e4c (patch) | |
tree | f665f58f22b2d26cfae1138c9e38fea58a01ef9e /testpar/t_pread.c | |
parent | a4e833c50c002239af8c16667b1b072e6c6fa27f (diff) | |
download | hdf5-7df729bbf3e0f0b53dda73829f05fdf709993e4c.zip hdf5-7df729bbf3e0f0b53dda73829f05fdf709993e4c.tar.gz hdf5-7df729bbf3e0f0b53dda73829f05fdf709993e4c.tar.bz2 |
1.12: Brings changes from testpar over from develop (#1614)
Diffstat (limited to 'testpar/t_pread.c')
-rw-r--r-- | testpar/t_pread.c | 62 |
1 files changed, 31 insertions, 31 deletions
diff --git a/testpar/t_pread.c b/testpar/t_pread.c index cf91b31..d79130b 100644 --- a/testpar/t_pread.c +++ b/testpar/t_pread.c @@ -35,7 +35,7 @@ const char *FILENAMES[NFILENAME + 1] = {"reloc_t_pread_data_file", "reloc_t_prea #define LIMIT_NPROC 6 -hbool_t pass = true; +hbool_t pass = TRUE; static const char *random_hdf5_text = "Now is the time for all first-time-users of HDF5 to read their \ manual or go thru the tutorials!\n\ While you\'re at it, now is also the time to read up on MPI-IO."; @@ -68,7 +68,7 @@ static char *test_argv0 = NULL; * more or less in half. Each sub group will operate * collectively on their assigned file. This split into * subgroups validates that parallel groups can successfully - * open and read data independantly from the other parallel + * open and read data independently from the other parallel * operations taking place. * * Return: Success: 0 @@ -112,7 +112,7 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) float nextValue; float * data_slice = NULL; - pass = true; + pass = TRUE; HDassert(comm != MPI_COMM_NULL); @@ -132,7 +132,7 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) } /* Setup the file names - * The test specfic filenames are stored as consecutive + * The test specific filenames are stored as consecutive * array entries in the global 'FILENAMES' array above. * Here, we simply decide on the starting index for * file construction. The reading portion of the test @@ -256,14 +256,14 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) if (pass) { if ((dset_id = H5Dcreate2(file_id, "dataset0", H5T_NATIVE_FLOAT, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT)) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dcreate2() failed.\n"; } } if (pass) { if ((H5Dwrite(dset_id, H5T_NATIVE_FLOAT, memspace, filespace, dxpl_id, data_slice)) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dwrite() failed.\n"; } } @@ -273,14 +273,14 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) if (pass) { if ((dcpl_id = H5Pcreate(H5P_DATASET_CREATE)) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pcreate() failed.\n"; } } if (pass) { if ((H5Pset_chunk(dcpl_id, 1, chunk)) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pset_chunk() failed.\n"; } } @@ -289,27 +289,27 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) if ((dset_id_ch = H5Dcreate2(file_id, "dataset0_chunked", H5T_NATIVE_FLOAT, filespace, H5P_DEFAULT, dcpl_id, H5P_DEFAULT)) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dcreate2() failed.\n"; } } if (pass) { if ((H5Dwrite(dset_id_ch, H5T_NATIVE_FLOAT, memspace, filespace, dxpl_id, data_slice)) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dwrite() failed.\n"; } } if (pass || (dcpl_id != -1)) { if (H5Pclose(dcpl_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pclose(dcpl_id) failed.\n"; } } if (pass || (dset_id_ch != -1)) { if (H5Dclose(dset_id_ch) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dclose(dset_id_ch) failed.\n"; } } @@ -317,49 +317,49 @@ generate_test_file(MPI_Comm comm, int mpi_rank, int group_id) /* close file, etc. */ if (pass || (dset_id != -1)) { if (H5Dclose(dset_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dclose(dset_id) failed.\n"; } } if (pass || (memspace != -1)) { if (H5Sclose(memspace) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Sclose(memspace) failed.\n"; } } if (pass || (filespace != -1)) { if (H5Sclose(filespace) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Sclose(filespace) failed.\n"; } } if (pass || (file_id != -1)) { if (H5Fclose(file_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Fclose(file_id) failed.\n"; } } if (pass || (dxpl_id != -1)) { if (H5Pclose(dxpl_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pclose(dxpl_id) failed.\n"; } } if (pass || (fapl_id != -1)) { if (H5Pclose(fapl_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pclose(fapl_id) failed.\n"; } } if (pass || (fctmpl != -1)) { if (H5Pclose(fctmpl) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pclose(fctmpl) failed.\n"; } } @@ -661,14 +661,14 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) if (pass || (memspace != -1)) { if (H5Sclose(memspace) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Sclose(memspace) failed.\n"; } } if (pass || (filespace != -1)) { if (H5Sclose(filespace) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Sclose(filespace) failed.\n"; } } @@ -859,7 +859,7 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) if (pass || (filespace != -1)) { if (H5Sclose(filespace) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Sclose(filespace) failed.\n"; } } @@ -956,7 +956,7 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) if (pass || (memspace != -1)) { if (H5Sclose(memspace) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Sclose(memspace) failed.\n"; } } @@ -969,7 +969,7 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) if (pass || (dxpl_id != -1)) { if (H5Pclose(dxpl_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pclose(dxpl_id) failed.\n"; } } @@ -978,28 +978,28 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) /* close file, etc. */ if (pass || (dset_id != -1)) { if (H5Dclose(dset_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dclose(dset_id) failed.\n"; } } if (pass || (dset_id_ch != -1)) { if (H5Dclose(dset_id_ch) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Dclose(dset_id_ch) failed.\n"; } } if (pass || (file_id != -1)) { if (H5Fclose(file_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Fclose(file_id) failed.\n"; } } if (pass || (fapl_id != -1)) { if (H5Pclose(fapl_id) < 0) { - pass = false; + pass = FALSE; failure_mssg = "H5Pclose(fapl_id) failed.\n"; } } @@ -1043,11 +1043,11 @@ test_parallel_read(MPI_Comm comm, int mpi_rank, int mpi_size, int group_id) * Purpose: To implement a parallel test which validates whether the * new superblock lookup functionality is working correctly. * - * The test consists of creating two seperate HDF datasets + * The test consists of creating two separate HDF datasets * in which random text is inserted at the start of each * file using the 'j5jam' application. This forces the * HDF5 file superblock to a non-zero offset. - * Having created the two independant files, we create two + * Having created the two independent files, we create two * non-overlapping MPI groups, each of which is then tasked * with the opening and validation of the data contained * therein. @@ -1152,7 +1152,7 @@ main(int argc, char **argv) goto finish; } - /* Now read the generated test file (stil using MPI_COMM_WORLD) */ + /* Now read the generated test file (still using MPI_COMM_WORLD) */ nerrs += test_parallel_read(MPI_COMM_WORLD, mpi_rank, mpi_size, which_group); if (nerrs > 0) { |