diff options
author | Allen Byrne <byrn@hdfgroup.org> | 2020-10-01 15:29:53 (GMT) |
---|---|---|
committer | Allen Byrne <byrn@hdfgroup.org> | 2020-10-01 15:29:53 (GMT) |
commit | ec8ad09125032c9c232878d02e90aa86df162841 (patch) | |
tree | 5d63105d3a0db7d5bbee41c5b74b097649a2115e /testpar/t_mdset.c | |
parent | 453238e90e1574ef1c15e3c79f7fb3d77920e77c (diff) | |
download | hdf5-ec8ad09125032c9c232878d02e90aa86df162841.zip hdf5-ec8ad09125032c9c232878d02e90aa86df162841.tar.gz hdf5-ec8ad09125032c9c232878d02e90aa86df162841.tar.bz2 |
Source formatted
Diffstat (limited to 'testpar/t_mdset.c')
-rw-r--r-- | testpar/t_mdset.c | 1677 |
1 files changed, 803 insertions, 874 deletions
diff --git a/testpar/t_mdset.c b/testpar/t_mdset.c index ed8867d..11ed85d 100644 --- a/testpar/t_mdset.c +++ b/testpar/t_mdset.c @@ -14,13 +14,12 @@ #include "testphdf5.h" #include "H5Dprivate.h" -#define DIM 2 -#define SIZE 32 -#define NDATASET 4 +#define DIM 2 +#define SIZE 32 +#define NDATASET 4 #define GROUP_DEPTH 128 enum obj_type { is_group, is_dset }; - static int get_size(void); static void write_dataset(hid_t, hid_t, hid_t); static int read_dataset(hid_t, hid_t, hid_t); @@ -32,7 +31,6 @@ static int read_attribute(hid_t, int, int); static int check_value(DATATYPE *, DATATYPE *, int); static void get_slab(hsize_t[], hsize_t[], hsize_t[], hsize_t[], int); - /* * The size value computed by this function is used extensively in * configuring tests for the current number of processes. @@ -53,10 +51,11 @@ get_size(void) MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); /* needed for VRFY */ MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); - if(mpi_size > size ) { - if((mpi_size % 2) == 0 ) { + if (mpi_size > size) { + if ((mpi_size % 2) == 0) { size = mpi_size; - } else { + } + else { size = mpi_size + 1; } } @@ -64,7 +63,7 @@ get_size(void) VRFY((mpi_size <= size), "mpi_size <= size"); VRFY(((size % 2) == 0), "size isn't even"); - return(size); + return (size); } /* get_size() */ @@ -72,14 +71,15 @@ get_size(void) * Example of using PHDF5 to create a zero sized dataset. * */ -void zero_dim_dset(void) +void +zero_dim_dset(void) { - int mpi_size, mpi_rank; - const char *filename; - hid_t fid, plist, dcpl, dsid, sid; - hsize_t dim, chunk_dim; - herr_t ret; - int data[1]; + int mpi_size, mpi_rank; + const char *filename; + hid_t fid, plist, dcpl, dsid, sid; + hsize_t dim, chunk_dim; + herr_t ret; + int data[1]; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); @@ -87,37 +87,37 @@ void zero_dim_dset(void) filename = GetTestParameters(); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - VRFY((plist>=0), "create_faccess_plist succeeded"); + VRFY((plist >= 0), "create_faccess_plist succeeded"); fid = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); - VRFY((fid>=0), "H5Fcreate succeeded"); + VRFY((fid >= 0), "H5Fcreate succeeded"); ret = H5Pclose(plist); - VRFY((ret>=0), "H5Pclose succeeded"); + VRFY((ret >= 0), "H5Pclose succeeded"); dcpl = H5Pcreate(H5P_DATASET_CREATE); - VRFY((dcpl>=0), "failed H5Pcreate"); + VRFY((dcpl >= 0), "failed H5Pcreate"); /* Set 1 chunk size */ chunk_dim = 1; - ret = H5Pset_chunk(dcpl, 1, &chunk_dim); - VRFY((ret>=0), "failed H5Pset_chunk"); + ret = H5Pset_chunk(dcpl, 1, &chunk_dim); + VRFY((ret >= 0), "failed H5Pset_chunk"); /* Create 1D dataspace with 0 dim size */ dim = 0; sid = H5Screate_simple(1, &dim, NULL); - VRFY((sid>=0), "failed H5Screate_simple"); + VRFY((sid >= 0), "failed H5Screate_simple"); /* Create chunked dataset */ dsid = H5Dcreate2(fid, "dset", H5T_NATIVE_INT, sid, H5P_DEFAULT, dcpl, H5P_DEFAULT); - VRFY((dsid>=0), "failed H5Dcreate2"); + VRFY((dsid >= 0), "failed H5Dcreate2"); /* write 0 elements from dataset */ ret = H5Dwrite(dsid, H5T_NATIVE_INT, sid, sid, H5P_DEFAULT, data); - VRFY((ret>=0), "failed H5Dwrite"); + VRFY((ret >= 0), "failed H5Dwrite"); /* Read 0 elements from dataset */ ret = H5Dread(dsid, H5T_NATIVE_INT, sid, sid, H5P_DEFAULT, data); - VRFY((ret>=0), "failed H5Dread"); + VRFY((ret >= 0), "failed H5Dread"); H5Pclose(dcpl); H5Dclose(dsid); @@ -135,24 +135,25 @@ void zero_dim_dset(void) * * JRM - 8/11/04 */ -void multiple_dset_write(void) +void +multiple_dset_write(void) { - int i, j, n, mpi_size, mpi_rank, size; - hid_t iof, plist, dataset, memspace, filespace; - hid_t dcpl; /* Dataset creation property list */ - hsize_t chunk_origin [DIM]; - hsize_t chunk_dims [DIM], file_dims [DIM]; - hsize_t count[DIM]={1,1}; - double *outme = NULL; - double fill=1.0; /* Fill value */ - char dname [100]; - herr_t ret; - const H5Ptest_param_t *pt; - char *filename; - int ndatasets; - - pt = GetTestParameters(); - filename = pt->name; + int i, j, n, mpi_size, mpi_rank, size; + hid_t iof, plist, dataset, memspace, filespace; + hid_t dcpl; /* Dataset creation property list */ + hsize_t chunk_origin[DIM]; + hsize_t chunk_dims[DIM], file_dims[DIM]; + hsize_t count[DIM] = {1, 1}; + double * outme = NULL; + double fill = 1.0; /* Fill value */ + char dname[100]; + herr_t ret; + const H5Ptest_param_t *pt; + char * filename; + int ndatasets; + + pt = GetTestParameters(); + filename = pt->name; ndatasets = pt->count; size = get_size(); @@ -164,45 +165,45 @@ void multiple_dset_write(void) VRFY((outme != NULL), "HDmalloc succeeded for outme"); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - VRFY((plist>=0), "create_faccess_plist succeeded"); + VRFY((plist >= 0), "create_faccess_plist succeeded"); iof = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); - VRFY((iof>=0), "H5Fcreate succeeded"); + VRFY((iof >= 0), "H5Fcreate succeeded"); ret = H5Pclose(plist); - VRFY((ret>=0), "H5Pclose succeeded"); + VRFY((ret >= 0), "H5Pclose succeeded"); /* decide the hyperslab according to process number. */ get_slab(chunk_origin, chunk_dims, count, file_dims, size); - memspace = H5Screate_simple(DIM, chunk_dims, NULL); + memspace = H5Screate_simple(DIM, chunk_dims, NULL); filespace = H5Screate_simple(DIM, file_dims, NULL); - ret = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); - VRFY((ret>=0), "mdata hyperslab selection"); + ret = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); + VRFY((ret >= 0), "mdata hyperslab selection"); /* Create a dataset creation property list */ dcpl = H5Pcreate(H5P_DATASET_CREATE); - VRFY((dcpl>=0), "dataset creation property list succeeded"); + VRFY((dcpl >= 0), "dataset creation property list succeeded"); ret = H5Pset_fill_value(dcpl, H5T_NATIVE_DOUBLE, &fill); - VRFY((ret>=0), "set fill-value succeeded"); + VRFY((ret >= 0), "set fill-value succeeded"); - for(n = 0; n < ndatasets; n++) { - HDsprintf(dname, "dataset %d", n); - dataset = H5Dcreate2(iof, dname, H5T_NATIVE_DOUBLE, filespace, H5P_DEFAULT, dcpl, H5P_DEFAULT); - VRFY((dataset > 0), dname); + for (n = 0; n < ndatasets; n++) { + HDsprintf(dname, "dataset %d", n); + dataset = H5Dcreate2(iof, dname, H5T_NATIVE_DOUBLE, filespace, H5P_DEFAULT, dcpl, H5P_DEFAULT); + VRFY((dataset > 0), dname); - /* calculate data to write */ - for(i = 0; i < size; i++) - for(j = 0; j < size; j++) - outme [(i * size) + j] = n*1000 + mpi_rank; + /* calculate data to write */ + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) + outme[(i * size) + j] = n * 1000 + mpi_rank; - H5Dwrite(dataset, H5T_NATIVE_DOUBLE, memspace, filespace, H5P_DEFAULT, outme); + H5Dwrite(dataset, H5T_NATIVE_DOUBLE, memspace, filespace, H5P_DEFAULT, outme); - H5Dclose(dataset); + H5Dclose(dataset); #ifdef BARRIER_CHECKS - if(!((n+1) % 10)) { - HDprintf("created %d datasets\n", n+1); - MPI_Barrier(MPI_COMM_WORLD); - } + if (!((n + 1) % 10)) { + HDprintf("created %d datasets\n", n + 1); + MPI_Barrier(MPI_COMM_WORLD); + } #endif /* BARRIER_CHECKS */ } @@ -214,7 +215,6 @@ void multiple_dset_write(void) HDfree(outme); } - /* Example of using PHDF5 to create, write, and read compact dataset. * * Changes: Updated function to use a dynamically calculated size, @@ -223,15 +223,16 @@ void multiple_dset_write(void) * * JRM - 8/11/04 */ -void compact_dataset(void) +void +compact_dataset(void) { - int i, j, mpi_size, mpi_rank, size, err_num=0; - hid_t iof, plist, dcpl, dxpl, dataset, filespace; - hsize_t file_dims [DIM]; - double *outme; - double *inme; - char dname[]="dataset"; - herr_t ret; + int i, j, mpi_size, mpi_rank, size, err_num = 0; + hid_t iof, plist, dcpl, dxpl, dataset, filespace; + hsize_t file_dims[DIM]; + double * outme; + double * inme; + char dname[] = "dataset"; + herr_t ret; const char *filename; #ifdef H5_HAVE_INSTRUMENTED_LIBRARY hbool_t prop_value; @@ -239,7 +240,7 @@ void compact_dataset(void) size = get_size(); - for(i = 0; i < DIM; i++ ) + for (i = 0; i < DIM; i++) file_dims[i] = size; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); @@ -255,14 +256,14 @@ void compact_dataset(void) VRFY((mpi_size <= size), "mpi_size <= size"); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - iof = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); + iof = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); /* Define data space */ filespace = H5Screate_simple(DIM, file_dims, NULL); /* Create a compact dataset */ dcpl = H5Pcreate(H5P_DATASET_CREATE); - VRFY((dcpl>=0), "dataset creation property list succeeded"); + VRFY((dcpl >= 0), "dataset creation property list succeeded"); ret = H5Pset_layout(dcpl, H5D_COMPACT); VRFY((dcpl >= 0), "set property list for compact dataset"); ret = H5Pset_alloc_time(dcpl, H5D_ALLOC_TIME_EARLY); @@ -276,16 +277,15 @@ void compact_dataset(void) VRFY((dxpl >= 0), ""); ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); VRFY((ret >= 0), "H5Pcreate xfer succeeded"); - if(dxfer_coll_type == DXFER_INDEPENDENT_IO) { + if (dxfer_coll_type == DXFER_INDEPENDENT_IO) { ret = H5Pset_dxpl_mpio_collective_opt(dxpl, H5FD_MPIO_INDIVIDUAL_IO); - VRFY((ret>= 0),"set independent IO collectively succeeded"); + VRFY((ret >= 0), "set independent IO collectively succeeded"); } - /* Recalculate data to write. Each process writes the same data. */ - for(i = 0; i < size; i++) - for(j = 0; j < size; j++) - outme[(i * size) + j] =(i + j) * 1000; + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) + outme[(i * size) + j] = (i + j) * 1000; ret = H5Dwrite(dataset, H5T_NATIVE_DOUBLE, H5S_ALL, H5S_ALL, dxpl, outme); VRFY((ret >= 0), "H5Dwrite succeeded"); @@ -298,7 +298,7 @@ void compact_dataset(void) /* Open the file and dataset, read and compare the data. */ plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - iof = H5Fopen(filename, H5F_ACC_RDONLY, plist); + iof = H5Fopen(filename, H5F_ACC_RDONLY, plist); VRFY((iof >= 0), "H5Fopen succeeded"); /* set up the collective transfer properties list */ @@ -306,9 +306,9 @@ void compact_dataset(void) VRFY((dxpl >= 0), ""); ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); VRFY((ret >= 0), "H5Pcreate xfer succeeded"); - if(dxfer_coll_type == DXFER_INDEPENDENT_IO) { - ret = H5Pset_dxpl_mpio_collective_opt(dxpl,H5FD_MPIO_INDIVIDUAL_IO); - VRFY((ret>= 0),"set independent IO collectively succeeded"); + if (dxfer_coll_type == DXFER_INDEPENDENT_IO) { + ret = H5Pset_dxpl_mpio_collective_opt(dxpl, H5FD_MPIO_INDIVIDUAL_IO); + VRFY((ret >= 0), "set independent IO collectively succeeded"); } dataset = H5Dopen2(iof, dname, H5P_DEFAULT); @@ -316,8 +316,8 @@ void compact_dataset(void) #ifdef H5_HAVE_INSTRUMENTED_LIBRARY prop_value = H5D_XFER_COLL_RANK0_BCAST_DEF; - ret = H5Pinsert2(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, H5D_XFER_COLL_RANK0_BCAST_SIZE, &prop_value, - NULL, NULL, NULL, NULL, NULL, NULL); + ret = H5Pinsert2(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, H5D_XFER_COLL_RANK0_BCAST_SIZE, &prop_value, NULL, + NULL, NULL, NULL, NULL, NULL); VRFY((ret >= 0), "H5Pinsert2() succeeded"); #endif /* H5_HAVE_INSTRUMENTED_LIBRARY */ @@ -326,17 +326,19 @@ void compact_dataset(void) #ifdef H5_HAVE_INSTRUMENTED_LIBRARY prop_value = FALSE; - ret = H5Pget(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); + ret = H5Pget(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); VRFY((ret >= 0), "H5Pget succeeded"); - VRFY((prop_value == FALSE && dxfer_coll_type == DXFER_COLLECTIVE_IO),"rank 0 Bcast optimization was performed for a compact dataset"); + VRFY((prop_value == FALSE && dxfer_coll_type == DXFER_COLLECTIVE_IO), + "rank 0 Bcast optimization was performed for a compact dataset"); #endif /* H5_HAVE_INSTRUMENTED_LIBRARY */ /* Verify data value */ - for(i = 0; i < size; i++) - for(j = 0; j < size; j++) - if(inme[(i * size) + j] != outme[(i * size) + j]) - if(err_num++ < MAX_ERR_REPORT || VERBOSE_MED) - HDprintf("Dataset Verify failed at [%d][%d]: expect %f, got %f\n", i, j, outme[(i * size) + j], inme[(i * size) + j]); + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) + if (inme[(i * size) + j] != outme[(i * size) + j]) + if (err_num++ < MAX_ERR_REPORT || VERBOSE_MED) + HDprintf("Dataset Verify failed at [%d][%d]: expect %f, got %f\n", i, j, + outme[(i * size) + j], inme[(i * size) + j]); H5Pclose(plist); H5Pclose(dxpl); @@ -358,16 +360,17 @@ void compact_dataset(void) * * JRM - 8/24/04 */ -void null_dataset(void) +void +null_dataset(void) { - int mpi_size, mpi_rank; - hid_t iof, plist, dxpl, dataset, attr, sid; - unsigned uval=2; /* Buffer for writing to dataset */ - int val=1; /* Buffer for writing to attribute */ - int nelem; - char dname[]="dataset"; - char attr_name[]="attribute"; - herr_t ret; + int mpi_size, mpi_rank; + hid_t iof, plist, dxpl, dataset, attr, sid; + unsigned uval = 2; /* Buffer for writing to dataset */ + int val = 1; /* Buffer for writing to attribute */ + int nelem; + char dname[] = "dataset"; + char attr_name[] = "attribute"; + herr_t ret; const char *filename; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); @@ -376,7 +379,7 @@ void null_dataset(void) filename = GetTestParameters(); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - iof = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); + iof = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); /* Define data space */ sid = H5Screate(H5S_NULL); @@ -394,12 +397,11 @@ void null_dataset(void) VRFY((dxpl >= 0), ""); ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); VRFY((ret >= 0), "H5Pcreate xfer succeeded"); - if(dxfer_coll_type == DXFER_INDEPENDENT_IO) { + if (dxfer_coll_type == DXFER_INDEPENDENT_IO) { ret = H5Pset_dxpl_mpio_collective_opt(dxpl, H5FD_MPIO_INDIVIDUAL_IO); - VRFY((ret>= 0),"set independent IO collectively succeeded"); + VRFY((ret >= 0), "set independent IO collectively succeeded"); } - /* Write "nothing" to the dataset(with type conversion) */ ret = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, &uval); VRFY((ret >= 0), "H5Dwrite succeeded"); @@ -420,7 +422,7 @@ void null_dataset(void) /* Open the file and dataset, read and compare the data. */ plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - iof = H5Fopen(filename, H5F_ACC_RDONLY, plist); + iof = H5Fopen(filename, H5F_ACC_RDONLY, plist); VRFY((iof >= 0), "H5Fopen succeeded"); /* set up the collective transfer properties list */ @@ -428,27 +430,27 @@ void null_dataset(void) VRFY((dxpl >= 0), ""); ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); VRFY((ret >= 0), "H5Pcreate xfer succeeded"); - if(dxfer_coll_type == DXFER_INDEPENDENT_IO) { - ret = H5Pset_dxpl_mpio_collective_opt(dxpl,H5FD_MPIO_INDIVIDUAL_IO); - VRFY((ret>= 0),"set independent IO collectively succeeded"); + if (dxfer_coll_type == DXFER_INDEPENDENT_IO) { + ret = H5Pset_dxpl_mpio_collective_opt(dxpl, H5FD_MPIO_INDIVIDUAL_IO); + VRFY((ret >= 0), "set independent IO collectively succeeded"); } - dataset = H5Dopen2(iof, dname, H5P_DEFAULT); VRFY((dataset >= 0), "H5Dopen2 succeeded"); /* Try reading from the dataset(make certain our buffer is unmodified) */ ret = H5Dread(dataset, H5T_NATIVE_UINT, H5S_ALL, H5S_ALL, dxpl, &uval); - VRFY((ret>=0), "H5Dread"); - VRFY((uval==2), "H5Dread"); + VRFY((ret >= 0), "H5Dread"); + VRFY((uval == 2), "H5Dread"); /* Open the attribute for the dataset */ attr = H5Aopen(dataset, attr_name, H5P_DEFAULT); VRFY((attr >= 0), "H5Aopen"); - /* Try reading from the attribute(make certain our buffer is unmodified) */ ret = H5Aread(attr, H5T_NATIVE_INT, &val); - VRFY((ret>=0), "H5Aread"); - VRFY((val==1), "H5Aread"); + /* Try reading from the attribute(make certain our buffer is unmodified) */ ret = + H5Aread(attr, H5T_NATIVE_INT, &val); + VRFY((ret >= 0), "H5Aread"); + VRFY((val == 1), "H5Aread"); H5Pclose(plist); H5Pclose(dxpl); @@ -470,17 +472,18 @@ void null_dataset(void) * * JRM - 8/11/04 */ -void big_dataset(void) +void +big_dataset(void) { - int mpi_size, mpi_rank; /* MPI info */ - hid_t iof, /* File ID */ - fapl, /* File access property list ID */ - dataset, /* Dataset ID */ - filespace; /* Dataset's dataspace ID */ - hsize_t file_dims [4]; /* Dimensions of dataspace */ - char dname[]="dataset"; /* Name of dataset */ - MPI_Offset file_size; /* Size of file on disk */ - herr_t ret; /* Generic return value */ + int mpi_size, mpi_rank; /* MPI info */ + hid_t iof, /* File ID */ + fapl, /* File access property list ID */ + dataset, /* Dataset ID */ + filespace; /* Dataset's dataspace ID */ + hsize_t file_dims[4]; /* Dimensions of dataspace */ + char dname[] = "dataset"; /* Name of dataset */ + MPI_Offset file_size; /* Size of file on disk */ + herr_t ret; /* Generic return value */ const char *filename; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); @@ -501,11 +504,11 @@ void big_dataset(void) VRFY((iof >= 0), "H5Fcreate succeeded"); /* Define dataspace for 2GB dataspace */ - file_dims[0]= 2; - file_dims[1]= 1024; - file_dims[2]= 1024; - file_dims[3]= 1024; - filespace = H5Screate_simple(4, file_dims, NULL); + file_dims[0] = 2; + file_dims[1] = 1024; + file_dims[2] = 1024; + file_dims[3] = 1024; + filespace = H5Screate_simple(4, file_dims, NULL); VRFY((filespace >= 0), "H5Screate_simple succeeded"); dataset = H5Dcreate2(iof, dname, H5T_NATIVE_UCHAR, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); @@ -530,11 +533,11 @@ void big_dataset(void) VRFY((iof >= 0), "H5Fcreate succeeded"); /* Define dataspace for 4GB dataspace */ - file_dims[0]= 4; - file_dims[1]= 1024; - file_dims[2]= 1024; - file_dims[3]= 1024; - filespace = H5Screate_simple(4, file_dims, NULL); + file_dims[0] = 4; + file_dims[1] = 1024; + file_dims[2] = 1024; + file_dims[3] = 1024; + filespace = H5Screate_simple(4, file_dims, NULL); VRFY((filespace >= 0), "H5Screate_simple succeeded"); dataset = H5Dcreate2(iof, dname, H5T_NATIVE_UCHAR, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); @@ -559,11 +562,11 @@ void big_dataset(void) VRFY((iof >= 0), "H5Fcreate succeeded"); /* Define dataspace for 8GB dataspace */ - file_dims[0]= 8; - file_dims[1]= 1024; - file_dims[2]= 1024; - file_dims[3]= 1024; - filespace = H5Screate_simple(4, file_dims, NULL); + file_dims[0] = 8; + file_dims[1] = 1024; + file_dims[2] = 1024; + file_dims[3] = 1024; + filespace = H5Screate_simple(4, file_dims, NULL); VRFY((filespace >= 0), "H5Screate_simple succeeded"); dataset = H5Dcreate2(iof, dname, H5T_NATIVE_UCHAR, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); @@ -600,25 +603,26 @@ void big_dataset(void) * * JRM - 8/11/04 */ -void dataset_fillvalue(void) +void +dataset_fillvalue(void) { - int mpi_size, mpi_rank; /* MPI info */ - int err_num; /* Number of errors */ - hid_t iof, /* File ID */ - fapl, /* File access property list ID */ - dxpl, /* Data transfer property list ID */ - dataset, /* Dataset ID */ - memspace, /* Memory dataspace ID */ - filespace; /* Dataset's dataspace ID */ - char dname[]="dataset"; /* Name of dataset */ + int mpi_size, mpi_rank; /* MPI info */ + int err_num; /* Number of errors */ + hid_t iof, /* File ID */ + fapl, /* File access property list ID */ + dxpl, /* Data transfer property list ID */ + dataset, /* Dataset ID */ + memspace, /* Memory dataspace ID */ + filespace; /* Dataset's dataspace ID */ + char dname[] = "dataset"; /* Name of dataset */ hsize_t dset_dims[4] = {0, 6, 7, 8}; hsize_t req_start[4] = {0, 0, 0, 0}; hsize_t req_count[4] = {1, 6, 7, 8}; - hsize_t dset_size; /* Dataset size */ - int *rdata, *wdata; /* Buffers for data to read and write */ - int *twdata, *trdata; /* Temporary pointer into buffer */ - int acc, i, ii, j, k, l; /* Local index variables */ - herr_t ret; /* Generic return value */ + hsize_t dset_size; /* Dataset size */ + int * rdata, *wdata; /* Buffers for data to read and write */ + int * twdata, *trdata; /* Temporary pointer into buffer */ + int acc, i, ii, j, k, l; /* Local index variables */ + herr_t ret; /* Generic return value */ const char *filename; #ifdef H5_HAVE_INSTRUMENTED_LIBRARY hbool_t prop_value; @@ -631,13 +635,13 @@ void dataset_fillvalue(void) /* Set the dataset dimension to be one row more than number of processes */ /* and calculate the actual dataset size. */ - dset_dims[0]=mpi_size+1; - dset_size=dset_dims[0]*dset_dims[1]*dset_dims[2]*dset_dims[3]; + dset_dims[0] = mpi_size + 1; + dset_size = dset_dims[0] * dset_dims[1] * dset_dims[2] * dset_dims[3]; /* Allocate space for the buffers */ - rdata=HDmalloc((size_t)(dset_size*sizeof(int))); + rdata = HDmalloc((size_t)(dset_size * sizeof(int))); VRFY((rdata != NULL), "HDcalloc succeeded for read buffer"); - wdata=HDmalloc((size_t)(dset_size*sizeof(int))); + wdata = HDmalloc((size_t)(dset_size * sizeof(int))); VRFY((wdata != NULL), "HDmalloc succeeded for write buffer"); fapl = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); @@ -667,53 +671,54 @@ void dataset_fillvalue(void) VRFY((dxpl >= 0), "H5Pcreate succeeded"); #ifdef H5_HAVE_INSTRUMENTED_LIBRARY - prop_value = H5D_XFER_COLL_RANK0_BCAST_DEF; - ret = H5Pinsert2(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, H5D_XFER_COLL_RANK0_BCAST_SIZE, &prop_value, - NULL, NULL, NULL, NULL, NULL, NULL); - VRFY((ret >= 0),"testing property list inserted succeeded"); + prop_value = H5D_XFER_COLL_RANK0_BCAST_DEF; + ret = H5Pinsert2(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, H5D_XFER_COLL_RANK0_BCAST_SIZE, &prop_value, NULL, + NULL, NULL, NULL, NULL, NULL); + VRFY((ret >= 0), "testing property list inserted succeeded"); #endif /* H5_HAVE_INSTRUMENTED_LIBRARY */ - for(ii = 0; ii < 2; ii++) { + for (ii = 0; ii < 2; ii++) { - if(ii == 0) - ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_INDEPENDENT); - else - ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); - VRFY((ret >= 0), "H5Pset_dxpl_mpio succeeded"); + if (ii == 0) + ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_INDEPENDENT); + else + ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); + VRFY((ret >= 0), "H5Pset_dxpl_mpio succeeded"); - /* set entire read buffer with the constant 2 */ - HDmemset(rdata,2,(size_t)(dset_size*sizeof(int))); + /* set entire read buffer with the constant 2 */ + HDmemset(rdata, 2, (size_t)(dset_size * sizeof(int))); - /* Read the entire dataset back */ - ret = H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, rdata); - VRFY((ret >= 0), "H5Dread succeeded"); + /* Read the entire dataset back */ + ret = H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, rdata); + VRFY((ret >= 0), "H5Dread succeeded"); #ifdef H5_HAVE_INSTRUMENTED_LIBRARY - prop_value = FALSE; - ret = H5Pget(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); - VRFY((ret >= 0), "testing property list get succeeded"); - if(ii == 0) - VRFY((prop_value == FALSE), "correctly handled rank 0 Bcast"); - else - VRFY((prop_value == TRUE), "correctly handled rank 0 Bcast"); + prop_value = FALSE; + ret = H5Pget(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); + VRFY((ret >= 0), "testing property list get succeeded"); + if (ii == 0) + VRFY((prop_value == FALSE), "correctly handled rank 0 Bcast"); + else + VRFY((prop_value == TRUE), "correctly handled rank 0 Bcast"); #endif /* H5_HAVE_INSTRUMENTED_LIBRARY */ - /* Verify all data read are the fill value 0 */ - trdata = rdata; - err_num = 0; - for(i = 0; i < (int)dset_dims[0]; i++) - for(j = 0; j < (int)dset_dims[1]; j++) - for(k = 0; k < (int)dset_dims[2]; k++) - for(l = 0; l < (int)dset_dims[3]; l++, twdata++, trdata++) - if(*trdata != 0) - if(err_num++ < MAX_ERR_REPORT || VERBOSE_MED) - HDprintf("Dataset Verify failed at [%d][%d][%d][%d]: expect 0, got %d\n", i, j, k, l, *trdata); - if(err_num > MAX_ERR_REPORT && !VERBOSE_MED) - HDprintf("[more errors ...]\n"); - if(err_num) { - HDprintf("%d errors found in check_value\n", err_num); - nerrors++; - } + /* Verify all data read are the fill value 0 */ + trdata = rdata; + err_num = 0; + for (i = 0; i < (int)dset_dims[0]; i++) + for (j = 0; j < (int)dset_dims[1]; j++) + for (k = 0; k < (int)dset_dims[2]; k++) + for (l = 0; l < (int)dset_dims[3]; l++, twdata++, trdata++) + if (*trdata != 0) + if (err_num++ < MAX_ERR_REPORT || VERBOSE_MED) + HDprintf("Dataset Verify failed at [%d][%d][%d][%d]: expect 0, got %d\n", i, + j, k, l, *trdata); + if (err_num > MAX_ERR_REPORT && !VERBOSE_MED) + HDprintf("[more errors ...]\n"); + if (err_num) { + HDprintf("%d errors found in check_value\n", err_num); + nerrors++; + } } /* Barrier to ensure all processes have completed the above test. */ @@ -723,26 +728,25 @@ void dataset_fillvalue(void) * Each process writes 1 row of data. Thus last row is not written. */ /* Create hyperslabs in memory and file dataspaces */ - req_start[0]=mpi_rank; - ret = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, req_start, NULL, req_count, NULL); + req_start[0] = mpi_rank; + ret = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, req_start, NULL, req_count, NULL); VRFY((ret >= 0), "H5Sselect_hyperslab succeeded on memory dataspace"); ret = H5Sselect_hyperslab(memspace, H5S_SELECT_SET, req_start, NULL, req_count, NULL); VRFY((ret >= 0), "H5Sselect_hyperslab succeeded on memory dataspace"); ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); VRFY((ret >= 0), "H5Pset_dxpl_mpio succeeded"); - if(dxfer_coll_type == DXFER_INDEPENDENT_IO) { - ret = H5Pset_dxpl_mpio_collective_opt(dxpl,H5FD_MPIO_INDIVIDUAL_IO); - VRFY((ret>= 0),"set independent IO collectively succeeded"); + if (dxfer_coll_type == DXFER_INDEPENDENT_IO) { + ret = H5Pset_dxpl_mpio_collective_opt(dxpl, H5FD_MPIO_INDIVIDUAL_IO); + VRFY((ret >= 0), "set independent IO collectively succeeded"); } - /* Fill write buffer with some values */ - twdata=wdata; - for(i=0, acc=0; i<(int)dset_dims[0]; i++) - for(j=0; j<(int)dset_dims[1]; j++) - for(k=0; k<(int)dset_dims[2]; k++) - for(l=0; l<(int)dset_dims[3]; l++) + twdata = wdata; + for (i = 0, acc = 0; i < (int)dset_dims[0]; i++) + for (j = 0; j < (int)dset_dims[1]; j++) + for (k = 0; k < (int)dset_dims[2]; k++) + for (l = 0; l < (int)dset_dims[3]; l++) *twdata++ = acc++; /* Collectively write a hyperslab of data to the dataset */ @@ -757,60 +761,62 @@ void dataset_fillvalue(void) */ #ifdef H5_HAVE_INSTRUMENTED_LIBRARY - prop_value = H5D_XFER_COLL_RANK0_BCAST_DEF; - ret = H5Pset(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); - VRFY((ret >= 0), " H5Pset succeeded"); + prop_value = H5D_XFER_COLL_RANK0_BCAST_DEF; + ret = H5Pset(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); + VRFY((ret >= 0), " H5Pset succeeded"); #endif /* H5_HAVE_INSTRUMENTED_LIBRARY */ - for(ii = 0; ii < 2; ii++) { + for (ii = 0; ii < 2; ii++) { - if(ii == 0) - ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_INDEPENDENT); - else - ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); - VRFY((ret >= 0), "H5Pset_dxpl_mpio succeeded"); + if (ii == 0) + ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_INDEPENDENT); + else + ret = H5Pset_dxpl_mpio(dxpl, H5FD_MPIO_COLLECTIVE); + VRFY((ret >= 0), "H5Pset_dxpl_mpio succeeded"); - /* set entire read buffer with the constant 2 */ - HDmemset(rdata,2,(size_t)(dset_size*sizeof(int))); + /* set entire read buffer with the constant 2 */ + HDmemset(rdata, 2, (size_t)(dset_size * sizeof(int))); - /* Read the entire dataset back */ - ret = H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, rdata); - VRFY((ret >= 0), "H5Dread succeeded"); + /* Read the entire dataset back */ + ret = H5Dread(dataset, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, dxpl, rdata); + VRFY((ret >= 0), "H5Dread succeeded"); #ifdef H5_HAVE_INSTRUMENTED_LIBRARY - prop_value = FALSE; - ret = H5Pget(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); - VRFY((ret >= 0), "testing property list get succeeded"); - if(ii == 0) - VRFY((prop_value == FALSE), "correctly handled rank 0 Bcast"); - else - VRFY((prop_value == TRUE), "correctly handled rank 0 Bcast"); + prop_value = FALSE; + ret = H5Pget(dxpl, H5D_XFER_COLL_RANK0_BCAST_NAME, &prop_value); + VRFY((ret >= 0), "testing property list get succeeded"); + if (ii == 0) + VRFY((prop_value == FALSE), "correctly handled rank 0 Bcast"); + else + VRFY((prop_value == TRUE), "correctly handled rank 0 Bcast"); #endif /* H5_HAVE_INSTRUMENTED_LIBRARY */ - /* Verify correct data read */ - twdata=wdata; - trdata=rdata; - err_num=0; - for(i=0; i<(int)dset_dims[0]; i++) - for(j=0; j<(int)dset_dims[1]; j++) - for(k=0; k<(int)dset_dims[2]; k++) - for(l=0; l<(int)dset_dims[3]; l++, twdata++, trdata++) - if(i<mpi_size) { - if(*twdata != *trdata ) - if(err_num++ < MAX_ERR_REPORT || VERBOSE_MED) - HDprintf("Dataset Verify failed at [%d][%d][%d][%d]: expect %d, got %d\n", i,j,k,l, *twdata, *trdata); - } /* end if */ - else { - if(*trdata != 0) - if(err_num++ < MAX_ERR_REPORT || VERBOSE_MED) - HDprintf("Dataset Verify failed at [%d][%d][%d][%d]: expect 0, got %d\n", i,j,k,l, *trdata); - } /* end else */ - if(err_num > MAX_ERR_REPORT && !VERBOSE_MED) - HDprintf("[more errors ...]\n"); - if(err_num){ - HDprintf("%d errors found in check_value\n", err_num); - nerrors++; - } + /* Verify correct data read */ + twdata = wdata; + trdata = rdata; + err_num = 0; + for (i = 0; i < (int)dset_dims[0]; i++) + for (j = 0; j < (int)dset_dims[1]; j++) + for (k = 0; k < (int)dset_dims[2]; k++) + for (l = 0; l < (int)dset_dims[3]; l++, twdata++, trdata++) + if (i < mpi_size) { + if (*twdata != *trdata) + if (err_num++ < MAX_ERR_REPORT || VERBOSE_MED) + HDprintf("Dataset Verify failed at [%d][%d][%d][%d]: expect %d, got %d\n", + i, j, k, l, *twdata, *trdata); + } /* end if */ + else { + if (*trdata != 0) + if (err_num++ < MAX_ERR_REPORT || VERBOSE_MED) + HDprintf("Dataset Verify failed at [%d][%d][%d][%d]: expect 0, got %d\n", + i, j, k, l, *trdata); + } /* end else */ + if (err_num > MAX_ERR_REPORT && !VERBOSE_MED) + HDprintf("[more errors ...]\n"); + if (err_num) { + HDprintf("%d errors found in check_value\n", err_num); + nerrors++; + } } /* Close all file objects */ @@ -839,7 +845,8 @@ void dataset_fillvalue(void) } /* combined cngrpw and ingrpr tests because ingrpr reads file created by cngrpw. */ -void collective_group_write_independent_group_read(void) +void +collective_group_write_independent_group_read(void) { collective_group_write(); independent_group_read(); @@ -854,38 +861,39 @@ void collective_group_write_independent_group_read(void) * * JRM - 8/16/04 */ -void collective_group_write(void) +void +collective_group_write(void) { - int mpi_rank, mpi_size, size; - int i, j, m; - char gname[64], dname[32]; - hid_t fid, gid, did, plist, dcpl, memspace, filespace; - DATATYPE *outme = NULL; - hsize_t chunk_origin[DIM]; - hsize_t chunk_dims[DIM], file_dims[DIM], count[DIM]; - hsize_t chunk_size[2]; /* Chunk dimensions - computed shortly */ - herr_t ret1, ret2; + int mpi_rank, mpi_size, size; + int i, j, m; + char gname[64], dname[32]; + hid_t fid, gid, did, plist, dcpl, memspace, filespace; + DATATYPE * outme = NULL; + hsize_t chunk_origin[DIM]; + hsize_t chunk_dims[DIM], file_dims[DIM], count[DIM]; + hsize_t chunk_size[2]; /* Chunk dimensions - computed shortly */ + herr_t ret1, ret2; const H5Ptest_param_t *pt; - char *filename; - int ngroups; + char * filename; + int ngroups; - pt = GetTestParameters(); + pt = GetTestParameters(); filename = pt->name; - ngroups = pt->count; + ngroups = pt->count; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); size = get_size(); - chunk_size[0] =(hsize_t)(size / 2); - chunk_size[1] =(hsize_t)(size / 2); + chunk_size[0] = (hsize_t)(size / 2); + chunk_size[1] = (hsize_t)(size / 2); outme = HDmalloc((size_t)(size * size * sizeof(DATATYPE))); VRFY((outme != NULL), "HDmalloc succeeded for outme"); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - fid = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); + fid = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); H5Pclose(plist); /* decide the hyperslab according to process number. */ @@ -894,24 +902,22 @@ void collective_group_write(void) /* select hyperslab in memory and file spaces. These two operations are * identical since the datasets are the same. */ memspace = H5Screate_simple(DIM, file_dims, NULL); - ret1 = H5Sselect_hyperslab(memspace, H5S_SELECT_SET, chunk_origin, - chunk_dims, count, chunk_dims); - filespace = H5Screate_simple(DIM, file_dims, NULL); - ret2 = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, - chunk_dims, count, chunk_dims); - VRFY((memspace>=0), "memspace"); - VRFY((filespace>=0), "filespace"); - VRFY((ret1>=0), "mgroup memspace selection"); - VRFY((ret2>=0), "mgroup filespace selection"); + ret1 = H5Sselect_hyperslab(memspace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); + filespace = H5Screate_simple(DIM, file_dims, NULL); + ret2 = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); + VRFY((memspace >= 0), "memspace"); + VRFY((filespace >= 0), "filespace"); + VRFY((ret1 >= 0), "mgroup memspace selection"); + VRFY((ret2 >= 0), "mgroup filespace selection"); dcpl = H5Pcreate(H5P_DATASET_CREATE); ret1 = H5Pset_chunk(dcpl, 2, chunk_size); - VRFY((dcpl>=0), "dataset creation property"); - VRFY((ret1>=0), "set chunk for dataset creation property"); + VRFY((dcpl >= 0), "dataset creation property"); + VRFY((ret1 >= 0), "set chunk for dataset creation property"); /* creates ngroups groups under the root group, writes chunked * datasets in parallel. */ - for(m = 0; m < ngroups; m++) { + for (m = 0; m < ngroups; m++) { HDsprintf(gname, "group%d", m); gid = H5Gcreate2(fid, gname, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); VRFY((gid > 0), gname); @@ -920,9 +926,9 @@ void collective_group_write(void) did = H5Dcreate2(gid, dname, H5T_NATIVE_INT, filespace, H5P_DEFAULT, dcpl, H5P_DEFAULT); VRFY((did > 0), dname); - for(i = 0; i < size; i++) - for(j = 0; j < size; j++) - outme[(i * size) + j] =(i + j) * 1000 + mpi_rank; + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) + outme[(i * size) + j] = (i + j) * 1000 + mpi_rank; H5Dwrite(did, H5T_NATIVE_INT, memspace, filespace, H5P_DEFAULT, outme); @@ -930,8 +936,8 @@ void collective_group_write(void) H5Gclose(gid); #ifdef BARRIER_CHECKS - if(!((m+1) % 10)) { - HDprintf("created %d groups\n", m+1); + if (!((m + 1) % 10)) { + HDprintf("created %d groups\n", m + 1); MPI_Barrier(MPI_COMM_WORLD); } #endif /* BARRIER_CHECKS */ @@ -948,17 +954,18 @@ void collective_group_write(void) /* Let two sets of processes open and read different groups and chunked * datasets independently. */ -void independent_group_read(void) +void +independent_group_read(void) { - int mpi_rank, m; - hid_t plist, fid; + int mpi_rank, m; + hid_t plist, fid; const H5Ptest_param_t *pt; - char *filename; - int ngroups; + char * filename; + int ngroups; - pt = GetTestParameters(); + pt = GetTestParameters(); filename = pt->name; - ngroups = pt->count; + ngroups = pt->count; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); @@ -971,11 +978,12 @@ void independent_group_read(void) /* open groups and read datasets. Odd number processes read even number * groups from the end; even number processes read odd number groups * from the beginning. */ - if(mpi_rank%2==0) { - for(m=ngroups-1; m==0; m-=2) + if (mpi_rank % 2 == 0) { + for (m = ngroups - 1; m == 0; m -= 2) group_dataset_read(fid, mpi_rank, m); - } else { - for(m=0; m<ngroups; m+=2) + } + else { + for (m = 0; m < ngroups; m += 2) group_dataset_read(fid, mpi_rank, m); } @@ -997,18 +1005,18 @@ void independent_group_read(void) static void group_dataset_read(hid_t fid, int mpi_rank, int m) { - int ret, i, j, size; - char gname[64], dname[32]; - hid_t gid, did; + int ret, i, j, size; + char gname[64], dname[32]; + hid_t gid, did; DATATYPE *outdata = NULL; - DATATYPE *indata = NULL; + DATATYPE *indata = NULL; size = get_size(); - indata =(DATATYPE*)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); + indata = (DATATYPE *)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); VRFY((indata != NULL), "HDmalloc succeeded for indata"); - outdata =(DATATYPE*)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); + outdata = (DATATYPE *)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); VRFY((outdata != NULL), "HDmalloc succeeded for outdata"); /* open every group under root group. */ @@ -1019,19 +1027,19 @@ group_dataset_read(hid_t fid, int mpi_rank, int m) /* check the data. */ HDsprintf(dname, "dataset%d", m); did = H5Dopen2(gid, dname, H5P_DEFAULT); - VRFY((did>0), dname); + VRFY((did > 0), dname); H5Dread(did, H5T_NATIVE_INT, H5S_ALL, H5S_ALL, H5P_DEFAULT, indata); /* this is the original value */ - for(i=0; i<size; i++) - for(j=0; j<size; j++) { - outdata[(i * size) + j] =(i+j)*1000 + mpi_rank; - } + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) { + outdata[(i * size) + j] = (i + j) * 1000 + mpi_rank; + } /* compare the original value(outdata) to the value in file(indata).*/ ret = check_value(indata, outdata, size); - VRFY((ret==0), "check the data"); + VRFY((ret == 0), "check the data"); H5Dclose(did); H5Gclose(gid); @@ -1072,22 +1080,23 @@ group_dataset_read(hid_t fid, int mpi_rank, int m) * * JRM - 8/16/04 */ -void multiple_group_write(void) +void +multiple_group_write(void) { - int mpi_rank, mpi_size, size; - int m; - char gname[64]; - hid_t fid, gid, plist, memspace, filespace; - hsize_t chunk_origin[DIM]; - hsize_t chunk_dims[DIM], file_dims[DIM], count[DIM]; - herr_t ret; + int mpi_rank, mpi_size, size; + int m; + char gname[64]; + hid_t fid, gid, plist, memspace, filespace; + hsize_t chunk_origin[DIM]; + hsize_t chunk_dims[DIM], file_dims[DIM], count[DIM]; + herr_t ret; const H5Ptest_param_t *pt; - char *filename; - int ngroups; + char * filename; + int ngroups; - pt = GetTestParameters(); + pt = GetTestParameters(); filename = pt->name; - ngroups = pt->count; + ngroups = pt->count; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); @@ -1095,7 +1104,7 @@ void multiple_group_write(void) size = get_size(); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - fid = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); + fid = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist); H5Pclose(plist); /* decide the hyperslab according to process number. */ @@ -1103,38 +1112,36 @@ void multiple_group_write(void) /* select hyperslab in memory and file spaces. These two operations are * identical since the datasets are the same. */ - memspace = H5Screate_simple(DIM, file_dims, NULL); - VRFY((memspace>=0), "memspace"); - ret = H5Sselect_hyperslab(memspace, H5S_SELECT_SET, chunk_origin, - chunk_dims, count, chunk_dims); - VRFY((ret>=0), "mgroup memspace selection"); + memspace = H5Screate_simple(DIM, file_dims, NULL); + VRFY((memspace >= 0), "memspace"); + ret = H5Sselect_hyperslab(memspace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); + VRFY((ret >= 0), "mgroup memspace selection"); - filespace = H5Screate_simple(DIM, file_dims, NULL); - VRFY((filespace>=0), "filespace"); - ret = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, - chunk_dims, count, chunk_dims); - VRFY((ret>=0), "mgroup filespace selection"); + filespace = H5Screate_simple(DIM, file_dims, NULL); + VRFY((filespace >= 0), "filespace"); + ret = H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); + VRFY((ret >= 0), "mgroup filespace selection"); /* creates ngroups groups under the root group, writes datasets in * parallel. */ - for(m = 0; m < ngroups; m++) { + for (m = 0; m < ngroups; m++) { HDsprintf(gname, "group%d", m); gid = H5Gcreate2(fid, gname, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); VRFY((gid > 0), gname); /* create attribute for these groups. */ - write_attribute(gid, is_group, m); + write_attribute(gid, is_group, m); - if(m != 0) - write_dataset(memspace, filespace, gid); + if (m != 0) + write_dataset(memspace, filespace, gid); H5Gclose(gid); #ifdef BARRIER_CHECKS - if(!((m+1) % 10)) { - HDprintf("created %d groups\n", m+1); + if (!((m + 1) % 10)) { + HDprintf("created %d groups\n", m + 1); MPI_Barrier(MPI_COMM_WORLD); - } + } #endif /* BARRIER_CHECKS */ } @@ -1142,14 +1149,14 @@ void multiple_group_write(void) gid = H5Gopen2(fid, "group0", H5P_DEFAULT); create_group_recursive(memspace, filespace, gid, 0); ret = H5Gclose(gid); - VRFY((ret>=0), "H5Gclose"); + VRFY((ret >= 0), "H5Gclose"); ret = H5Sclose(filespace); - VRFY((ret>=0), "H5Sclose"); + VRFY((ret >= 0), "H5Sclose"); ret = H5Sclose(memspace); - VRFY((ret>=0), "H5Sclose"); + VRFY((ret >= 0), "H5Sclose"); ret = H5Fclose(fid); - VRFY((ret>=0), "H5Fclose"); + VRFY((ret >= 0), "H5Fclose"); } /* @@ -1179,21 +1186,21 @@ write_dataset(hid_t memspace, hid_t filespace, hid_t gid) outme = HDmalloc((size_t)(size * size * sizeof(double))); VRFY((outme != NULL), "HDmalloc succeeded for outme"); - for(n = 0; n < NDATASET; n++) { - HDsprintf(dname, "dataset%d", n); - did = H5Dcreate2(gid, dname, H5T_NATIVE_INT, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); - VRFY((did > 0), dname); + for (n = 0; n < NDATASET; n++) { + HDsprintf(dname, "dataset%d", n); + did = H5Dcreate2(gid, dname, H5T_NATIVE_INT, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); + VRFY((did > 0), dname); - for(i = 0; i < size; i++) - for(j = 0; j < size; j++) + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) outme[(i * size) + j] = n * 1000 + mpi_rank; - H5Dwrite(did, H5T_NATIVE_INT, memspace, filespace, H5P_DEFAULT, outme); + H5Dwrite(did, H5T_NATIVE_INT, memspace, filespace, H5P_DEFAULT, outme); - /* create attribute for these datasets.*/ - write_attribute(did, is_dset, n); + /* create attribute for these datasets.*/ + write_attribute(did, is_dset, n); - H5Dclose(did); + H5Dclose(did); } HDfree(outme); } @@ -1205,30 +1212,30 @@ write_dataset(hid_t memspace, hid_t filespace, hid_t gid) static void create_group_recursive(hid_t memspace, hid_t filespace, hid_t gid, int counter) { - hid_t child_gid; - int mpi_rank; - char gname[64]; + hid_t child_gid; + int mpi_rank; + char gname[64]; - MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); + MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); #ifdef BARRIER_CHECKS - if(!((counter+1) % 10)) { - HDprintf("created %dth child groups\n", counter+1); + if (!((counter + 1) % 10)) { + HDprintf("created %dth child groups\n", counter + 1); MPI_Barrier(MPI_COMM_WORLD); - } + } #endif /* BARRIER_CHECKS */ - HDsprintf(gname, "%dth_child_group", counter+1); - child_gid = H5Gcreate2(gid, gname, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); - VRFY((child_gid > 0), gname); + HDsprintf(gname, "%dth_child_group", counter + 1); + child_gid = H5Gcreate2(gid, gname, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); + VRFY((child_gid > 0), gname); - /* write datasets in parallel. */ - write_dataset(memspace, filespace, gid); + /* write datasets in parallel. */ + write_dataset(memspace, filespace, gid); - if(counter < GROUP_DEPTH ) - create_group_recursive(memspace, filespace, child_gid, counter+1); + if (counter < GROUP_DEPTH) + create_group_recursive(memspace, filespace, child_gid, counter + 1); - H5Gclose(child_gid); + H5Gclose(child_gid); } /* @@ -1241,21 +1248,22 @@ create_group_recursive(hid_t memspace, hid_t filespace, hid_t gid, int counter) * * JRM - 8/11/04 */ -void multiple_group_read(void) +void +multiple_group_read(void) { - int mpi_rank, mpi_size, error_num, size; - int m; - char gname[64]; - hid_t plist, fid, gid, memspace, filespace; - hsize_t chunk_origin[DIM]; - hsize_t chunk_dims[DIM], file_dims[DIM], count[DIM]; + int mpi_rank, mpi_size, error_num, size; + int m; + char gname[64]; + hid_t plist, fid, gid, memspace, filespace; + hsize_t chunk_origin[DIM]; + hsize_t chunk_dims[DIM], file_dims[DIM], count[DIM]; const H5Ptest_param_t *pt; - char *filename; - int ngroups; + char * filename; + int ngroups; - pt = GetTestParameters(); + pt = GetTestParameters(); filename = pt->name; - ngroups = pt->count; + ngroups = pt->count; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); @@ -1263,54 +1271,51 @@ void multiple_group_read(void) size = get_size(); plist = create_faccess_plist(MPI_COMM_WORLD, MPI_INFO_NULL, facc_type); - fid = H5Fopen(filename, H5F_ACC_RDONLY, plist); + fid = H5Fopen(filename, H5F_ACC_RDONLY, plist); H5Pclose(plist); /* decide hyperslab for each process */ get_slab(chunk_origin, chunk_dims, count, file_dims, size); /* select hyperslab for memory and file space */ - memspace = H5Screate_simple(DIM, file_dims, NULL); - H5Sselect_hyperslab(memspace, H5S_SELECT_SET, chunk_origin, chunk_dims, - count, chunk_dims); + memspace = H5Screate_simple(DIM, file_dims, NULL); + H5Sselect_hyperslab(memspace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); filespace = H5Screate_simple(DIM, file_dims, NULL); - H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, chunk_dims, - count, chunk_dims); + H5Sselect_hyperslab(filespace, H5S_SELECT_SET, chunk_origin, chunk_dims, count, chunk_dims); /* open every group under root group. */ - for(m=0; m<ngroups; m++) { + for (m = 0; m < ngroups; m++) { HDsprintf(gname, "group%d", m); gid = H5Gopen2(fid, gname, H5P_DEFAULT); VRFY((gid > 0), gname); /* check the data. */ - if(m != 0) - if((error_num = read_dataset(memspace, filespace, gid))>0) - nerrors += error_num; + if (m != 0) + if ((error_num = read_dataset(memspace, filespace, gid)) > 0) + nerrors += error_num; /* check attribute.*/ error_num = 0; - if((error_num = read_attribute(gid, is_group, m))>0 ) - nerrors += error_num; + if ((error_num = read_attribute(gid, is_group, m)) > 0) + nerrors += error_num; H5Gclose(gid); #ifdef BARRIER_CHECKS - if(!((m+1)%10)) + if (!((m + 1) % 10)) MPI_Barrier(MPI_COMM_WORLD); #endif /* BARRIER_CHECKS */ } /* open all the groups in vertical direction. */ gid = H5Gopen2(fid, "group0", H5P_DEFAULT); - VRFY((gid>0), "group0"); + VRFY((gid > 0), "group0"); recursive_read_group(memspace, filespace, gid, 0); H5Gclose(gid); H5Sclose(filespace); H5Sclose(memspace); H5Fclose(fid); - } /* @@ -1326,7 +1331,7 @@ void multiple_group_read(void) static int read_dataset(hid_t memspace, hid_t filespace, hid_t gid) { - int i, j, n, mpi_rank, mpi_size, size, attr_errors=0, vrfy_errors=0; + int i, j, n, mpi_rank, mpi_size, size, attr_errors = 0, vrfy_errors = 0; char dname[32]; DATATYPE *outdata = NULL, *indata = NULL; hid_t did; @@ -1336,32 +1341,32 @@ read_dataset(hid_t memspace, hid_t filespace, hid_t gid) size = get_size(); - indata =(DATATYPE*)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); + indata = (DATATYPE *)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); VRFY((indata != NULL), "HDmalloc succeeded for indata"); - outdata =(DATATYPE*)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); + outdata = (DATATYPE *)HDmalloc((size_t)(size * size * sizeof(DATATYPE))); VRFY((outdata != NULL), "HDmalloc succeeded for outdata"); - for(n=0; n<NDATASET; n++) { + for (n = 0; n < NDATASET; n++) { HDsprintf(dname, "dataset%d", n); did = H5Dopen2(gid, dname, H5P_DEFAULT); - VRFY((did>0), dname); + VRFY((did > 0), dname); H5Dread(did, H5T_NATIVE_INT, memspace, filespace, H5P_DEFAULT, indata); /* this is the original value */ - for(i=0; i<size; i++) - for(j=0; j<size; j++) { - *outdata = n*1000 + mpi_rank; - outdata++; - } + for (i = 0; i < size; i++) + for (j = 0; j < size; j++) { + *outdata = n * 1000 + mpi_rank; + outdata++; + } outdata -= size * size; /* compare the original value(outdata) to the value in file(indata).*/ vrfy_errors = check_value(indata, outdata, size); /* check attribute.*/ - if((attr_errors = read_attribute(did, is_dset, n))>0 ) + if ((attr_errors = read_attribute(did, is_dset, n)) > 0) vrfy_errors += attr_errors; H5Dclose(did); @@ -1381,23 +1386,23 @@ static void recursive_read_group(hid_t memspace, hid_t filespace, hid_t gid, int counter) { hid_t child_gid; - int mpi_rank, err_num=0; + int mpi_rank, err_num = 0; char gname[64]; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); #ifdef BARRIER_CHECKS - if((counter+1) % 10) + if ((counter + 1) % 10) MPI_Barrier(MPI_COMM_WORLD); #endif /* BARRIER_CHECKS */ - if((err_num = read_dataset(memspace, filespace, gid)) ) + if ((err_num = read_dataset(memspace, filespace, gid))) nerrors += err_num; - if(counter < GROUP_DEPTH ) { - HDsprintf(gname, "%dth_child_group", counter+1); + if (counter < GROUP_DEPTH) { + HDsprintf(gname, "%dth_child_group", counter + 1); child_gid = H5Gopen2(gid, gname, H5P_DEFAULT); - VRFY((child_gid>0), gname); - recursive_read_group(memspace, filespace, child_gid, counter+1); + VRFY((child_gid > 0), gname); + recursive_read_group(memspace, filespace, child_gid, counter + 1); H5Gclose(child_gid); } } @@ -1409,23 +1414,23 @@ static void write_attribute(hid_t obj_id, int this_type, int num) { hid_t sid, aid; - hsize_t dspace_dims[1]={8}; - int i, mpi_rank, attr_data[8], dspace_rank=1; + hsize_t dspace_dims[1] = {8}; + int i, mpi_rank, attr_data[8], dspace_rank = 1; char attr_name[32]; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); - if(this_type == is_group) { + if (this_type == is_group) { HDsprintf(attr_name, "Group Attribute %d", num); sid = H5Screate(H5S_SCALAR); aid = H5Acreate2(obj_id, attr_name, H5T_NATIVE_INT, sid, H5P_DEFAULT, H5P_DEFAULT); - H5Awrite(aid, H5T_NATIVE_INT, &num); + H5Awrite(aid, H5T_NATIVE_INT, &num); H5Aclose(aid); H5Sclose(sid); } /* end if */ - else if(this_type == is_dset) { + else if (this_type == is_dset) { HDsprintf(attr_name, "Dataset Attribute %d", num); - for(i=0; i<8; i++) + for (i = 0; i < 8; i++) attr_data[i] = i; sid = H5Screate_simple(dspace_rank, dspace_dims, NULL); aid = H5Acreate2(obj_id, attr_name, H5T_NATIVE_INT, sid, H5P_DEFAULT, H5P_DEFAULT); @@ -1433,35 +1438,34 @@ write_attribute(hid_t obj_id, int this_type, int num) H5Aclose(aid); H5Sclose(sid); } /* end else-if */ - } /* Read and verify attribute for group or dataset. */ static int read_attribute(hid_t obj_id, int this_type, int num) { - hid_t aid; - hsize_t group_block[2]={1,1}, dset_block[2]={1, 8}; - int i, mpi_rank, in_num, in_data[8], out_data[8], vrfy_errors = 0; - char attr_name[32]; + hid_t aid; + hsize_t group_block[2] = {1, 1}, dset_block[2] = {1, 8}; + int i, mpi_rank, in_num, in_data[8], out_data[8], vrfy_errors = 0; + char attr_name[32]; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); - if(this_type == is_group) { + if (this_type == is_group) { HDsprintf(attr_name, "Group Attribute %d", num); aid = H5Aopen(obj_id, attr_name, H5P_DEFAULT); - if(MAINPROCESS) { + if (MAINPROCESS) { H5Aread(aid, H5T_NATIVE_INT, &in_num); - vrfy_errors = dataset_vrfy(NULL, NULL, NULL, group_block, &in_num, &num); + vrfy_errors = dataset_vrfy(NULL, NULL, NULL, group_block, &in_num, &num); } H5Aclose(aid); } - else if(this_type == is_dset) { + else if (this_type == is_dset) { HDsprintf(attr_name, "Dataset Attribute %d", num); - for(i=0; i<8; i++) + for (i = 0; i < 8; i++) out_data[i] = i; aid = H5Aopen(obj_id, attr_name, H5P_DEFAULT); - if(MAINPROCESS) { + if (MAINPROCESS) { H5Aread(aid, H5T_NATIVE_INT, in_data); vrfy_errors = dataset_vrfy(NULL, NULL, NULL, dset_block, in_data, out_data); } @@ -1483,27 +1487,29 @@ read_attribute(hid_t obj_id, int this_type, int num) static int check_value(DATATYPE *indata, DATATYPE *outdata, int size) { - int mpi_rank, mpi_size, err_num=0; - hsize_t i, j; - hsize_t chunk_origin[DIM]; - hsize_t chunk_dims[DIM], count[DIM]; + int mpi_rank, mpi_size, err_num = 0; + hsize_t i, j; + hsize_t chunk_origin[DIM]; + hsize_t chunk_dims[DIM], count[DIM]; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); get_slab(chunk_origin, chunk_dims, count, NULL, size); - indata += chunk_origin[0]*size; - outdata += chunk_origin[0]*size; - for(i=chunk_origin[0]; i<(chunk_origin[0]+chunk_dims[0]); i++) - for(j=chunk_origin[1]; j<(chunk_origin[1]+chunk_dims[1]); j++) { - if(*indata != *outdata ) - if(err_num++ < MAX_ERR_REPORT || VERBOSE_MED) - HDprintf("Dataset Verify failed at [%lu][%lu](row %lu, col%lu): expect %d, got %d\n",(unsigned long)i,(unsigned long)j,(unsigned long)i,(unsigned long)j, *outdata, *indata); + indata += chunk_origin[0] * size; + outdata += chunk_origin[0] * size; + for (i = chunk_origin[0]; i < (chunk_origin[0] + chunk_dims[0]); i++) + for (j = chunk_origin[1]; j < (chunk_origin[1] + chunk_dims[1]); j++) { + if (*indata != *outdata) + if (err_num++ < MAX_ERR_REPORT || VERBOSE_MED) + HDprintf("Dataset Verify failed at [%lu][%lu](row %lu, col%lu): expect %d, got %d\n", + (unsigned long)i, (unsigned long)j, (unsigned long)i, (unsigned long)j, *outdata, + *indata); } - if(err_num > MAX_ERR_REPORT && !VERBOSE_MED) + if (err_num > MAX_ERR_REPORT && !VERBOSE_MED) HDprintf("[more errors ...]\n"); - if(err_num) + if (err_num) HDprintf("%d errors found in check_value\n", err_num); return err_num; } @@ -1518,25 +1524,24 @@ check_value(DATATYPE *indata, DATATYPE *outdata, int size) */ static void -get_slab(hsize_t chunk_origin[], hsize_t chunk_dims[], hsize_t count[], - hsize_t file_dims[], int size) +get_slab(hsize_t chunk_origin[], hsize_t chunk_dims[], hsize_t count[], hsize_t file_dims[], int size) { int mpi_rank, mpi_size; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); MPI_Comm_size(MPI_COMM_WORLD, &mpi_size); - if(chunk_origin != NULL) { - chunk_origin[0] = mpi_rank *(size/mpi_size); + if (chunk_origin != NULL) { + chunk_origin[0] = mpi_rank * (size / mpi_size); chunk_origin[1] = 0; } - if(chunk_dims != NULL) { - chunk_dims[0] = size/mpi_size; - chunk_dims[1] = size; + if (chunk_dims != NULL) { + chunk_dims[0] = size / mpi_size; + chunk_dims[1] = size; } - if(file_dims != NULL) + if (file_dims != NULL) file_dims[0] = file_dims[1] = size; - if(count != NULL) + if (count != NULL) count[0] = count[1] = 1; } @@ -1563,24 +1568,24 @@ get_slab(hsize_t chunk_origin[], hsize_t chunk_dims[], hsize_t count[], #define N 4 -void io_mode_confusion(void) +void +io_mode_confusion(void) { /* * HDF5 APIs definitions */ - const int rank = 1; + const int rank = 1; const char *dataset_name = "IntArray"; - hid_t file_id, dset_id; /* file and dataset identifiers */ - hid_t filespace, memspace; /* file and memory dataspace */ - /* identifiers */ - hsize_t dimsf[1]; /* dataset dimensions */ - int data[N] = {1}; /* pointer to data buffer to write */ - hsize_t coord[N] = {0L,1L,2L,3L}; - hid_t plist_id; /* property list identifier */ - herr_t status; - + hid_t file_id, dset_id; /* file and dataset identifiers */ + hid_t filespace, memspace; /* file and memory dataspace */ + /* identifiers */ + hsize_t dimsf[1]; /* dataset dimensions */ + int data[N] = {1}; /* pointer to data buffer to write */ + hsize_t coord[N] = {0L, 1L, 2L, 3L}; + hid_t plist_id; /* property list identifier */ + herr_t status; /* * MPI variables @@ -1588,18 +1593,16 @@ void io_mode_confusion(void) int mpi_size, mpi_rank; - /* * test bed related variables */ - const char * fcn_name = "io_mode_confusion"; - const hbool_t verbose = FALSE; - const H5Ptest_param_t * pt; - char * filename; - + const char * fcn_name = "io_mode_confusion"; + const hbool_t verbose = FALSE; + const H5Ptest_param_t *pt; + char * filename; - pt = GetTestParameters(); + pt = GetTestParameters(); filename = pt->name; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); @@ -1609,183 +1612,154 @@ void io_mode_confusion(void) * Set up file access property list with parallel I/O access */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Setting up property list.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Setting up property list.\n", mpi_rank, fcn_name); plist_id = H5Pcreate(H5P_FILE_ACCESS); VRFY((plist_id != -1), "H5Pcreate() failed"); status = H5Pset_fapl_mpio(plist_id, MPI_COMM_WORLD, MPI_INFO_NULL); - VRFY((status >= 0 ), "H5Pset_fapl_mpio() failed"); - + VRFY((status >= 0), "H5Pset_fapl_mpio() failed"); /* * Create a new file collectively and release property list identifier. */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Creating new file.\n", mpi_rank, fcn_name); file_id = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, plist_id); - VRFY((file_id >= 0 ), "H5Fcreate() failed"); + VRFY((file_id >= 0), "H5Fcreate() failed"); status = H5Pclose(plist_id); - VRFY((status >= 0 ), "H5Pclose() failed"); - + VRFY((status >= 0), "H5Pclose() failed"); /* * Create the dataspace for the dataset. */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Creating the dataspace for the dataset.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Creating the dataspace for the dataset.\n", mpi_rank, fcn_name); - dimsf[0] = N; + dimsf[0] = N; filespace = H5Screate_simple(rank, dimsf, NULL); - VRFY((filespace >= 0 ), "H5Screate_simple() failed."); - + VRFY((filespace >= 0), "H5Screate_simple() failed."); /* * Create the dataset with default properties and close filespace. */ - if(verbose ) - HDfprintf(stdout, - "%0d:%s: Creating the dataset, and closing filespace.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Creating the dataset, and closing filespace.\n", mpi_rank, fcn_name); - dset_id = H5Dcreate2(file_id, dataset_name, H5T_NATIVE_INT, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); - VRFY((dset_id >= 0 ), "H5Dcreate2() failed"); + dset_id = + H5Dcreate2(file_id, dataset_name, H5T_NATIVE_INT, filespace, H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); + VRFY((dset_id >= 0), "H5Dcreate2() failed"); status = H5Sclose(filespace); - VRFY((status >= 0 ), "H5Sclose() failed"); - + VRFY((status >= 0), "H5Sclose() failed"); - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Screate_simple().\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Screate_simple().\n", mpi_rank, fcn_name); memspace = H5Screate_simple(rank, dimsf, NULL); - VRFY((memspace >= 0 ), "H5Screate_simple() failed."); - + VRFY((memspace >= 0), "H5Screate_simple() failed."); - if(mpi_rank == 0 ) { - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Sselect_all(memspace).\n", - mpi_rank, fcn_name); + if (mpi_rank == 0) { + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Sselect_all(memspace).\n", mpi_rank, fcn_name); status = H5Sselect_all(memspace); - VRFY((status >= 0 ), "H5Sselect_all() failed"); - } else { - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Sselect_none(memspace).\n", - mpi_rank, fcn_name); + VRFY((status >= 0), "H5Sselect_all() failed"); + } + else { + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Sselect_none(memspace).\n", mpi_rank, fcn_name); status = H5Sselect_none(memspace); - VRFY((status >= 0 ), "H5Sselect_none() failed"); + VRFY((status >= 0), "H5Sselect_none() failed"); } - - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling MPI_Barrier().\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling MPI_Barrier().\n", mpi_rank, fcn_name); MPI_Barrier(MPI_COMM_WORLD); - - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Dget_space().\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Dget_space().\n", mpi_rank, fcn_name); filespace = H5Dget_space(dset_id); - VRFY((filespace >= 0 ), "H5Dget_space() failed"); - + VRFY((filespace >= 0), "H5Dget_space() failed"); /* select all */ - if(mpi_rank == 0 ) { - if(verbose ) - HDfprintf(stdout, - "%0d:%s: Calling H5Sselect_elements() -- set up hang?\n", - mpi_rank, fcn_name); + if (mpi_rank == 0) { + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Sselect_elements() -- set up hang?\n", mpi_rank, fcn_name); status = H5Sselect_elements(filespace, H5S_SELECT_SET, N, (const hsize_t *)&coord); - VRFY((status >= 0 ), "H5Sselect_elements() failed"); - } else { /* select nothing */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Sselect_none().\n", - mpi_rank, fcn_name); + VRFY((status >= 0), "H5Sselect_elements() failed"); + } + else { /* select nothing */ + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Sselect_none().\n", mpi_rank, fcn_name); status = H5Sselect_none(filespace); - VRFY((status >= 0 ), "H5Sselect_none() failed"); + VRFY((status >= 0), "H5Sselect_none() failed"); } - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling MPI_Barrier().\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling MPI_Barrier().\n", mpi_rank, fcn_name); MPI_Barrier(MPI_COMM_WORLD); - - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Calling H5Pcreate().\n", mpi_rank, fcn_name); plist_id = H5Pcreate(H5P_DATASET_XFER); - VRFY((plist_id != -1 ), "H5Pcreate() failed"); - + VRFY((plist_id != -1), "H5Pcreate() failed"); - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Pset_dxpl_mpio().\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Pset_dxpl_mpio().\n", mpi_rank, fcn_name); status = H5Pset_dxpl_mpio(plist_id, H5FD_MPIO_COLLECTIVE); - VRFY((status >= 0 ), "H5Pset_dxpl_mpio() failed"); - if(dxfer_coll_type == DXFER_INDEPENDENT_IO) { + VRFY((status >= 0), "H5Pset_dxpl_mpio() failed"); + if (dxfer_coll_type == DXFER_INDEPENDENT_IO) { status = H5Pset_dxpl_mpio_collective_opt(plist_id, H5FD_MPIO_INDIVIDUAL_IO); - VRFY((status>= 0),"set independent IO collectively succeeded"); + VRFY((status >= 0), "set independent IO collectively succeeded"); } + if (verbose) + HDfprintf(stdout, "%0d:%s: Calling H5Dwrite() -- hang here?.\n", mpi_rank, fcn_name); + status = H5Dwrite(dset_id, H5T_NATIVE_INT, memspace, filespace, plist_id, data); - - if(verbose ) - HDfprintf(stdout, "%0d:%s: Calling H5Dwrite() -- hang here?.\n", - mpi_rank, fcn_name); - - status = H5Dwrite(dset_id, H5T_NATIVE_INT, memspace, filespace, - plist_id, data); - - if(verbose ) - HDfprintf(stdout, "%0d:%s: Returned from H5Dwrite(), status=%d.\n", - mpi_rank, fcn_name, status); - VRFY((status >= 0 ), "H5Dwrite() failed"); + if (verbose) + HDfprintf(stdout, "%0d:%s: Returned from H5Dwrite(), status=%d.\n", mpi_rank, fcn_name, status); + VRFY((status >= 0), "H5Dwrite() failed"); /* * Close/release resources. */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Cleaning up from test.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Cleaning up from test.\n", mpi_rank, fcn_name); status = H5Dclose(dset_id); - VRFY((status >= 0 ), "H5Dclose() failed"); + VRFY((status >= 0), "H5Dclose() failed"); status = H5Sclose(filespace); - VRFY((status >= 0 ), "H5Dclose() failed"); + VRFY((status >= 0), "H5Dclose() failed"); status = H5Sclose(memspace); - VRFY((status >= 0 ), "H5Sclose() failed"); + VRFY((status >= 0), "H5Sclose() failed"); status = H5Pclose(plist_id); - VRFY((status >= 0 ), "H5Pclose() failed"); + VRFY((status >= 0), "H5Pclose() failed"); status = H5Fclose(file_id); - VRFY((status >= 0 ), "H5Fclose() failed"); - + VRFY((status >= 0), "H5Fclose() failed"); - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Done.\n", mpi_rank, fcn_name); return; @@ -1839,62 +1813,46 @@ void io_mode_confusion(void) * the correctness of the writer. AKC -- 2010/10/27 */ -#define NUM_DATA_SETS 4 -#define LOCAL_DATA_SIZE 4 -#define LARGE_ATTR_SIZE 256 +#define NUM_DATA_SETS 4 +#define LOCAL_DATA_SIZE 4 +#define LARGE_ATTR_SIZE 256 /* Since all even and odd processes are split into writer and reader comm * respectively, process 0 and 1 in COMM_WORLD become the root process of * the writer and reader comm respectively. */ -#define Writer_Root 0 -#define Reader_Root 1 -#define Reader_wait(mpi_err, xsteps) \ - mpi_err = MPI_Bcast(&xsteps, 1, MPI_INT, Writer_Root, MPI_COMM_WORLD) -#define Reader_result(mpi_err, xsteps_done) \ +#define Writer_Root 0 +#define Reader_Root 1 +#define Reader_wait(mpi_err, xsteps) mpi_err = MPI_Bcast(&xsteps, 1, MPI_INT, Writer_Root, MPI_COMM_WORLD) +#define Reader_result(mpi_err, xsteps_done) \ mpi_err = MPI_Bcast(&xsteps_done, 1, MPI_INT, Reader_Root, MPI_COMM_WORLD) -#define Reader_check(mpi_err, xsteps, xsteps_done) \ - { Reader_wait(mpi_err, xsteps); \ - Reader_result(mpi_err, xsteps_done);} +#define Reader_check(mpi_err, xsteps, xsteps_done) \ + { \ + Reader_wait(mpi_err, xsteps); \ + Reader_result(mpi_err, xsteps_done); \ + } /* object names used by both rr_obj_hdr_flush_confusion and * rr_obj_hdr_flush_confusion_reader. */ -const char * dataset_name[NUM_DATA_SETS] = - { - "dataset_0", - "dataset_1", - "dataset_2", - "dataset_3" - }; -const char * att_name[NUM_DATA_SETS] = - { - "attribute_0", - "attribute_1", - "attribute_2", - "attribute_3" - }; -const char * lg_att_name[NUM_DATA_SETS] = - { - "large_attribute_0", - "large_attribute_1", - "large_attribute_2", - "large_attribute_3" - }; - -void rr_obj_hdr_flush_confusion(void) +const char *dataset_name[NUM_DATA_SETS] = {"dataset_0", "dataset_1", "dataset_2", "dataset_3"}; +const char *att_name[NUM_DATA_SETS] = {"attribute_0", "attribute_1", "attribute_2", "attribute_3"}; +const char *lg_att_name[NUM_DATA_SETS] = {"large_attribute_0", "large_attribute_1", "large_attribute_2", + "large_attribute_3"}; + +void +rr_obj_hdr_flush_confusion(void) { /* MPI variables */ /* private communicator size and rank */ - int mpi_size; - int mpi_rank; - int mrc; /* mpi error code */ - int is_reader; /* 1 for reader process; 0 for writer process. */ + int mpi_size; + int mpi_rank; + int mrc; /* mpi error code */ + int is_reader; /* 1 for reader process; 0 for writer process. */ MPI_Comm comm; - /* test bed related variables */ - const char * fcn_name = "rr_obj_hdr_flush_confusion"; - const hbool_t verbose = FALSE; + const char * fcn_name = "rr_obj_hdr_flush_confusion"; + const hbool_t verbose = FALSE; /* Create two new private communicators from MPI_COMM_WORLD. * Even and odd ranked processes go to comm_writers and comm_readers @@ -1906,8 +1864,8 @@ void rr_obj_hdr_flush_confusion(void) HDassert(mpi_size > 2); is_reader = mpi_rank % 2; - mrc = MPI_Comm_split(MPI_COMM_WORLD, is_reader, mpi_rank, &comm); - VRFY((mrc==MPI_SUCCESS), "MPI_Comm_split"); + mrc = MPI_Comm_split(MPI_COMM_WORLD, is_reader, mpi_rank, &comm); + VRFY((mrc == MPI_SUCCESS), "MPI_Comm_split"); /* The reader proocesses branches off to do reading * while the writer processes continues to do writing @@ -1917,32 +1875,33 @@ void rr_obj_hdr_flush_confusion(void) * step. When all steps are done, they inform readers to end. */ if (is_reader) - rr_obj_hdr_flush_confusion_reader(comm); + rr_obj_hdr_flush_confusion_reader(comm); else - rr_obj_hdr_flush_confusion_writer(comm); + rr_obj_hdr_flush_confusion_writer(comm); MPI_Comm_free(&comm); - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Done.\n", mpi_rank, fcn_name); return; } /* rr_obj_hdr_flush_confusion() */ -void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) +void +rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) { - int i; - int j; - hid_t file_id = -1; - hid_t fapl_id = -1; - hid_t dxpl_id = -1; - hid_t att_id[NUM_DATA_SETS]; - hid_t att_space[NUM_DATA_SETS]; - hid_t lg_att_id[NUM_DATA_SETS]; - hid_t lg_att_space[NUM_DATA_SETS]; - hid_t disk_space[NUM_DATA_SETS]; - hid_t mem_space[NUM_DATA_SETS]; - hid_t dataset[NUM_DATA_SETS]; + int i; + int j; + hid_t file_id = -1; + hid_t fapl_id = -1; + hid_t dxpl_id = -1; + hid_t att_id[NUM_DATA_SETS]; + hid_t att_space[NUM_DATA_SETS]; + hid_t lg_att_id[NUM_DATA_SETS]; + hid_t lg_att_space[NUM_DATA_SETS]; + hid_t disk_space[NUM_DATA_SETS]; + hid_t mem_space[NUM_DATA_SETS]; + hid_t dataset[NUM_DATA_SETS]; hsize_t att_size[1]; hsize_t lg_att_size[1]; hsize_t disk_count[1]; @@ -1951,10 +1910,10 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) hsize_t mem_count[1]; hsize_t mem_size[1]; hsize_t mem_start[1]; - herr_t err; - double data[LOCAL_DATA_SIZE]; - double att[LOCAL_DATA_SIZE]; - double lg_att[LARGE_ATTR_SIZE]; + herr_t err; + double data[LOCAL_DATA_SIZE]; + double att[LOCAL_DATA_SIZE]; + double lg_att[LARGE_ATTR_SIZE]; /* MPI variables */ /* world communication size and rank */ @@ -1963,22 +1922,22 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) /* private communicator size and rank */ int mpi_size; int mpi_rank; - int mrc; /* mpi error code */ + int mrc; /* mpi error code */ /* steps to verify and have been verified */ - int steps = 0; + int steps = 0; int steps_done = 0; /* test bed related variables */ - const char *fcn_name = "rr_obj_hdr_flush_confusion_writer"; - const hbool_t verbose = FALSE; + const char * fcn_name = "rr_obj_hdr_flush_confusion_writer"; + const hbool_t verbose = FALSE; const H5Ptest_param_t *pt; - char *filename; + char * filename; /* * setup test bed related variables: */ - pt = (const H5Ptest_param_t *)GetTestParameters(); + pt = (const H5Ptest_param_t *)GetTestParameters(); filename = pt->name; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_world_rank); @@ -1990,50 +1949,45 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * Set up file access property list with parallel I/O access */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Setting up property list.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Setting up property list.\n", mpi_rank, fcn_name); fapl_id = H5Pcreate(H5P_FILE_ACCESS); VRFY((fapl_id != -1), "H5Pcreate(H5P_FILE_ACCESS) failed"); err = H5Pset_fapl_mpio(fapl_id, comm, MPI_INFO_NULL); - VRFY((err >= 0 ), "H5Pset_fapl_mpio() failed"); - + VRFY((err >= 0), "H5Pset_fapl_mpio() failed"); /* * Create a new file collectively and release property list identifier. */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Creating new file \"%s\".\n", - mpi_rank, fcn_name, filename); + if (verbose) + HDfprintf(stdout, "%0d:%s: Creating new file \"%s\".\n", mpi_rank, fcn_name, filename); file_id = H5Fcreate(filename, H5F_ACC_TRUNC, H5P_DEFAULT, fapl_id); - VRFY((file_id >= 0 ), "H5Fcreate() failed"); + VRFY((file_id >= 0), "H5Fcreate() failed"); err = H5Pclose(fapl_id); - VRFY((err >= 0 ), "H5Pclose(fapl_id) failed"); - + VRFY((err >= 0), "H5Pclose(fapl_id) failed"); /* * Step 1: create the data sets and write data. */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Creating the datasets.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Creating the datasets.\n", mpi_rank, fcn_name); disk_size[0] = (hsize_t)(LOCAL_DATA_SIZE * mpi_size); - mem_size[0] = (hsize_t)(LOCAL_DATA_SIZE); + mem_size[0] = (hsize_t)(LOCAL_DATA_SIZE); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { disk_space[i] = H5Screate_simple(1, disk_size, NULL); - VRFY((disk_space[i] >= 0), "H5Screate_simple(1) failed.\n"); + VRFY((disk_space[i] >= 0), "H5Screate_simple(1) failed.\n"); - dataset[i] = H5Dcreate2(file_id, dataset_name[i], H5T_NATIVE_DOUBLE, - disk_space[i], H5P_DEFAULT, H5P_DEFAULT, H5P_DEFAULT); + dataset[i] = H5Dcreate2(file_id, dataset_name[i], H5T_NATIVE_DOUBLE, disk_space[i], H5P_DEFAULT, + H5P_DEFAULT, H5P_DEFAULT); VRFY((dataset[i] >= 0), "H5Dcreate(1) failed.\n"); } @@ -2042,45 +1996,41 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * setup data transfer property list */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Setting up dxpl.\n", mpi_rank, fcn_name); dxpl_id = H5Pcreate(H5P_DATASET_XFER); VRFY((dxpl_id != -1), "H5Pcreate(H5P_DATASET_XFER) failed.\n"); err = H5Pset_dxpl_mpio(dxpl_id, H5FD_MPIO_COLLECTIVE); - VRFY((err >= 0), - "H5Pset_dxpl_mpio(dxpl_id, H5FD_MPIO_COLLECTIVE) failed.\n"); + VRFY((err >= 0), "H5Pset_dxpl_mpio(dxpl_id, H5FD_MPIO_COLLECTIVE) failed.\n"); /* * write data to the data sets */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Writing datasets.\n", mpi_rank, fcn_name); disk_count[0] = (hsize_t)(LOCAL_DATA_SIZE); disk_start[0] = (hsize_t)(LOCAL_DATA_SIZE * mpi_rank); - mem_count[0] = (hsize_t)(LOCAL_DATA_SIZE); - mem_start[0] = (hsize_t)(0); + mem_count[0] = (hsize_t)(LOCAL_DATA_SIZE); + mem_start[0] = (hsize_t)(0); - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) { + for (j = 0; j < LOCAL_DATA_SIZE; j++) { data[j] = (double)(mpi_rank + 1); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { - err = H5Sselect_hyperslab(disk_space[i], H5S_SELECT_SET, disk_start, - NULL, disk_count, NULL); + for (i = 0; i < NUM_DATA_SETS; i++) { + err = H5Sselect_hyperslab(disk_space[i], H5S_SELECT_SET, disk_start, NULL, disk_count, NULL); VRFY((err >= 0), "H5Sselect_hyperslab(1) failed.\n"); mem_space[i] = H5Screate_simple(1, mem_size, NULL); VRFY((mem_space[i] >= 0), "H5Screate_simple(2) failed.\n"); - err = H5Sselect_hyperslab(mem_space[i], H5S_SELECT_SET, - mem_start, NULL, mem_count, NULL); + err = H5Sselect_hyperslab(mem_space[i], H5S_SELECT_SET, mem_start, NULL, mem_count, NULL); VRFY((err >= 0), "H5Sselect_hyperslab(2) failed.\n"); - err = H5Dwrite(dataset[i], H5T_NATIVE_DOUBLE, mem_space[i], - disk_space[i], dxpl_id, data); + err = H5Dwrite(dataset[i], H5T_NATIVE_DOUBLE, mem_space[i], disk_space[i], dxpl_id, data); VRFY((err >= 0), "H5Dwrite(1) failed.\n"); - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) + for (j = 0; j < LOCAL_DATA_SIZE; j++) data[j] *= 10.0; } @@ -2088,10 +2038,10 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * close the data spaces */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: closing dataspaces.\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { err = H5Sclose(disk_space[i]); VRFY((err >= 0), "H5Sclose(disk_space[i]) failed.\n"); err = H5Sclose(mem_space[i]); @@ -2104,9 +2054,8 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * flush the metadata cache */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", mpi_rank, fcn_name); err = H5Fflush(file_id, H5F_SCOPE_GLOBAL); VRFY((err >= 0), "H5Fflush(1) failed.\n"); @@ -2118,23 +2067,23 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * Step 2: write attributes to each dataset */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: writing attributes.\n", mpi_rank, fcn_name); att_size[0] = (hsize_t)(LOCAL_DATA_SIZE); - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) { + for (j = 0; j < LOCAL_DATA_SIZE; j++) { att[j] = (double)(j + 1); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { att_space[i] = H5Screate_simple(1, att_size, NULL); VRFY((att_space[i] >= 0), "H5Screate_simple(3) failed.\n"); - att_id[i] = H5Acreate2(dataset[i], att_name[i], H5T_NATIVE_DOUBLE, - att_space[i], H5P_DEFAULT, H5P_DEFAULT); + att_id[i] = + H5Acreate2(dataset[i], att_name[i], H5T_NATIVE_DOUBLE, att_space[i], H5P_DEFAULT, H5P_DEFAULT); VRFY((att_id[i] >= 0), "H5Acreate(1) failed.\n"); err = H5Awrite(att_id[i], H5T_NATIVE_DOUBLE, att); VRFY((err >= 0), "H5Awrite(1) failed.\n"); - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) { + for (j = 0; j < LOCAL_DATA_SIZE; j++) { att[j] /= 10.0; } } @@ -2143,11 +2092,10 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * close attribute IDs and spaces */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: closing attr ids and spaces .\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: closing attr ids and spaces .\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { err = H5Sclose(att_space[i]); VRFY((err >= 0), "H5Sclose(att_space[i]) failed.\n"); err = H5Aclose(att_id[i]); @@ -2160,9 +2108,8 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * flush the metadata cache again */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", mpi_rank, fcn_name); err = H5Fflush(file_id, H5F_SCOPE_GLOBAL); VRFY((err >= 0), "H5Fflush(2) failed.\n"); @@ -2174,25 +2121,24 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * Step 3: write large attributes to each dataset */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: writing large attributes.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: writing large attributes.\n", mpi_rank, fcn_name); lg_att_size[0] = (hsize_t)(LARGE_ATTR_SIZE); - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) { + for (j = 0; j < LARGE_ATTR_SIZE; j++) { lg_att[j] = (double)(j + 1); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { lg_att_space[i] = H5Screate_simple(1, lg_att_size, NULL); VRFY((lg_att_space[i] >= 0), "H5Screate_simple(4) failed.\n"); - lg_att_id[i] = H5Acreate2(dataset[i], lg_att_name[i], H5T_NATIVE_DOUBLE, - lg_att_space[i], H5P_DEFAULT, H5P_DEFAULT); + lg_att_id[i] = H5Acreate2(dataset[i], lg_att_name[i], H5T_NATIVE_DOUBLE, lg_att_space[i], H5P_DEFAULT, + H5P_DEFAULT); VRFY((lg_att_id[i] >= 0), "H5Acreate(2) failed.\n"); err = H5Awrite(lg_att_id[i], H5T_NATIVE_DOUBLE, lg_att); VRFY((err >= 0), "H5Awrite(2) failed.\n"); - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) { + for (j = 0; j < LARGE_ATTR_SIZE; j++) { lg_att[j] /= 10.0; } } @@ -2209,9 +2155,8 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * code is going to change a lot in the near future. */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", mpi_rank, fcn_name); err = H5Fflush(file_id, H5F_SCOPE_GLOBAL); VRFY((err >= 0), "H5Fflush(3) failed.\n"); @@ -2223,18 +2168,17 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * Step 4: write different large attributes to each dataset */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: writing different large attributes.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: writing different large attributes.\n", mpi_rank, fcn_name); - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) { + for (j = 0; j < LARGE_ATTR_SIZE; j++) { lg_att[j] = (double)(j + 2); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { err = H5Awrite(lg_att_id[i], H5T_NATIVE_DOUBLE, lg_att); VRFY((err >= 0), "H5Awrite(2) failed.\n"); - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) { + for (j = 0; j < LARGE_ATTR_SIZE; j++) { lg_att[j] /= 10.0; } } @@ -2244,9 +2188,8 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) /* * flush the metadata cache again */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: flushing metadata cache.\n", mpi_rank, fcn_name); err = H5Fflush(file_id, H5F_SCOPE_GLOBAL); VRFY((err >= 0), "H5Fflush(3) failed.\n"); @@ -2260,11 +2203,10 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * close large attribute IDs and spaces */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: closing large attr ids and spaces .\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: closing large attr ids and spaces .\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { err = H5Sclose(lg_att_space[i]); VRFY((err >= 0), "H5Sclose(lg_att_space[i]) failed.\n"); @@ -2272,15 +2214,14 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) VRFY((err >= 0), "H5Aclose(lg_att_id[i]) failed.\n"); } - /* * close the data sets */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: closing datasets .\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { err = H5Dclose(dataset[i]); VRFY((err >= 0), "H5Dclose(dataset[i])1 failed.\n"); } @@ -2289,88 +2230,87 @@ void rr_obj_hdr_flush_confusion_writer(MPI_Comm comm) * close the data transfer property list. */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: closing dxpl .\n", mpi_rank, fcn_name); err = H5Pclose(dxpl_id); VRFY((err >= 0), "H5Pclose(dxpl_id) failed.\n"); - /* * Close file. */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: closing file.\n", mpi_rank, fcn_name); err = H5Fclose(file_id); - VRFY((err >= 0 ), "H5Fclose(1) failed"); + VRFY((err >= 0), "H5Fclose(1) failed"); /* End of Step 5: Close all objects and the file */ /* Tell the reader to check the file up to steps. */ steps++; Reader_check(mrc, steps, steps_done); - /* All done. Inform reader to end. */ - steps=0; + steps = 0; Reader_check(mrc, steps, steps_done); - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Done.\n", mpi_rank, fcn_name); return; } /* rr_obj_hdr_flush_confusion_writer() */ -void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) +void +rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) { - int i; - int j; - hid_t file_id = -1; - hid_t fapl_id = -1; - hid_t dxpl_id = -1; - hid_t lg_att_id[NUM_DATA_SETS]; - hid_t lg_att_type[NUM_DATA_SETS]; - hid_t disk_space[NUM_DATA_SETS]; - hid_t mem_space[NUM_DATA_SETS]; - hid_t dataset[NUM_DATA_SETS]; + int i; + int j; + hid_t file_id = -1; + hid_t fapl_id = -1; + hid_t dxpl_id = -1; + hid_t lg_att_id[NUM_DATA_SETS]; + hid_t lg_att_type[NUM_DATA_SETS]; + hid_t disk_space[NUM_DATA_SETS]; + hid_t mem_space[NUM_DATA_SETS]; + hid_t dataset[NUM_DATA_SETS]; hsize_t disk_count[1]; hsize_t disk_start[1]; hsize_t mem_count[1]; hsize_t mem_size[1]; hsize_t mem_start[1]; - herr_t err; - htri_t tri_err; - double data[LOCAL_DATA_SIZE]; - double data_read[LOCAL_DATA_SIZE]; - double att[LOCAL_DATA_SIZE]; - double att_read[LOCAL_DATA_SIZE]; - double lg_att[LARGE_ATTR_SIZE]; - double lg_att_read[LARGE_ATTR_SIZE]; + herr_t err; + htri_t tri_err; + double data[LOCAL_DATA_SIZE]; + double data_read[LOCAL_DATA_SIZE]; + double att[LOCAL_DATA_SIZE]; + double att_read[LOCAL_DATA_SIZE]; + double lg_att[LARGE_ATTR_SIZE]; + double lg_att_read[LARGE_ATTR_SIZE]; /* MPI variables */ /* world communication size and rank */ - int mpi_world_size; - int mpi_world_rank; + int mpi_world_size; + int mpi_world_rank; /* private communicator size and rank */ - int mpi_size; - int mpi_rank; - int mrc; /* mpi error code */ - int steps = -1; /* How far (steps) to verify the file */ - int steps_done = -1; /* How far (steps) have been verified */ + int mpi_size; + int mpi_rank; + int mrc; /* mpi error code */ + int steps = -1; /* How far (steps) to verify the file */ + int steps_done = -1; /* How far (steps) have been verified */ /* test bed related variables */ - const char *fcn_name = "rr_obj_hdr_flush_confusion_reader"; - const hbool_t verbose = FALSE; + const char * fcn_name = "rr_obj_hdr_flush_confusion_reader"; + const hbool_t verbose = FALSE; const H5Ptest_param_t *pt; - char *filename; + char * filename; /* * setup test bed related variables: */ - pt = (const H5Ptest_param_t *)GetTestParameters(); + pt = (const H5Ptest_param_t *)GetTestParameters(); filename = pt->name; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_world_rank); @@ -2380,53 +2320,50 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) /* Repeatedly re-open the file and verify its contents until it is */ /* told to end (when steps=0). */ - while (steps_done != 0){ + while (steps_done != 0) { Reader_wait(mrc, steps); VRFY((mrc >= 0), "Reader_wait failed"); steps_done = 0; - if (steps > 0 ){ + if (steps > 0) { /* - * Set up file access property list with parallel I/O access - */ + * Set up file access property list with parallel I/O access + */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Setting up property list.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: Setting up property list.\n", mpi_rank, fcn_name); fapl_id = H5Pcreate(H5P_FILE_ACCESS); VRFY((fapl_id != -1), "H5Pcreate(H5P_FILE_ACCESS) failed"); err = H5Pset_fapl_mpio(fapl_id, comm, MPI_INFO_NULL); - VRFY((err >= 0 ), "H5Pset_fapl_mpio() failed"); + VRFY((err >= 0), "H5Pset_fapl_mpio() failed"); /* - * Create a new file collectively and release property list identifier. - */ + * Create a new file collectively and release property list identifier. + */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: Re-open file \"%s\".\n", - mpi_rank, fcn_name, filename); + if (verbose) + HDfprintf(stdout, "%0d:%s: Re-open file \"%s\".\n", mpi_rank, fcn_name, filename); file_id = H5Fopen(filename, H5F_ACC_RDONLY, fapl_id); - VRFY((file_id >= 0 ), "H5Fopen() failed"); + VRFY((file_id >= 0), "H5Fopen() failed"); err = H5Pclose(fapl_id); - VRFY((err >= 0 ), "H5Pclose(fapl_id) failed"); + VRFY((err >= 0), "H5Pclose(fapl_id) failed"); #if 1 - if (steps >= 1){ + if (steps >= 1) { /*=====================================================* - * Step 1: open the data sets and read data. - *=====================================================*/ + * Step 1: open the data sets and read data. + *=====================================================*/ - if(verbose ) - HDfprintf(stdout, "%0d:%s: opening the datasets.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: opening the datasets.\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { dataset[i] = -1; } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { dataset[i] = H5Dopen2(file_id, dataset_name[i], H5P_DEFAULT); VRFY((dataset[i] >= 0), "H5Dopen(1) failed.\n"); disk_space[i] = H5Dget_space(dataset[i]); @@ -2434,23 +2371,22 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) } /* - * setup data transfer property list - */ + * setup data transfer property list + */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Setting up dxpl.\n", mpi_rank, fcn_name); dxpl_id = H5Pcreate(H5P_DATASET_XFER); VRFY((dxpl_id != -1), "H5Pcreate(H5P_DATASET_XFER) failed.\n"); err = H5Pset_dxpl_mpio(dxpl_id, H5FD_MPIO_COLLECTIVE); - VRFY((err >= 0), - "H5Pset_dxpl_mpio(dxpl_id, H5FD_MPIO_COLLECTIVE) failed.\n"); + VRFY((err >= 0), "H5Pset_dxpl_mpio(dxpl_id, H5FD_MPIO_COLLECTIVE) failed.\n"); /* - * read data from the data sets - */ + * read data from the data sets + */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Reading datasets.\n", mpi_rank, fcn_name); disk_count[0] = (hsize_t)(LOCAL_DATA_SIZE); @@ -2462,44 +2398,43 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) mem_start[0] = (hsize_t)(0); /* set up expected data for verification */ - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) { + for (j = 0; j < LOCAL_DATA_SIZE; j++) { data[j] = (double)(mpi_rank + 1); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { - err = H5Sselect_hyperslab(disk_space[i], H5S_SELECT_SET, disk_start, - NULL, disk_count, NULL); + for (i = 0; i < NUM_DATA_SETS; i++) { + err = H5Sselect_hyperslab(disk_space[i], H5S_SELECT_SET, disk_start, NULL, disk_count, + NULL); VRFY((err >= 0), "H5Sselect_hyperslab(1) failed.\n"); mem_space[i] = H5Screate_simple(1, mem_size, NULL); VRFY((mem_space[i] >= 0), "H5Screate_simple(2) failed.\n"); - err = H5Sselect_hyperslab(mem_space[i], H5S_SELECT_SET, - mem_start, NULL, mem_count, NULL); + err = H5Sselect_hyperslab(mem_space[i], H5S_SELECT_SET, mem_start, NULL, mem_count, NULL); VRFY((err >= 0), "H5Sselect_hyperslab(2) failed.\n"); - err = H5Dread(dataset[i], H5T_NATIVE_DOUBLE, mem_space[i], - disk_space[i], dxpl_id, data_read); + err = H5Dread(dataset[i], H5T_NATIVE_DOUBLE, mem_space[i], disk_space[i], dxpl_id, + data_read); VRFY((err >= 0), "H5Dread(1) failed.\n"); /* compare read data with expected data */ - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) - if (data_read[j] != data[j]){ + for (j = 0; j < LOCAL_DATA_SIZE; j++) + if (data_read[j] != data[j]) { HDfprintf(stdout, - "%0d:%s: Reading datasets value failed in " - "Dataset %d, at position %d: expect %f, got %f.\n", - mpi_rank, fcn_name, i, j, data[j], data_read[j]); + "%0d:%s: Reading datasets value failed in " + "Dataset %d, at position %d: expect %f, got %f.\n", + mpi_rank, fcn_name, i, j, data[j], data_read[j]); nerrors++; } - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) + for (j = 0; j < LOCAL_DATA_SIZE; j++) data[j] *= 10.0; } /* - * close the data spaces - */ + * close the data spaces + */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: closing dataspaces.\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { err = H5Sclose(disk_space[i]); VRFY((err >= 0), "H5Sclose(disk_space[i]) failed.\n"); err = H5Sclose(mem_space[i]); @@ -2512,18 +2447,18 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) #if 1 /*=====================================================* - * Step 2: reading attributes from each dataset - *=====================================================*/ + * Step 2: reading attributes from each dataset + *=====================================================*/ - if (steps >= 2){ - if(verbose ) + if (steps >= 2) { + if (verbose) HDfprintf(stdout, "%0d:%s: reading attributes.\n", mpi_rank, fcn_name); - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) { + for (j = 0; j < LOCAL_DATA_SIZE; j++) { att[j] = (double)(j + 1); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { hid_t att_id, att_type; att_id = H5Aopen(dataset[i], att_name[i], H5P_DEFAULT); @@ -2532,10 +2467,9 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) VRFY((att_type >= 0), "H5Aget_type failed.\n"); tri_err = H5Tequal(att_type, H5T_NATIVE_DOUBLE); VRFY((tri_err >= 0), "H5Tequal failed.\n"); - if (tri_err==0){ - HDfprintf(stdout, - "%0d:%s: Mismatched Attribute type of Dataset %d.\n", - mpi_rank, fcn_name, i); + if (tri_err == 0) { + HDfprintf(stdout, "%0d:%s: Mismatched Attribute type of Dataset %d.\n", mpi_rank, + fcn_name, i); nerrors++; } else { @@ -2543,14 +2477,15 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) err = H5Aread(att_id, H5T_NATIVE_DOUBLE, att_read); VRFY((err >= 0), "H5Aread failed.\n"); /* compare read attribute data with expected data */ - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) - if (att_read[j] != att[j]){ + for (j = 0; j < LOCAL_DATA_SIZE; j++) + if (att_read[j] != att[j]) { HDfprintf(stdout, - "%0d:%s: Mismatched attribute data read in Dataset %d, at position %d: expect %f, got %f.\n", - mpi_rank, fcn_name, i, j, att[j], att_read[j]); + "%0d:%s: Mismatched attribute data read in Dataset %d, at position " + "%d: expect %f, got %f.\n", + mpi_rank, fcn_name, i, j, att[j], att_read[j]); nerrors++; } - for ( j = 0; j < LOCAL_DATA_SIZE; j++ ) { + for (j = 0; j < LOCAL_DATA_SIZE; j++) { att[j] /= 10.0; } } @@ -2562,47 +2497,46 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) /* End of Step 2: reading attributes from each dataset */ #endif - #if 1 /*=====================================================* - * Step 3 or 4: read large attributes from each dataset. - * Step 4 has different attribute value from step 3. - *=====================================================*/ + * Step 3 or 4: read large attributes from each dataset. + * Step 4 has different attribute value from step 3. + *=====================================================*/ - if (steps >= 3){ - if(verbose ) + if (steps >= 3) { + if (verbose) HDfprintf(stdout, "%0d:%s: reading large attributes.\n", mpi_rank, fcn_name); - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) { - lg_att[j] = (steps==3) ? (double)(j + 1) : (double)(j+2); + for (j = 0; j < LARGE_ATTR_SIZE; j++) { + lg_att[j] = (steps == 3) ? (double)(j + 1) : (double)(j + 2); } - for ( i = 0; i < NUM_DATA_SETS; i++ ) { + for (i = 0; i < NUM_DATA_SETS; i++) { lg_att_id[i] = H5Aopen(dataset[i], lg_att_name[i], H5P_DEFAULT); VRFY((lg_att_id[i] >= 0), "H5Aopen(2) failed.\n"); lg_att_type[i] = H5Aget_type(lg_att_id[i]); VRFY((err >= 0), "H5Aget_type failed.\n"); tri_err = H5Tequal(lg_att_type[i], H5T_NATIVE_DOUBLE); VRFY((tri_err >= 0), "H5Tequal failed.\n"); - if (tri_err==0){ - HDfprintf(stdout, - "%0d:%s: Mismatched Large attribute type of Dataset %d.\n", - mpi_rank, fcn_name, i); + if (tri_err == 0) { + HDfprintf(stdout, "%0d:%s: Mismatched Large attribute type of Dataset %d.\n", + mpi_rank, fcn_name, i); nerrors++; } - else{ + else { /* should verify large attribute size before H5Aread */ err = H5Aread(lg_att_id[i], H5T_NATIVE_DOUBLE, lg_att_read); VRFY((err >= 0), "H5Aread failed.\n"); /* compare read attribute data with expected data */ - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) - if (lg_att_read[j] != lg_att[j]){ + for (j = 0; j < LARGE_ATTR_SIZE; j++) + if (lg_att_read[j] != lg_att[j]) { HDfprintf(stdout, - "%0d:%s: Mismatched large attribute data read in Dataset %d, at position %d: expect %f, got %f.\n", - mpi_rank, fcn_name, i, j, lg_att[j], lg_att_read[j]); + "%0d:%s: Mismatched large attribute data read in Dataset %d, at " + "position %d: expect %f, got %f.\n", + mpi_rank, fcn_name, i, j, lg_att[j], lg_att_read[j]); nerrors++; } - for ( j = 0; j < LARGE_ATTR_SIZE; j++ ) { + for (j = 0; j < LARGE_ATTR_SIZE; j++) { lg_att[j] /= 10.0; } @@ -2613,61 +2547,58 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) VRFY((err >= 0), "H5Aclose failed.\n"); } /* Both step 3 and 4 use this same read checking code. */ - steps_done = (steps==3) ? 3 : 4; + steps_done = (steps == 3) ? 3 : 4; } /* End of Step 3 or 4: read large attributes from each dataset */ #endif - /*=====================================================* - * Step 5: read all objects from the file - *=====================================================*/ - if (steps>=5){ + * Step 5: read all objects from the file + *=====================================================*/ + if (steps >= 5) { /* nothing extra to verify. The file is closed normally. */ /* Just increment steps_done */ steps_done++; } /* - * Close the data sets - */ + * Close the data sets + */ - if(verbose ) - HDfprintf(stdout, "%0d:%s: closing datasets again.\n", - mpi_rank, fcn_name); + if (verbose) + HDfprintf(stdout, "%0d:%s: closing datasets again.\n", mpi_rank, fcn_name); - for ( i = 0; i < NUM_DATA_SETS; i++ ) { - if ( dataset[i] >= 0 ) { + for (i = 0; i < NUM_DATA_SETS; i++) { + if (dataset[i] >= 0) { err = H5Dclose(dataset[i]); VRFY((err >= 0), "H5Dclose(dataset[i])1 failed.\n"); } } /* - * close the data transfer property list. - */ + * close the data transfer property list. + */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: closing dxpl .\n", mpi_rank, fcn_name); err = H5Pclose(dxpl_id); VRFY((err >= 0), "H5Pclose(dxpl_id) failed.\n"); /* - * Close the file - */ - if(verbose) - HDfprintf(stdout, "%0d:%s: closing file again.\n", - mpi_rank, fcn_name); + * Close the file + */ + if (verbose) + HDfprintf(stdout, "%0d:%s: closing file again.\n", mpi_rank, fcn_name); err = H5Fclose(file_id); - VRFY((err >= 0 ), "H5Fclose(1) failed"); + VRFY((err >= 0), "H5Fclose(1) failed"); } /* else if (steps_done==0) */ Reader_result(mrc, steps_done); } /* end while(1) */ - if(verbose ) + if (verbose) HDfprintf(stdout, "%0d:%s: Done.\n", mpi_rank, fcn_name); return; @@ -2682,27 +2613,27 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) #undef Writer_Root #undef Reader_Root - /* * Test creating a chunked dataset in parallel in a file with an alignment set * and an alignment threshold large enough to avoid aligning the chunks but * small enough that the raw data aggregator will be aligned if it is treated as * an object that must be aligned by the library */ -#define CHUNK_SIZE 72 -#define NCHUNKS 32 -#define AGGR_SIZE 2048 +#define CHUNK_SIZE 72 +#define NCHUNKS 32 +#define AGGR_SIZE 2048 #define EXTRA_ALIGN 100 - void chunk_align_bug_1(void) - { - int mpi_rank; - hid_t file_id, dset_id, fapl_id, dcpl_id, space_id; - hsize_t dims = CHUNK_SIZE * NCHUNKS, cdims = CHUNK_SIZE; - h5_stat_size_t file_size; - hsize_t align; - herr_t ret; - const char *filename; +void +chunk_align_bug_1(void) +{ + int mpi_rank; + hid_t file_id, dset_id, fapl_id, dcpl_id, space_id; + hsize_t dims = CHUNK_SIZE * NCHUNKS, cdims = CHUNK_SIZE; + h5_stat_size_t file_size; + hsize_t align; + herr_t ret; + const char * filename; MPI_Comm_rank(MPI_COMM_WORLD, &mpi_rank); @@ -2767,8 +2698,6 @@ void rr_obj_hdr_flush_confusion_reader(MPI_Comm comm) return; } /* end chunk_align_bug_1() */ - /*============================================================================= * End of t_mdset.c *===========================================================================*/ - |