diff options
author | Quincey Koziol <koziol@hdfgroup.org> | 2006-06-27 14:45:06 (GMT) |
---|---|---|
committer | Quincey Koziol <koziol@hdfgroup.org> | 2006-06-27 14:45:06 (GMT) |
commit | 7be3afb278aea67ba09a97f4b41c0aaaf5c47983 (patch) | |
tree | 24ed86ab2a5c982fbf182d2ac8cd892c3813bc34 /testpar | |
parent | 8d72542a50fac7a747fe0bfec8d2285de8efd29f (diff) | |
download | hdf5-7be3afb278aea67ba09a97f4b41c0aaaf5c47983.zip hdf5-7be3afb278aea67ba09a97f4b41c0aaaf5c47983.tar.gz hdf5-7be3afb278aea67ba09a97f4b41c0aaaf5c47983.tar.bz2 |
[svn-r12440] Purpose:
Code cleanup
Description:
Trim trailing whitespace in Makefile.am and C/C++ source files to make
diffing changes easier.
Platforms tested:
None necessary, whitespace only change
Diffstat (limited to 'testpar')
-rw-r--r-- | testpar/t_cache.c | 550 | ||||
-rw-r--r-- | testpar/t_chunk_alloc.c | 34 | ||||
-rw-r--r-- | testpar/t_coll_chunk.c | 110 | ||||
-rw-r--r-- | testpar/t_mpi.c | 18 | ||||
-rw-r--r-- | testpar/t_posix_compliant.c | 90 | ||||
-rw-r--r-- | testpar/t_span_tree.c | 54 | ||||
-rw-r--r-- | testpar/testphdf5.c | 8 |
7 files changed, 432 insertions, 432 deletions
diff --git a/testpar/t_cache.c b/testpar/t_cache.c index 3bacb8e..4b7d298 100644 --- a/testpar/t_cache.c +++ b/testpar/t_cache.c @@ -39,8 +39,8 @@ int nerrors = 0; int failures = 0; hbool_t verbose = TRUE; /* used to control error messages */ #if 0 -/* So far we haven't needed this, but that may change. - * Keep it around for now +/* So far we haven't needed this, but that may change. + * Keep it around for now */ hid_t noblock_dxpl_id=(-1); #endif @@ -80,11 +80,11 @@ long local_pins = 0; * struct datum * * Instances of struct datum are used to store information on entries - * that may be loaded into the cache. The individual fields are + * that may be loaded into the cache. The individual fields are * discussed below: * * header: Instance of H5C_cache_entry_t used by the for its data. - * This field is only used on the file processes, not on the + * This field is only used on the file processes, not on the * server process. * * This field MUST be the first entry in this structure. @@ -98,27 +98,27 @@ long local_pins = 0; * * dirty: Boolean flag indicating whether the entry is dirty. * - * For current purposes, an entry is clean until it is + * For current purposes, an entry is clean until it is * modified, and dirty until written to the server (cache - * on process 0) or until it is marked clean (all other + * on process 0) or until it is marked clean (all other * caches). * * valid: Boolean flag indicating whether the entry contains * valid data. Attempts to read an entry whose valid * flag is not set should trigger an error. * - * locked: Boolean flag that is set to true iff the entry is in + * locked: Boolean flag that is set to true iff the entry is in * the cache and locked. * - * global_pinned: Boolean flag that is set to true iff the entry has + * global_pinned: Boolean flag that is set to true iff the entry has * been pinned collectively in all caches. Since writes must * be collective across all processes, only entries pinned * in this fashion may be marked dirty. * * local_pinned: Boolean flag that is set to true iff the entry * has been pinned in the local cache, but probably not all - * caches. Such pins will typically not be consistant across - * processes, and thus cannot be marked as dirty unless they + * caches. Such pins will typically not be consistant across + * processes, and thus cannot be marked as dirty unless they * happen to overlap some collective operation. * * index: Index of this instance of datum in the data_index[] array @@ -126,7 +126,7 @@ long local_pins = 0; * *****************************************************************************/ -struct datum +struct datum { H5C_cache_entry_t header; haddr_t base_addr; @@ -143,14 +143,14 @@ struct datum /***************************************************************************** * data array * - * The data array is an array of instances of datum of size + * The data array is an array of instances of datum of size * NUM_DATA_ENTRIES that is used to track the particulars of all * the entries that may be loaded into the cache. * * It exists on all processes, although the master copy is maintained * by the server process. If the cache is performing correctly, all * versions should be effectively identical. By that I mean that - * the data received from the server should always match that in + * the data received from the server should always match that in * the local version of the data array. * *****************************************************************************/ @@ -163,8 +163,8 @@ struct datum data[NUM_DATA_ENTRIES]; /* Many tests use the size of data array as the size of test loops. * On some machines, this results in unacceptably long test runs. * - * To deal with this issue, I have introduced the virt_num_data_entries - * global, which can be set to a lower value to throtle the length of + * To deal with this issue, I have introduced the virt_num_data_entries + * global, which can be set to a lower value to throtle the length of * tests. * * Note that this value must always be divisible by 40, and must be an @@ -172,7 +172,7 @@ struct datum data[NUM_DATA_ENTRIES]; * powers of 10 that meet these criteria. * * Further, this value must be consistant across all processes. - */ + */ #define STD_VIRT_NUM_DATA_ENTRIES NUM_DATA_ENTRIES #define EXPRESS_VIRT_NUM_DATA_ENTRIES (NUM_DATA_ENTRIES / 10) @@ -191,9 +191,9 @@ int virt_num_data_entries = NUM_DATA_ENTRIES; * This array is necessary, as rename operations can swap the values * of the base_addr fields of two instances of datum. Without this * array, we would no longer be able to use a binary search on a sorted - * list to find the indexes of instances of datum given the values of + * list to find the indexes of instances of datum given the values of * their base_addr fields. - * + * *****************************************************************************/ int data_index[NUM_DATA_ENTRIES]; @@ -202,7 +202,7 @@ int data_index[NUM_DATA_ENTRIES]; /***************************************************************************** * struct mssg * - * The mssg structure is used as a generic container for messages to + * The mssg structure is used as a generic container for messages to * and from the server. Not all fields are used in all cases. * * req: Integer field containing the type of the message. @@ -219,7 +219,7 @@ int data_index[NUM_DATA_ENTRIES]; * * ver: Version number of a datum. Not used in all mssgs. * - * magic: Magic number for error detection. Must be set to + * magic: Magic number for error detection. Must be set to * MSSG_MAGIC. * *****************************************************************************/ @@ -300,7 +300,7 @@ hbool_t serve_write_request(struct mssg_t * mssg_ptr); herr_t clear_datum(H5F_t * f, void * thing, hbool_t dest); herr_t destroy_datum(H5F_t UNUSED * f, void * thing); -herr_t flush_datum(H5F_t *f, hid_t UNUSED dxpl_id, hbool_t dest, haddr_t addr, +herr_t flush_datum(H5F_t *f, hid_t UNUSED dxpl_id, hbool_t dest, haddr_t addr, void *thing); void * load_datum(H5F_t UNUSED *f, hid_t UNUSED dxpl_id, haddr_t addr, const void UNUSED *udata1, void UNUSED *udata2); @@ -325,10 +325,10 @@ const H5C_class_t types[NUMBER_OF_ENTRY_TYPES] = /* test utility functions */ -void insert_entry(H5C_t * cache_ptr, H5F_t * file_ptr, +void insert_entry(H5C_t * cache_ptr, H5F_t * file_ptr, int32_t idx, unsigned int flags); void local_pin_and_unpin_random_entries(H5C_t * cache_ptr, H5F_t * file_ptr, - int min_idx, int max_idx, + int min_idx, int max_idx, int min_count, int max_count); void local_pin_random_entry(H5C_t * cache_ptr, H5F_t * file_ptr, int min_idx, int max_idx); @@ -355,7 +355,7 @@ hbool_t setup_cache_for_test(hid_t * fid_ptr, H5F_t ** file_ptr_ptr, H5C_t ** cache_ptr_ptr); void setup_rand(void); hbool_t take_down_cache(hid_t fid); -void unlock_entry(H5C_t * cache_ptr, H5F_t * file_ptr, +void unlock_entry(H5C_t * cache_ptr, H5F_t * file_ptr, int32_t type, unsigned int flags); void unpin_entry(H5C_t * cache_ptr, H5F_t * file_ptr, int32_t idx, hbool_t global, hbool_t dirty, hbool_t via_unprotect); @@ -391,19 +391,19 @@ hbool_t smoke_check_5(void); * Modifications: * * None. - * + * *****************************************************************************/ -void +void print_stats(void) { - HDfprintf(stdout, + HDfprintf(stdout, "%d: datum clears / pinned clears / destroys = %ld / %ld / %ld\n", - world_mpi_rank, datum_clears, datum_pinned_clears, + world_mpi_rank, datum_clears, datum_pinned_clears, datum_destroys ); - HDfprintf(stdout, + HDfprintf(stdout, "%d: datum flushes / pinned flushes / loads = %ld / %ld / %ld\n", - world_mpi_rank, datum_flushes, datum_pinned_flushes, + world_mpi_rank, datum_flushes, datum_pinned_flushes, datum_loads ); HDfprintf(stdout, "%d: pins: global / global dirty / local = %ld / %ld / %ld\n", @@ -427,7 +427,7 @@ print_stats(void) * Modifications: * * None. - * + * *****************************************************************************/ void @@ -468,7 +468,7 @@ reset_stats(void) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -491,7 +491,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: MPI_Comm_group() failed with error %d.\n", world_mpi_rank, fcn_name, mpi_result); } @@ -499,10 +499,10 @@ set_up_file_communicator(void) } if ( success ) { - + num_excluded_ranks = 1; excluded_ranks[0] = world_server_mpi_rank; - mpi_result = MPI_Group_excl(world_group, num_excluded_ranks, + mpi_result = MPI_Group_excl(world_group, num_excluded_ranks, excluded_ranks, &file_group); if ( mpi_result != MPI_SUCCESS ) { @@ -510,7 +510,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: MPI_Group_excl() failed with error %d.\n", world_mpi_rank, fcn_name, mpi_result); } @@ -519,7 +519,7 @@ set_up_file_communicator(void) if ( success ) { - mpi_result = MPI_Comm_create(world_mpi_comm, file_group, + mpi_result = MPI_Comm_create(world_mpi_comm, file_group, &file_mpi_comm); if ( mpi_result != MPI_SUCCESS ) { @@ -527,7 +527,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: MPI_Comm_create() failed with error %d.\n", world_mpi_rank, fcn_name, mpi_result); } @@ -541,7 +541,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: file_mpi_comm == MPI_COMM_NULL.\n", world_mpi_rank, fcn_name); } @@ -555,7 +555,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: file_mpi_comm != MPI_COMM_NULL.\n", world_mpi_rank, fcn_name); } @@ -573,7 +573,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: MPI_Comm_size() failed with error %d.\n", world_mpi_rank, fcn_name, mpi_result); } @@ -589,7 +589,7 @@ set_up_file_communicator(void) nerrors++; success = FALSE; if ( verbose ) { - fprintf(stdout, + fprintf(stdout, "%d:%s: MPI_Comm_rank() failed with error %d.\n", world_mpi_rank, fcn_name, mpi_result); } @@ -609,7 +609,7 @@ set_up_file_communicator(void) * * Function: addr_to_datum_index() * - * Purpose: Given the base address of a datum, find and return its index + * Purpose: Given the base address of a datum, find and return its index * in the data array. * * Return: Success: index of target datum. @@ -621,7 +621,7 @@ set_up_file_communicator(void) * Modifications: * * None. - * + * *****************************************************************************/ int @@ -650,7 +650,7 @@ addr_to_datum_index(haddr_t base_addr) ret_value = data_index[middle]; bottom = top + 1; /* to force exit from while loop */ - } + } } return(ret_value); @@ -662,7 +662,7 @@ addr_to_datum_index(haddr_t base_addr) * * Function: init_data() * - * Purpose: Initialize the data array, from which cache entries are + * Purpose: Initialize the data array, from which cache entries are * loaded. * * Return: Success: TRUE @@ -674,24 +674,24 @@ addr_to_datum_index(haddr_t base_addr) * Modifications: * * None. - * + * *****************************************************************************/ void init_data(void) { const char * fcn_name = "init_data()"; - /* The set of address offsets is chosen so as to avoid allowing the + /* The set of address offsets is chosen so as to avoid allowing the * base addresses to fall in a pattern of that will annoy the hash * table, and to give a good range of entry sizes. * - * At present, I am using the first 20 entries of the Fibonacci + * At present, I am using the first 20 entries of the Fibonacci * sequence multiplied by 2. We will see how it works. */ const int num_addr_offsets = 20; - const haddr_t addr_offsets[20] = { 2, 2, 4, 6, 10, - 16, 26, 42, 68, 110, - 178, 288, 466, 754, 1220, + const haddr_t addr_offsets[20] = { 2, 2, 4, 6, 10, + 16, 26, 42, 68, 110, + 178, 288, 466, 754, 1220, 1974, 3194, 5168, 8362, 13539}; int i; int j = 0; @@ -720,7 +720,7 @@ init_data(void) j = (j + 1) % num_addr_offsets; } - + return; } /* init_data() */ @@ -742,7 +742,7 @@ init_data(void) * processes. This function ensures that all processes agree * on whether to do an express test. * - * Return: Success: Maximum of the values returned by + * Return: Success: Maximum of the values returned by * GetTestExpress() across all processes. * * Failure: -1 @@ -752,7 +752,7 @@ init_data(void) * Modifications: * * None. - * + * *****************************************************************************/ int @@ -777,10 +777,10 @@ do_express_test(void) nerrors++; max_express_test = -1; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Allreduce() failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Allreduce() failed.\n", world_mpi_rank, fcn_name ); } - } + } return(max_express_test); @@ -794,7 +794,7 @@ do_express_test(void) * Purpose: Ensure that all messages sent by this process have been * processed before proceeding. * - * Do this by exchanging sync req / sync ack messages with + * Do this by exchanging sync req / sync ack messages with * the server. * * Do nothing if nerrors is greater than zero. @@ -806,7 +806,7 @@ do_express_test(void) * Modifications: * * None. - * + * *****************************************************************************/ void @@ -830,7 +830,7 @@ do_sync(void) mssg.magic = MSSG_MAGIC; if ( ! send_mssg(&mssg, FALSE) ) { - + nerrors++; if ( verbose ) { HDfprintf(stdout, "%d:%s: send_mssg() failed.\n", @@ -870,7 +870,7 @@ do_sync(void) * * Function: get_max_nerrors() * - * Purpose: Do an MPI_Allreduce to obtain the maximum value of nerrors + * Purpose: Do an MPI_Allreduce to obtain the maximum value of nerrors * across all processes. Return this value. * * Return: Success: Maximum of the nerrors global variables across @@ -883,7 +883,7 @@ do_sync(void) * Modifications: * * None. - * + * *****************************************************************************/ int @@ -905,10 +905,10 @@ get_max_nerrors(void) nerrors++; max_nerrors = -1; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Allreduce() failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Allreduce() failed.\n", world_mpi_rank, fcn_name ); } - } + } return(max_nerrors); @@ -923,7 +923,7 @@ get_max_nerrors(void) * * Function: recv_mssg() * - * Purpose: Receive a message from any process in the provided instance + * Purpose: Receive a message from any process in the provided instance * of struct mssg. * * Return: Success: TRUE @@ -936,7 +936,7 @@ get_max_nerrors(void) * * JRM -- 5/10/06 * Added mssg_tag_offset parameter and supporting code. - * + * *****************************************************************************/ #define CACHE_TEST_TAG 99 /* different from any used by the library */ @@ -958,7 +958,7 @@ recv_mssg(struct mssg_t *mssg_ptr, nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: bad param(s) on entry.\n", + HDfprintf(stdout, "%d:%s: bad param(s) on entry.\n", world_mpi_rank, fcn_name); } } else { @@ -976,7 +976,7 @@ recv_mssg(struct mssg_t *mssg_ptr, nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Recv() failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Recv() failed.\n", world_mpi_rank, fcn_name ); } } else if ( mssg_ptr->magic != MSSG_MAGIC ) { @@ -992,8 +992,8 @@ recv_mssg(struct mssg_t *mssg_ptr, nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, - "%d:%s: mssg_ptr->src != status.MPI_SOURCE.\n", + HDfprintf(stdout, + "%d:%s: mssg_ptr->src != status.MPI_SOURCE.\n", world_mpi_rank, fcn_name); } } @@ -1010,7 +1010,7 @@ recv_mssg(struct mssg_t *mssg_ptr, * * Purpose: Send the provided instance of mssg to the indicated target. * - * Note that all source and destination ranks are in the + * Note that all source and destination ranks are in the * global communicator. * * Return: Success: TRUE @@ -1023,7 +1023,7 @@ recv_mssg(struct mssg_t *mssg_ptr, * * JRM -- 5/10/06 * Added the add_req_to_tag parameter and supporting code. - * + * *****************************************************************************/ hbool_t @@ -1048,7 +1048,7 @@ send_mssg(struct mssg_t *mssg_ptr, nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Invalid mssg on entry.\n", + HDfprintf(stdout, "%d:%s: Invalid mssg on entry.\n", world_mpi_rank, fcn_name); } } @@ -1070,7 +1070,7 @@ send_mssg(struct mssg_t *mssg_ptr, nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Send() failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Send() failed.\n", world_mpi_rank, fcn_name); } } @@ -1097,7 +1097,7 @@ send_mssg(struct mssg_t *mssg_ptr, * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -1108,7 +1108,7 @@ setup_derived_types(void) int i; int result; MPI_Datatype mpi_types[8] = {MPI_INT, MPI_INT, MPI_INT, MPI_LONG, - HADDR_AS_MPI_TYPE, MPI_INT, MPI_INT, + HADDR_AS_MPI_TYPE, MPI_INT, MPI_INT, MPI_UNSIGNED}; int block_len[8] = {1, 1, 1, 1, 1, 1, 1, 1}; MPI_Aint displs[8]; @@ -1127,7 +1127,7 @@ setup_derived_types(void) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Address() call failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Address() call failed.\n", world_mpi_rank, fcn_name); } @@ -1149,7 +1149,7 @@ setup_derived_types(void) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Type_struct() call failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Type_struct() call failed.\n", world_mpi_rank, fcn_name); } } @@ -1164,7 +1164,7 @@ setup_derived_types(void) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Type_commit() call failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Type_commit() call failed.\n", world_mpi_rank, fcn_name); } } @@ -1191,7 +1191,7 @@ setup_derived_types(void) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -1208,7 +1208,7 @@ takedown_derived_types(void) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: MPI_Type_free() call failed.\n", + HDfprintf(stdout, "%d:%s: MPI_Type_free() call failed.\n", world_mpi_rank, fcn_name); } } @@ -1243,10 +1243,10 @@ takedown_derived_types(void) * * JRM -- 5/10/06 * Updated for sync message. - * + * *****************************************************************************/ -hbool_t +hbool_t server_main(void) { const char * fcn_name = "server_main()"; @@ -1260,7 +1260,7 @@ server_main(void) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: This isn't the server process?!?!?\n", + HDfprintf(stdout, "%d:%s: This isn't the server process?!?!?\n", world_mpi_rank, fcn_name); } } @@ -1289,7 +1289,7 @@ server_main(void) case READ_REQ_REPLY_CODE: success = FALSE; - HDfprintf(stdout, "%s: Received read req reply?!?.\n", + HDfprintf(stdout, "%s: Received read req reply?!?.\n", fcn_name); break; @@ -1299,7 +1299,7 @@ server_main(void) case SYNC_ACK_CODE: success = FALSE; - HDfprintf(stdout, "%s: Received sync ack?!?.\n", + HDfprintf(stdout, "%s: Received sync ack?!?.\n", fcn_name); break; @@ -1317,7 +1317,7 @@ server_main(void) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Unknown request code.\n", + HDfprintf(stdout, "%d:%s: Unknown request code.\n", world_mpi_rank, fcn_name); } break; @@ -1336,7 +1336,7 @@ server_main(void) * * Purpose: Serve a read request. * - * The function accepts a pointer to an instance of struct + * The function accepts a pointer to an instance of struct * mssg_t as input. If all sanity checks pass, it sends * a copy of the indicated datum from the data array to * the requesting process. @@ -1350,7 +1350,7 @@ server_main(void) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -1362,14 +1362,14 @@ serve_read_request(struct mssg_t * mssg_ptr) haddr_t target_addr; struct mssg_t reply; - if ( ( mssg_ptr == NULL ) || + if ( ( mssg_ptr == NULL ) || ( mssg_ptr->req != READ_REQ_CODE ) || ( mssg_ptr->magic != MSSG_MAGIC ) ) { nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Bad mssg on entry.\n", + HDfprintf(stdout, "%d:%s: Bad mssg on entry.\n", world_mpi_rank, fcn_name); } } @@ -1384,7 +1384,7 @@ serve_read_request(struct mssg_t * mssg_ptr) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: addr lookup failed for %a.\n", + HDfprintf(stdout, "%d:%s: addr lookup failed for %a.\n", world_mpi_rank, fcn_name, target_addr); } } else if ( data[target_index].len != mssg_ptr->len ) { @@ -1392,9 +1392,9 @@ serve_read_request(struct mssg_t * mssg_ptr) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: data[i].len = %d != mssg->len = %d.\n", - world_mpi_rank, fcn_name, + world_mpi_rank, fcn_name, data[target_index].len, mssg_ptr->len); } } else if ( ! (data[target_index].valid) ) { @@ -1402,9 +1402,9 @@ serve_read_request(struct mssg_t * mssg_ptr) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: proc %d read invalid entry. idx/base_addr = %d/%a.\n", - world_mpi_rank, fcn_name, + world_mpi_rank, fcn_name, mssg_ptr->src, target_index, target_index, data[target_index].base_addr); @@ -1427,7 +1427,7 @@ serve_read_request(struct mssg_t * mssg_ptr) success = send_mssg(&reply, TRUE); } - + return(success); } /* serve_read_request() */ @@ -1439,12 +1439,12 @@ serve_read_request(struct mssg_t * mssg_ptr) * * Purpose: Serve a sync request. * - * The function accepts a pointer to an instance of struct - * mssg_t as input. If all sanity checks pass, it sends a + * The function accepts a pointer to an instance of struct + * mssg_t as input. If all sanity checks pass, it sends a * sync ack to the requesting process. * * This service exist to allow the sending process to ensure - * that all previous messages have been processed before + * that all previous messages have been processed before * proceeding. * * Return: Success: TRUE @@ -1456,7 +1456,7 @@ serve_read_request(struct mssg_t * mssg_ptr) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -1466,14 +1466,14 @@ serve_sync_request(struct mssg_t * mssg_ptr) hbool_t success = TRUE; struct mssg_t reply; - if ( ( mssg_ptr == NULL ) || + if ( ( mssg_ptr == NULL ) || ( mssg_ptr->req != SYNC_REQ_CODE ) || ( mssg_ptr->magic != MSSG_MAGIC ) ) { nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Bad mssg on entry.\n", + HDfprintf(stdout, "%d:%s: Bad mssg on entry.\n", world_mpi_rank, fcn_name); } } @@ -1485,7 +1485,7 @@ serve_sync_request(struct mssg_t * mssg_ptr) reply.src = world_mpi_rank; reply.dest = mssg_ptr->src; reply.mssg_num = -1; /* set by send function */ - reply.base_addr = 0; + reply.base_addr = 0; reply.len = 0; reply.ver = 0; reply.magic = MSSG_MAGIC; @@ -1495,7 +1495,7 @@ serve_sync_request(struct mssg_t * mssg_ptr) success = send_mssg(&reply, TRUE); } - + return(success); } /* serve_sync_request() */ @@ -1507,7 +1507,7 @@ serve_sync_request(struct mssg_t * mssg_ptr) * * Purpose: Serve a write request. * - * The function accepts a pointer to an instance of struct + * The function accepts a pointer to an instance of struct * mssg_t as input. If all sanity checks pass, it updates * the version number of the target data array entry as * specified in the message. @@ -1521,10 +1521,10 @@ serve_sync_request(struct mssg_t * mssg_ptr) * Modifications: * * JRM -- 5/9/06 - * Added code supporting a write ack message. This is a - * speculative fix to a bug observed on Cobalt. If it + * Added code supporting a write ack message. This is a + * speculative fix to a bug observed on Cobalt. If it * doesn't work, it will help narrow down the possibilities. - * + * *****************************************************************************/ hbool_t @@ -1539,14 +1539,14 @@ serve_write_request(struct mssg_t * mssg_ptr) struct mssg_t reply; #endif /* DO_WRITE_REQ_ACK */ - if ( ( mssg_ptr == NULL ) || + if ( ( mssg_ptr == NULL ) || ( mssg_ptr->req != WRITE_REQ_CODE ) || ( mssg_ptr->magic != MSSG_MAGIC ) ) { nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Bad mssg on entry.\n", + HDfprintf(stdout, "%d:%s: Bad mssg on entry.\n", world_mpi_rank, fcn_name); } } @@ -1561,7 +1561,7 @@ serve_write_request(struct mssg_t * mssg_ptr) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, "%d:%s: addr lookup failed for %a.\n", + HDfprintf(stdout, "%d:%s: addr lookup failed for %a.\n", world_mpi_rank, fcn_name, target_addr); } } else if ( data[target_index].len != mssg_ptr->len ) { @@ -1569,9 +1569,9 @@ serve_write_request(struct mssg_t * mssg_ptr) nerrors++; success = FALSE; if ( verbose ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: data[i].len = %d != mssg->len = %d.\n", - world_mpi_rank, fcn_name, + world_mpi_rank, fcn_name, data[target_index].len, mssg_ptr->len); } } @@ -1587,7 +1587,7 @@ serve_write_request(struct mssg_t * mssg_ptr) success = FALSE; if ( verbose ) { HDfprintf(stdout, "%d:%s: new ver = %d <= old ver = %d.\n", - world_mpi_rank, fcn_name, + world_mpi_rank, fcn_name, new_ver_num, data[target_index].ver); } } @@ -1617,7 +1617,7 @@ serve_write_request(struct mssg_t * mssg_ptr) #endif /* DO_WRITE_REQ_ACK */ } - + return(success); } /* serve_write_request() */ @@ -1630,7 +1630,7 @@ serve_write_request(struct mssg_t * mssg_ptr) /*------------------------------------------------------------------------- * Function: clear_datum * - * Purpose: Mark the datum as clean and destroy it if requested. + * Purpose: Mark the datum as clean and destroy it if requested. * Do not write it to the server, or increment the version. * * Return: SUCCEED @@ -1690,7 +1690,7 @@ clear_datum(H5F_t * f, * Function: destroy_datum() * * Purpose: Destroy the entry. At present, this means do nothing other - * than verify that the entry is clean. In particular, do not + * than verify that the entry is clean. In particular, do not * write it to the server process. * * Return: SUCCEED @@ -1739,7 +1739,7 @@ destroy_datum(H5F_t UNUSED * f, /*------------------------------------------------------------------------- * Function: flush_datum * - * Purpose: Flush the entry to the server process and mark it as clean. + * Purpose: Flush the entry to the server process and mark it as clean. * Then destroy the entry if requested. * * Return: SUCCEED if successful, and FAIL otherwise. @@ -1790,7 +1790,7 @@ flush_datum(H5F_t *f, if ( ( file_mpi_rank != 0 ) && ( entry_ptr->dirty ) ) { ret_value = FAIL; - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: Flushed dirty entry from non-zero file process.", world_mpi_rank, fcn_name); } @@ -1868,7 +1868,7 @@ flush_datum(H5F_t *f, datum_flushes++; if ( entry_ptr->header.is_pinned ) { - + datum_pinned_flushes++; HDassert( entry_ptr->global_pinned || entry_ptr->local_pinned ); } @@ -1880,7 +1880,7 @@ flush_datum(H5F_t *f, /*------------------------------------------------------------------------- * Function: load_datum * - * Purpose: Read the requested entry from the server and mark it as + * Purpose: Read the requested entry from the server and mark it as * clean. * * Return: SUCCEED if successful, FAIL otherwise. @@ -1990,7 +1990,7 @@ load_datum(H5F_t UNUSED *f, if ( mssg.base_addr != entry_ptr->base_addr ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: mssg.base_addr != entry_ptr->base_addr.\n", world_mpi_rank, fcn_name); HDfprintf(stdout, "%d:%s: mssg.base_addr = %a.\n", @@ -2042,7 +2042,7 @@ load_datum(H5F_t UNUSED *f, /*------------------------------------------------------------------------- * Function: size_datum * - * Purpose: Get the size of the specified entry. Just look at the + * Purpose: Get the size of the specified entry. Just look at the * local copy, as size can't change. * * Return: SUCCEED @@ -2131,7 +2131,7 @@ insert_entry(H5C_t * cache_ptr, (entry_ptr->ver)++; entry_ptr->dirty = TRUE; - result = H5AC_set(file_ptr, -1, &(types[0]), entry_ptr->base_addr, + result = H5AC_set(file_ptr, -1, &(types[0]), entry_ptr->base_addr, (void *)(&(entry_ptr->header)), flags); if ( ( result < 0 ) || @@ -2149,7 +2149,7 @@ insert_entry(H5C_t * cache_ptr, if ( ! (entry_ptr->header.is_dirty) ) { /* it is possible that we just exceeded the dirty bytes - * threshold, triggering a write of the newly inserted + * threshold, triggering a write of the newly inserted * entry. Test for this, and only flag an error if this * is not the case. */ @@ -2161,11 +2161,11 @@ insert_entry(H5C_t * cache_ptr, if ( ! ( ( aux_ptr != NULL ) && ( aux_ptr->magic == H5AC__H5AC_AUX_T_MAGIC ) && ( aux_ptr->dirty_bytes == 0 ) ) ) { - + nerrors++; if ( verbose ) { HDfprintf(stdout, "%d:%s: data[%d].header.is_dirty = %d.\n", - world_mpi_rank, fcn_name, idx, + world_mpi_rank, fcn_name, idx, (int)(data[idx].header.is_dirty)); } } @@ -2183,7 +2183,7 @@ insert_entry(H5C_t * cache_ptr, /***************************************************************************** * Function: local_pin_and_unpin_random_entries() * - * Purpose: Pin a random number of randomly selected entries in cache, and + * Purpose: Pin a random number of randomly selected entries in cache, and * then unpin a random number of entries. * * Do nothing if nerrors is non-zero on entry. @@ -2201,8 +2201,8 @@ void local_pin_and_unpin_random_entries(H5C_t * cache_ptr, H5F_t * file_ptr, int min_idx, - int max_idx, - int min_count, + int max_idx, + int min_count, int max_count) { const char * fcn_name = "local_pin_and_unpin_random_entries()"; @@ -2259,8 +2259,8 @@ local_pin_and_unpin_random_entries(H5C_t * cache_ptr, * Function: local_pin_random_entry() * * Purpose: Pin a randomly selected entry in cache, and mark the entry - * as being locally pinned. Since this entry will not in - * general be pinned in any other cache, we can't mark it + * as being locally pinned. Since this entry will not in + * general be pinned in any other cache, we can't mark it * dirty. * * Do nothing if nerrors is non-zero on entry. @@ -2340,7 +2340,7 @@ local_unpin_all_entries(H5C_t * cache_ptr, idx = 0; - while ( idx >= 0 ) + while ( idx >= 0 ) { idx = local_unpin_next_pinned_entry(cache_ptr, file_ptr, idx, via_unprotect); @@ -2360,8 +2360,8 @@ local_unpin_all_entries(H5C_t * cache_ptr, * * Do nothing if nerrors is non-zero on entry. * - * Return: Index of the unpinned entry if there is one, or -1 if - * nerrors is non-zero on entry, or if there is no locally + * Return: Index of the unpinned entry if there is one, or -1 if + * nerrors is non-zero on entry, or if there is no locally * pinned entry. * * Programmer: John Mainzer @@ -2391,7 +2391,7 @@ local_unpin_next_pinned_entry(H5C_t * cache_ptr, idx = start_idx; - while ( ( i < virt_num_data_entries ) && + while ( ( i < virt_num_data_entries ) && ( ! ( data[idx].local_pinned ) ) ) { i++; @@ -2420,7 +2420,7 @@ local_unpin_next_pinned_entry(H5C_t * cache_ptr, * Function: lock_and_unlock_random_entries() * * Purpose: Obtain a random number in the closed interval [min_count, - * max_count]. Then protect and unprotect that number of + * max_count]. Then protect and unprotect that number of * random entries. * * Do nothing if nerrors is non-zero on entry. @@ -2554,7 +2554,7 @@ lock_entry(H5C_t * cache_ptr, HDassert( ! (entry_ptr->locked) ); cache_entry_ptr = H5AC_protect(file_ptr, -1, &(types[0]), - entry_ptr->base_addr, + entry_ptr->base_addr, NULL, NULL, H5AC_WRITE); if ( ( cache_entry_ptr != (void *)(&(entry_ptr->header)) ) || @@ -2633,12 +2633,12 @@ mark_pinned_entry_dirty(H5C_t * cache_ptr, nerrors++; if ( verbose ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: error in H5AC_mark_pinned_entry_dirty().\n", world_mpi_rank, fcn_name); } - } - else + } + else { global_dirty_pins++; } @@ -2700,7 +2700,7 @@ mark_pinned_or_protected_entry_dirty(H5C_t * cache_ptr, world_mpi_rank, fcn_name, "H5AC_mark_pinned_or_protected_entry_dirty()"); } - } + } else if ( ! ( entry_ptr->locked ) ) { global_dirty_pins++; @@ -2891,9 +2891,9 @@ rename_entry(H5C_t * cache_ptr, * Purpose: Setup the parallel cache for a test, and return the file id * and a pointer to the cache's internal data structures. * - * To do this, we must create a file, flush it (so that we + * To do this, we must create a file, flush it (so that we * don't have to worry about entries in the metadata cache), - * look up the address of the metadata cache, and then instruct + * look up the address of the metadata cache, and then instruct * the cache to omit sanity checks on dxpl IDs. * * Return: Success: TRUE @@ -2905,7 +2905,7 @@ rename_entry(H5C_t * cache_ptr, * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -2930,13 +2930,13 @@ setup_cache_for_test(hid_t * fid_ptr, if ( fid < 0 ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Fcreate() failed.\n", + HDfprintf(stdout, "%d:%s: H5Fcreate() failed.\n", world_mpi_rank, fcn_name); } } else if ( H5Fflush(fid, H5F_SCOPE_GLOBAL) < 0 ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Fflush() failed.\n", + HDfprintf(stdout, "%d:%s: H5Fflush() failed.\n", world_mpi_rank, fcn_name); } } else { @@ -2946,7 +2946,7 @@ setup_cache_for_test(hid_t * fid_ptr, if ( file_ptr == NULL ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Can't get file_ptr.\n", + HDfprintf(stdout, "%d:%s: Can't get file_ptr.\n", world_mpi_rank, fcn_name); } } else { @@ -2956,13 +2956,13 @@ setup_cache_for_test(hid_t * fid_ptr, if ( cache_ptr == NULL ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Can't get cache_ptr.\n", + HDfprintf(stdout, "%d:%s: Can't get cache_ptr.\n", world_mpi_rank, fcn_name); } } else if ( cache_ptr->magic != H5C__H5C_T_MAGIC ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: Bad cache_ptr magic.\n", + HDfprintf(stdout, "%d:%s: Bad cache_ptr magic.\n", world_mpi_rank, fcn_name); } } else { @@ -2978,21 +2978,21 @@ setup_cache_for_test(hid_t * fid_ptr, config.version = H5AC__CURR_CACHE_CONFIG_VERSION; - if ( H5AC_get_cache_auto_resize_config(cache_ptr, &config) + if ( H5AC_get_cache_auto_resize_config(cache_ptr, &config) != SUCCEED ) { - HDfprintf(stdout, - "%d:%s: H5AC_get_cache_auto_resize_config() failed.\n", + HDfprintf(stdout, + "%d:%s: H5AC_get_cache_auto_resize_config() failed.\n", world_mpi_rank, fcn_name); } else { config.rpt_fcn_enabled = TRUE; - if ( H5AC_set_cache_auto_resize_config(cache_ptr, &config) + if ( H5AC_set_cache_auto_resize_config(cache_ptr, &config) != SUCCEED ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: H5AC_set_cache_auto_resize_config() failed.\n", world_mpi_rank, fcn_name); } else { @@ -3011,8 +3011,8 @@ setup_cache_for_test(hid_t * fid_ptr, nerrors++; if ( verbose ) { - HDfprintf(stdout, - "%d:%s: H5C_set_write_done_callback failed.\n", + HDfprintf(stdout, + "%d:%s: H5C_set_write_done_callback failed.\n", world_mpi_rank, fcn_name); } } @@ -3029,7 +3029,7 @@ setup_cache_for_test(hid_t * fid_ptr, * * Function: setup_noblock_dxpl_id() * - * Purpose: Setup the noblock_dxpl_id global. Increment nerrors if + * Purpose: Setup the noblock_dxpl_id global. Increment nerrors if * errors are detected. Do nothing if nerrors is non-zero * on entry. * @@ -3040,22 +3040,22 @@ setup_cache_for_test(hid_t * fid_ptr, * Modifications: * * None. - * + * *****************************************************************************/ -/* So far we haven't needed this, but that may change. - * Keep it around for now +/* So far we haven't needed this, but that may change. + * Keep it around for now */ #if 0 void setup_noblock_dxpl_id(void) { const char * fcn_name = "setup_noblock_dxpl_id()"; - H5P_genclass_t *xfer_pclass; /* Dataset transfer property list - * class object + H5P_genclass_t *xfer_pclass; /* Dataset transfer property list + * class object */ H5P_genplist_t *xfer_plist; /* Dataset transfer property list object */ - unsigned block_before_meta_write; /* "block before meta write" - * property value + unsigned block_before_meta_write; /* "block before meta write" + * property value */ unsigned library_internal = 1; /* "library internal" property value */ H5FD_mpio_xfer_t xfer_mode; /* I/O transfer mode property value */ @@ -3100,13 +3100,13 @@ setup_noblock_dxpl_id(void) block_before_meta_write=0; if ( ( nerrors == 0 ) && ( H5P_insert(xfer_plist, H5AC_BLOCK_BEFORE_META_WRITE_NAME, - H5AC_BLOCK_BEFORE_META_WRITE_SIZE, + H5AC_BLOCK_BEFORE_META_WRITE_SIZE, &block_before_meta_write, NULL, NULL, NULL, NULL, NULL, NULL) < 0 ) ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: can't insert metadata cache dxpl property 1.\n", world_mpi_rank, fcn_name); } @@ -3120,7 +3120,7 @@ setup_noblock_dxpl_id(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, + HDfprintf(stdout, "%d:%s: can't insert metadata cache dxpl property 2.\n", world_mpi_rank, fcn_name); } @@ -3133,7 +3133,7 @@ setup_noblock_dxpl_id(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: unable to set value.\n", world_mpi_rank, + HDfprintf(stdout, "%d:%s: unable to set value.\n", world_mpi_rank, fcn_name); } } @@ -3148,7 +3148,7 @@ setup_noblock_dxpl_id(void) * * Function: setup_rand() * - * Purpose: Use gettimeofday() to obtain a seed for rand(), print the + * Purpose: Use gettimeofday() to obtain a seed for rand(), print the * seed to stdout, and then pass it to srand(). * * Increment nerrors if any errors are detected. @@ -3161,7 +3161,7 @@ setup_noblock_dxpl_id(void) * * JRM -- 5/9/06 * Modified function to facilitate setting predefined seeds. - * + * *****************************************************************************/ void @@ -3175,14 +3175,14 @@ setup_rand(void) struct timeval tv; struct timezone tz; - if ( ( use_predefined_seeds ) && + if ( ( use_predefined_seeds ) && ( world_mpi_size == num_predefined_seeds ) ) { HDassert( world_mpi_rank >= 0 ); HDassert( world_mpi_rank < world_mpi_size ); seed = predefined_seeds[world_mpi_rank]; - HDfprintf(stdout, "%d:%s: predefined_seed = %d.\n", + HDfprintf(stdout, "%d:%s: predefined_seed = %d.\n", world_mpi_rank, fcn_name, seed); fflush(stdout); HDsrand(seed); @@ -3198,7 +3198,7 @@ setup_rand(void) } } else { seed = (unsigned)tv.tv_usec; - HDfprintf(stdout, "%d:%s: seed = %d.\n", + HDfprintf(stdout, "%d:%s: seed = %d.\n", world_mpi_rank, fcn_name, seed); fflush(stdout); HDsrand(seed); @@ -3216,7 +3216,7 @@ setup_rand(void) * * Purpose: Take down the parallel cache after a test. * - * To do this, we must close the file, and delete if if + * To do this, we must close the file, and delete if if * possible. * * Return: Success: TRUE @@ -3228,7 +3228,7 @@ setup_rand(void) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -3242,7 +3242,7 @@ take_down_cache(hid_t fid) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Fclose() failed.\n", + HDfprintf(stdout, "%d:%s: H5Fclose() failed.\n", world_mpi_rank, fcn_name); } @@ -3252,7 +3252,7 @@ take_down_cache(hid_t fid) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: HDremove() failed.\n", + HDfprintf(stdout, "%d:%s: HDremove() failed.\n", world_mpi_rank, fcn_name); } } else { @@ -3315,7 +3315,7 @@ unlock_entry(H5C_t * cache_ptr, entry_ptr->dirty = TRUE; } - result = H5AC_unprotect(file_ptr, -1, &(types[0]), + result = H5AC_unprotect(file_ptr, -1, &(types[0]), entry_ptr->base_addr, (void *)(&(entry_ptr->header)), flags); @@ -3462,15 +3462,15 @@ unpin_entry(H5C_t * cache_ptr, * * JRM -- 5/9/06 * Added code supporting the write request ack message. This - * message was added to eliminate one possible cause of a + * message was added to eliminate one possible cause of a * bug spotted on cobalt. If this doesn't fix the problem, * it will narrow things down a bit. * * JRM -- 5/10/06 - * Added call to do_sync(). This is part of an attempt to + * Added call to do_sync(). This is part of an attempt to * optimize out the slowdown caused by the addition of the * write request ack message. - * + * *****************************************************************************/ hbool_t @@ -3518,7 +3518,7 @@ server_smoke_check(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on write.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on write.\n", world_mpi_rank, fcn_name); } } @@ -3534,7 +3534,7 @@ server_smoke_check(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: recv_mssg() failed.\n", + HDfprintf(stdout, "%d:%s: recv_mssg() failed.\n", world_mpi_rank, fcn_name); } } @@ -3582,7 +3582,7 @@ server_smoke_check(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on write.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on write.\n", world_mpi_rank, fcn_name); } } @@ -3597,7 +3597,7 @@ server_smoke_check(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: recv_mssg() failed.\n", + HDfprintf(stdout, "%d:%s: recv_mssg() failed.\n", world_mpi_rank, fcn_name); } } @@ -3641,13 +3641,13 @@ server_smoke_check(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", world_mpi_rank, fcn_name); } } } } - + max_nerrors = get_max_nerrors(); if ( world_mpi_rank == 0 ) { @@ -3684,7 +3684,7 @@ server_smoke_check(void) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -3749,7 +3749,7 @@ smoke_check_1(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); } @@ -3758,7 +3758,7 @@ smoke_check_1(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); } @@ -3774,7 +3774,7 @@ smoke_check_1(void) } } - /* verify that all instance of datum are back where the started + /* verify that all instance of datum are back where the started * and are clean. */ @@ -3802,7 +3802,7 @@ smoke_check_1(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", world_mpi_rank, fcn_name); } } @@ -3853,7 +3853,7 @@ smoke_check_1(void) * * JRM -- 4/28/06 * Modified test to rename pinned entries. - * + * *****************************************************************************/ hbool_t @@ -3927,11 +3927,11 @@ smoke_check_2(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 20), + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 20), 0, 100); - local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 4), + local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 4), 0, 3); } @@ -3939,8 +3939,8 @@ smoke_check_2(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__DIRTIED_FLAG); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 10), + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 10), 0, 100); } @@ -3952,10 +3952,10 @@ smoke_check_2(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, - ((virt_num_data_entries / 50) - 1), + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + ((virt_num_data_entries / 50) - 1), 0, 100); } @@ -3964,10 +3964,10 @@ smoke_check_2(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__DIRTIED_FLAG); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 100), + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 100), 0, 100); } @@ -3991,7 +3991,7 @@ smoke_check_2(void) } } - /* verify that all instance of datum are back where the started + /* verify that all instance of datum are back where the started * and are clean. */ @@ -4019,13 +4019,13 @@ smoke_check_2(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", world_mpi_rank, fcn_name); } } } } - + max_nerrors = get_max_nerrors(); if ( world_mpi_rank == 0 ) { @@ -4054,7 +4054,7 @@ smoke_check_2(void) * * Purpose: Third smoke check for the parallel cache. * - * Use random reads to vary the loads on the diffferent + * Use random reads to vary the loads on the diffferent * processors. Also force different cache size adjustments. * * In this test, load process 0 heavily, and the other @@ -4073,7 +4073,7 @@ smoke_check_2(void) * JRM - 1/31/06 * * Added pinned entry tests. JRM - 4/14/06 - * + * *****************************************************************************/ hbool_t @@ -4136,7 +4136,7 @@ smoke_check_3(void) if ( i > 100 ) { lock_and_unlock_random_entries(cache_ptr, file_ptr, - (i - 100), i, + (i - 100), i, min_count, max_count); } } @@ -4145,8 +4145,8 @@ smoke_check_3(void) min_count = 100 / ((file_mpi_rank + 2) * (file_mpi_rank + 2)); max_count = min_count + 50; - for ( i = (virt_num_data_entries / 4); - i < (virt_num_data_entries / 2); + for ( i = (virt_num_data_entries / 4); + i < (virt_num_data_entries / 2); i++ ) { @@ -4176,7 +4176,7 @@ smoke_check_3(void) } local_pin_and_unpin_random_entries(cache_ptr, file_ptr, - 0, virt_num_data_entries / 4, + 0, virt_num_data_entries / 4, 0, (file_mpi_rank + 2)); } @@ -4188,14 +4188,14 @@ smoke_check_3(void) if ( H5Fflush(fid, H5F_SCOPE_GLOBAL) < 0 ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Fflush() failed.\n", + HDfprintf(stdout, "%d:%s: H5Fflush() failed.\n", world_mpi_rank, fcn_name); } } min_idx = 0; - max_idx = ((virt_num_data_entries / 10) / + max_idx = ((virt_num_data_entries / 10) / ((file_mpi_rank + 1) * (file_mpi_rank + 1))) - 1; if ( max_idx <= min_idx ) { @@ -4212,15 +4212,15 @@ smoke_check_3(void) HDassert( data[i].global_pinned ); HDassert( ! data[i].local_pinned ); - unpin_entry(cache_ptr, file_ptr, i, TRUE, dirty, + unpin_entry(cache_ptr, file_ptr, i, TRUE, dirty, via_unprotect); } if ( i % 2 == 0 ) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, - virt_num_data_entries / 2, + local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, + virt_num_data_entries / 2, 0, 2); lock_and_unlock_random_entries(cache_ptr, file_ptr, min_idx, max_idx, 0, 100); @@ -4228,7 +4228,7 @@ smoke_check_3(void) } min_idx = 0; - max_idx = ((virt_num_data_entries / 10) / + max_idx = ((virt_num_data_entries / 10) / ((file_mpi_rank + 3) * (file_mpi_rank + 3))) - 1; if ( max_idx <= min_idx ) { @@ -4254,10 +4254,10 @@ smoke_check_3(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 20), + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 20), min_count, max_count); } @@ -4266,9 +4266,9 @@ smoke_check_3(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__DIRTIED_FLAG); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, (virt_num_data_entries / 40), min_count, max_count); } @@ -4281,8 +4281,8 @@ smoke_check_3(void) for ( i = 0; i < (virt_num_data_entries / 2); i+=2 ) { - local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 2), + local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 2), 0, 5); lock_entry(cache_ptr, file_ptr, i); @@ -4291,7 +4291,7 @@ smoke_check_3(void) if ( i > 100 ) { lock_and_unlock_random_entries(cache_ptr, file_ptr, - (i - 100), i, + (i - 100), i, min_count, max_count); } } @@ -4311,7 +4311,7 @@ smoke_check_3(void) } } - /* verify that all instances of datum are back where the started + /* verify that all instances of datum are back where the started * and are clean. */ @@ -4340,13 +4340,13 @@ smoke_check_3(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", world_mpi_rank, fcn_name); } } } } - + max_nerrors = get_max_nerrors(); if ( world_mpi_rank == 0 ) { @@ -4375,7 +4375,7 @@ smoke_check_3(void) * * Purpose: Fourth smoke check for the parallel cache. * - * Use random reads to vary the loads on the diffferent + * Use random reads to vary the loads on the diffferent * processors. Also force different cache size adjustments. * * In this test, load process 0 lightly, and the other @@ -4392,7 +4392,7 @@ smoke_check_3(void) * Added code intended to insure correct operation with large * numbers of processors. * JRM - 1/31/06 - * + * *****************************************************************************/ hbool_t @@ -4445,7 +4445,7 @@ smoke_check_4(void) } } - + min_count = 100 * (file_mpi_rank % 4); max_count = min_count + 50; @@ -4456,7 +4456,7 @@ smoke_check_4(void) if ( i > 100 ) { lock_and_unlock_random_entries(cache_ptr, file_ptr, - (i - 100), i, + (i - 100), i, min_count, max_count); } } @@ -4464,8 +4464,8 @@ smoke_check_4(void) min_count = 10 * (file_mpi_rank % 4); max_count = min_count + 100; - for ( i = (virt_num_data_entries / 4); - i < (virt_num_data_entries / 2); + for ( i = (virt_num_data_entries / 4); + i < (virt_num_data_entries / 2); i++ ) { insert_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); @@ -4493,7 +4493,7 @@ smoke_check_4(void) min_count, max_count); } - local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, + local_pin_and_unpin_random_entries(cache_ptr, file_ptr, 0, (virt_num_data_entries / 4), 0, (file_mpi_rank + 2)); } @@ -4512,7 +4512,7 @@ smoke_check_4(void) min_idx = 0; - max_idx = (((virt_num_data_entries / 10) / 4) * + max_idx = (((virt_num_data_entries / 10) / 4) * ((file_mpi_rank % 4) + 1)) - 1; for ( i = (virt_num_data_entries / 2) - 1; i >= 0; i-- ) @@ -4538,7 +4538,7 @@ smoke_check_4(void) } min_idx = 0; - max_idx = (((virt_num_data_entries / 10) / 8) * + max_idx = (((virt_num_data_entries / 10) / 8) * ((file_mpi_rank % 4) + 1)) - 1; for ( i = 0; i < (virt_num_data_entries / 2); i+=2 ) @@ -4560,10 +4560,10 @@ smoke_check_4(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, - (virt_num_data_entries / 20), + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + (virt_num_data_entries / 20), min_count, max_count); } @@ -4572,9 +4572,9 @@ smoke_check_4(void) { lock_entry(cache_ptr, file_ptr, i); unlock_entry(cache_ptr, file_ptr, i, H5AC__DIRTIED_FLAG); - rename_entry(cache_ptr, file_ptr, i, + rename_entry(cache_ptr, file_ptr, i, (i + (virt_num_data_entries / 2))); - lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, + lock_and_unlock_random_entries(cache_ptr, file_ptr, 0, (virt_num_data_entries / 40), min_count, max_count); } @@ -4593,7 +4593,7 @@ smoke_check_4(void) if ( i > 100 ) { lock_and_unlock_random_entries(cache_ptr, file_ptr, - (i - 100), i, + (i - 100), i, min_count, max_count); } } @@ -4610,7 +4610,7 @@ smoke_check_4(void) } } - /* verify that all instance of datum are back where the started + /* verify that all instance of datum are back where the started * and are clean. */ @@ -4639,13 +4639,13 @@ smoke_check_4(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", world_mpi_rank, fcn_name); } } } } - + max_nerrors = get_max_nerrors(); if ( world_mpi_rank == 0 ) { @@ -4685,7 +4685,7 @@ smoke_check_4(void) * Modifications: * * None. - * + * *****************************************************************************/ hbool_t @@ -4743,7 +4743,7 @@ smoke_check_5(void) if ( H5Fflush(fid, H5F_SCOPE_GLOBAL) < 0 ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Fflush() failed.\n", + HDfprintf(stdout, "%d:%s: H5Fflush() failed.\n", world_mpi_rank, fcn_name); } } @@ -4752,7 +4752,7 @@ smoke_check_5(void) { lock_entry(cache_ptr, file_ptr, i); - if ( i % 2 == 0 ) + if ( i % 2 == 0 ) { mark_pinned_or_protected_entry_dirty(cache_ptr, file_ptr, i); } @@ -4760,22 +4760,22 @@ smoke_check_5(void) unlock_entry(cache_ptr, file_ptr, i, H5AC__NO_FLAGS_SET); } - for ( i = (virt_num_data_entries / 2) - 1; - i >= (virt_num_data_entries / 4); + for ( i = (virt_num_data_entries / 2) - 1; + i >= (virt_num_data_entries / 4); i-- ) { pin_entry(cache_ptr, file_ptr, i, TRUE, FALSE); - if ( i % 2 == 0 ) + if ( i % 2 == 0 ) { if ( i % 4 == 0 ) { - mark_pinned_or_protected_entry_dirty(cache_ptr, + mark_pinned_or_protected_entry_dirty(cache_ptr, file_ptr, i); } else { - mark_pinned_entry_dirty(cache_ptr, file_ptr, i, + mark_pinned_entry_dirty(cache_ptr, file_ptr, i, FALSE, (size_t)0); } } @@ -4794,7 +4794,7 @@ smoke_check_5(void) } } - /* verify that all instance of datum are back where the started + /* verify that all instance of datum are back where the started * and are clean. */ @@ -4822,7 +4822,7 @@ smoke_check_5(void) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", + HDfprintf(stdout, "%d:%s: send_mssg() failed on done.\n", world_mpi_rank, fcn_name); } } @@ -4866,7 +4866,7 @@ smoke_check_5(void) * Modifications: * * None. - * + * *****************************************************************************/ int @@ -4928,7 +4928,7 @@ main(int argc, char **argv) /* h5_fixname() will hang some processes don't participate. * - * Thus we set up the fapl global with the world communicator, + * Thus we set up the fapl global with the world communicator, * make our calls to h5_fixname(), discard the fapl, and then * create it again with the file communicator. */ @@ -4937,7 +4937,7 @@ main(int argc, char **argv) if ( FAIL == (fapl = H5Pcreate(H5P_FILE_ACCESS)) ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Pcreate() failed 1.\n", + HDfprintf(stdout, "%d:%s: H5Pcreate() failed 1.\n", world_mpi_rank, fcn_name); } } @@ -4946,20 +4946,20 @@ main(int argc, char **argv) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Pset_fapl_mpio() failed 1.\n", + HDfprintf(stdout, "%d:%s: H5Pset_fapl_mpio() failed 1.\n", world_mpi_rank, fcn_name); } } /* fix the file names */ - for ( u = 0; u < sizeof(FILENAME) / sizeof(FILENAME[0]) - 1; ++u ) + for ( u = 0; u < sizeof(FILENAME) / sizeof(FILENAME[0]) - 1; ++u ) { - if ( h5_fixname(FILENAME[u], fapl, filenames[u], + if ( h5_fixname(FILENAME[u], fapl, filenames[u], sizeof(filenames[u])) == NULL ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: h5_fixname() failed.\n", + HDfprintf(stdout, "%d:%s: h5_fixname() failed.\n", world_mpi_rank, fcn_name); } break; @@ -4970,7 +4970,7 @@ main(int argc, char **argv) if ( H5Pclose(fapl) < 0 ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Pclose() failed.\n", + HDfprintf(stdout, "%d:%s: H5Pclose() failed.\n", world_mpi_rank, fcn_name); } } @@ -4982,7 +4982,7 @@ main(int argc, char **argv) if ( FAIL == (fapl = H5Pcreate(H5P_FILE_ACCESS)) ) { nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Pcreate() failed 2.\n", + HDfprintf(stdout, "%d:%s: H5Pcreate() failed 2.\n", world_mpi_rank, fcn_name); } } @@ -4991,14 +4991,14 @@ main(int argc, char **argv) nerrors++; if ( verbose ) { - HDfprintf(stdout, "%d:%s: H5Pset_fapl_mpio() failed 2.\n", + HDfprintf(stdout, "%d:%s: H5Pset_fapl_mpio() failed 2.\n", world_mpi_rank, fcn_name); } } } setup_rand(); - + max_nerrors = get_max_nerrors(); if ( max_nerrors != 0 ) { @@ -5027,10 +5027,10 @@ main(int argc, char **argv) #endif #if 1 smoke_check_4(); -#endif +#endif #if 1 smoke_check_5(); -#endif +#endif finish: /* make sure all processes are finished before final report, cleanup @@ -5040,7 +5040,7 @@ finish: if (MAINPROCESS){ /* only process 0 reports */ printf("===================================\n"); if (failures){ - printf("***metadata cache tests detected %d failures***\n", + printf("***metadata cache tests detected %d failures***\n", failures); } else{ diff --git a/testpar/t_chunk_alloc.c b/testpar/t_chunk_alloc.c index 7186099..70091db 100644 --- a/testpar/t_chunk_alloc.c +++ b/testpar/t_chunk_alloc.c @@ -12,7 +12,7 @@ * access to either file, you may request a copy from hdfhelp@ncsa.uiuc.edu. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ -/* +/* * This verifies if the storage space allocation methods are compatible between * serial and parallel modes. * @@ -23,7 +23,7 @@ #include "testphdf5.h" static int mpi_size, mpi_rank; -#define DATASETNAME "ExtendibleArray" +#define DATASETNAME "ExtendibleArray" #define CHUNKSIZE 1000 /* #elements per chunk */ #define DSETCHUNKS 20000 #define CLOSE 1 @@ -71,11 +71,11 @@ void create_chunked_dataset(const char *filename, int nchunks, write_type write_pattern) { hid_t file_id, dataset; /* handles */ - hid_t dataspace,memspace; + hid_t dataspace,memspace; hid_t cparms; hsize_t dims[1]; hsize_t maxdims[1] = {H5S_UNLIMITED}; - + hsize_t chunk_dims[1] ={CHUNKSIZE}; hsize_t count[1]; hsize_t stride[1]; @@ -85,7 +85,7 @@ create_chunked_dataset(const char *filename, int nchunks, write_type write_patte char buffer[CHUNKSIZE]; int i; - herr_t hrc; + herr_t hrc; MPI_Offset filesize, /* actual file size */ est_filesize; /* estimated file size */ @@ -99,12 +99,12 @@ create_chunked_dataset(const char *filename, int nchunks, write_type write_patte dims[0]=nchunks*CHUNKSIZE; /* Create the data space with unlimited dimensions. */ - dataspace = H5Screate_simple (1, dims, maxdims); + dataspace = H5Screate_simple (1, dims, maxdims); VRFY((dataspace >= 0), ""); memspace = H5Screate_simple(1, chunk_dims, NULL); VRFY((memspace >= 0), ""); - + /* Create a new file. If file exists its contents will be overwritten. */ file_id = H5Fcreate(h5_rmprefix(filename), H5F_ACC_TRUNC, H5P_DEFAULT, H5P_DEFAULT); @@ -183,7 +183,7 @@ create_chunked_dataset(const char *filename, int nchunks, write_type write_patte */ MPI_Barrier(MPI_COMM_WORLD); -} +} /* @@ -242,7 +242,7 @@ parallel_access_dataset(const char *filename, int nchunks, access_type action, h memspace = H5Screate_simple(1, chunk_dims, NULL); VRFY((memspace >= 0), ""); - dataspace = H5Dget_space(*dataset); + dataspace = H5Dget_space(*dataset); VRFY((dataspace >= 0), ""); size[0] = nchunks*CHUNKSIZE; @@ -256,7 +256,7 @@ parallel_access_dataset(const char *filename, int nchunks, access_type action, h count[0] = 1; stride[0] = 1; block[0] = chunk_dims[0]; - for (i=0; i<(nchunks+mpi_size-1)/mpi_size; i++){ + for (i=0; i<(nchunks+mpi_size-1)/mpi_size; i++){ if (i*mpi_size+mpi_rank < nchunks){ offset[0] = (i*mpi_size+mpi_rank)*chunk_dims[0]; @@ -373,14 +373,14 @@ void verify_data(const char *filename, int nchunks, write_type write_pattern, in memspace = H5Screate_simple(1, chunk_dims, NULL); VRFY((memspace >= 0), ""); - dataspace = H5Dget_space(*dataset); + dataspace = H5Dget_space(*dataset); VRFY((dataspace >= 0), ""); /* all processes check all chunks. */ count[0] = 1; stride[0] = 1; block[0] = chunk_dims[0]; - for (i=0; i<nchunks; i++){ + for (i=0; i<nchunks; i++){ /* reset buffer values */ memset(buffer, -1, CHUNKSIZE); @@ -411,7 +411,7 @@ void verify_data(const char *filename, int nchunks, write_type write_pattern, in /* verify content of the chunk */ for (index = 0; index < CHUNKSIZE; index++) VRFY((buffer[index] == value), "data verification"); - + } hrc = H5Sclose (dataspace); @@ -469,7 +469,7 @@ test_chunk_alloc(void) hid_t file_id, dataset; file_id = dataset = -1; - + filename = GetTestParameters(); if (VERBOSE_MED) printf("Extend Chunked allocation test on file %s\n", filename); @@ -480,7 +480,7 @@ test_chunk_alloc(void) /* reopen dataset in parallel and check for file size */ parallel_access_dataset(filename, DSETCHUNKS, open_only, &file_id, &dataset); /* reopen dataset in parallel, read and verify the data */ - verify_data(filename, DSETCHUNKS, none, CLOSE, &file_id, &dataset); + verify_data(filename, DSETCHUNKS, none, CLOSE, &file_id, &dataset); /* Case 2 */ /* Create chunked dataset without writing anything */ @@ -494,10 +494,10 @@ test_chunk_alloc(void) /* Create chunked dataset and write in the second to last chunk */ create_chunked_dataset(filename, DSETCHUNKS, sec_last); /* Reopen dataset in parallel, read and verify the data. The file and dataset are not closed*/ - verify_data(filename, DSETCHUNKS, sec_last, NO_CLOSE, &file_id, &dataset); + verify_data(filename, DSETCHUNKS, sec_last, NO_CLOSE, &file_id, &dataset); /* All processes write in all the chunks in a interleaved way */ parallel_access_dataset(filename, DSETCHUNKS, write_all, &file_id, &dataset); /* reopen dataset in parallel, read and verify the data */ - verify_data(filename, DSETCHUNKS, all, CLOSE, &file_id, &dataset); + verify_data(filename, DSETCHUNKS, all, CLOSE, &file_id, &dataset); } diff --git a/testpar/t_coll_chunk.c b/testpar/t_coll_chunk.c index 5fe6207..1476fc7 100644 --- a/testpar/t_coll_chunk.c +++ b/testpar/t_coll_chunk.c @@ -35,7 +35,7 @@ static void coll_chunktest(const char* filename,int chunk_factor,int select_fact /*------------------------------------------------------------------------- * Function: coll_chunk1 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with a single chunk * * Return: Success: 0 @@ -52,7 +52,7 @@ static void coll_chunktest(const char* filename,int chunk_factor,int select_fact /* ------------------------------------------------------------------------ * Descriptions for the selection: One big singluar selection inside one chunk - * Two dimensions, + * Two dimensions, * * dim1 = SPACE_DIM1(5760)*mpi_size * dim2 = SPACE_DIM2(3) @@ -82,7 +82,7 @@ coll_chunk1(void) /*------------------------------------------------------------------------- * Function: coll_chunk2 * - * Purpose: Wrapper to test the collective chunk IO for regular DISJOINT + * Purpose: Wrapper to test the collective chunk IO for regular DISJOINT selection with a single chunk * * Return: Success: 0 @@ -99,7 +99,7 @@ coll_chunk1(void) /* ------------------------------------------------------------------------ * Descriptions for the selection: many disjoint selections inside one chunk - * Two dimensions, + * Two dimensions, * * dim1 = SPACE_DIM1*mpi_size(5760) * dim2 = SPACE_DIM2(3) @@ -111,7 +111,7 @@ coll_chunk1(void) * count1 = SPACE_DIM2/stride(3/3 = 1) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ void @@ -129,7 +129,7 @@ coll_chunk2(void) /*------------------------------------------------------------------------- * Function: coll_chunk3 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -158,7 +158,7 @@ coll_chunk2(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -180,7 +180,7 @@ coll_chunk3(void) /*------------------------------------------------------------------------- * Function: coll_chunk4 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -209,7 +209,7 @@ coll_chunk3(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -230,7 +230,7 @@ coll_chunk4(void) /*------------------------------------------------------------------------- * Function: coll_chunk4 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -259,7 +259,7 @@ coll_chunk4(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -280,7 +280,7 @@ coll_chunk5(void) /*------------------------------------------------------------------------- * Function: coll_chunk6 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -309,7 +309,7 @@ coll_chunk5(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -330,7 +330,7 @@ coll_chunk6(void) /*------------------------------------------------------------------------- * Function: coll_chunk7 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -359,7 +359,7 @@ coll_chunk6(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -380,7 +380,7 @@ coll_chunk7(void) /*------------------------------------------------------------------------- * Function: coll_chunk8 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -409,7 +409,7 @@ coll_chunk7(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -430,7 +430,7 @@ coll_chunk8(void) /*------------------------------------------------------------------------- * Function: coll_chunk9 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -459,7 +459,7 @@ coll_chunk8(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -480,7 +480,7 @@ coll_chunk9(void) /*------------------------------------------------------------------------- * Function: coll_chunk10 * - * Purpose: Wrapper to test the collective chunk IO for regular JOINT + * Purpose: Wrapper to test the collective chunk IO for regular JOINT selection with at least number of 2*mpi_size chunks * * Return: Success: 0 @@ -509,7 +509,7 @@ coll_chunk9(void) * count1 = SPACE_DIM2(3) * start0 = mpi_rank*SPACE_DIM1 * start1 = 0 - * + * * ------------------------------------------------------------------------ */ @@ -534,7 +534,7 @@ coll_chunk10(void) * Purpose: The real testing routine for regular selection of collective chunking storage testing both write and read, - If anything fails, it may be read or write. There is no + If anything fails, it may be read or write. There is no separation test between read and write. * * Return: Success: 0 @@ -613,7 +613,7 @@ coll_chunktest(const char* filename, chunk_dims[0] = dims[0]/chunk_factor; /* to decrease the testing time, maintain bigger chunk size */ - + (chunk_factor == 1) ? (chunk_dims[1] = SPACE_DIM2) : (chunk_dims[1] = SPACE_DIM2/2); status = H5Pset_chunk(crp_plist, 2, chunk_dims); VRFY((status >= 0),"chunk creation property list succeeded"); @@ -671,8 +671,8 @@ coll_chunktest(const char* filename, break; default: ; - } - + } + #ifdef H5_HAVE_INSTRUMENTED_LIBRARY if(facc_type == FACC_MPIO) { switch(api_option){ @@ -683,44 +683,44 @@ coll_chunktest(const char* filename, VRFY((status >= 0),"testing property list inserted succeeded"); break; case API_MULTI_HARD: - prop_value = H5D_XFER_COLL_CHUNK_DEF; + prop_value = H5D_XFER_COLL_CHUNK_DEF; status = H5Pinsert(xfer_plist,H5D_XFER_COLL_CHUNK_MULTI_HARD_NAME,H5D_XFER_COLL_CHUNK_SIZE,&prop_value, NULL,NULL,NULL,NULL,NULL,NULL); VRFY((status >= 0),"testing property list inserted succeeded"); break; case API_LINK_TRUE: - prop_value = H5D_XFER_COLL_CHUNK_DEF; + prop_value = H5D_XFER_COLL_CHUNK_DEF; status = H5Pinsert(xfer_plist,H5D_XFER_COLL_CHUNK_LINK_NUM_TRUE_NAME,H5D_XFER_COLL_CHUNK_SIZE,&prop_value, NULL,NULL,NULL,NULL,NULL,NULL); VRFY((status >= 0),"testing property list inserted succeeded"); break; case API_LINK_FALSE: - prop_value = H5D_XFER_COLL_CHUNK_DEF; + prop_value = H5D_XFER_COLL_CHUNK_DEF; status = H5Pinsert(xfer_plist,H5D_XFER_COLL_CHUNK_LINK_NUM_FALSE_NAME,H5D_XFER_COLL_CHUNK_SIZE,&prop_value, NULL,NULL,NULL,NULL,NULL,NULL); VRFY((status >= 0),"testing property list inserted succeeded"); - + break; case API_MULTI_COLL: - prop_value = H5D_XFER_COLL_CHUNK_DEF; + prop_value = H5D_XFER_COLL_CHUNK_DEF; status = H5Pinsert(xfer_plist,H5D_XFER_COLL_CHUNK_MULTI_RATIO_COLL_NAME,H5D_XFER_COLL_CHUNK_SIZE,&prop_value, NULL,NULL,NULL,NULL,NULL,NULL); VRFY((status >= 0),"testing property list inserted succeeded"); - + break; case API_MULTI_IND: - prop_value = H5D_XFER_COLL_CHUNK_DEF; + prop_value = H5D_XFER_COLL_CHUNK_DEF; status = H5Pinsert(xfer_plist,H5D_XFER_COLL_CHUNK_MULTI_RATIO_IND_NAME,H5D_XFER_COLL_CHUNK_SIZE,&prop_value, NULL,NULL,NULL,NULL,NULL,NULL); VRFY((status >= 0),"testing property list inserted succeeded"); - + break; default: ; - } + } } #endif - + /* write data collectively */ status = H5Dwrite(dataset, H5T_NATIVE_INT, H5S_ALL, file_dataspace, xfer_plist, data_array1); @@ -761,7 +761,7 @@ coll_chunktest(const char* filename, break; default: ; - } + } } #endif @@ -779,7 +779,7 @@ coll_chunktest(const char* filename, if (data_array1) HDfree(data_array1); - + /* Use collective read to verify the correctness of collective write. */ /* allocate memory for data buffer */ @@ -851,12 +851,12 @@ coll_chunktest(const char* filename, /* Set up the selection */ static void -ccslab_set(int mpi_rank, - int mpi_size, - hsize_t start[], +ccslab_set(int mpi_rank, + int mpi_size, + hsize_t start[], hsize_t count[], - hsize_t stride[], - hsize_t block[], + hsize_t stride[], + hsize_t block[], int mode) { @@ -905,7 +905,7 @@ ccslab_set(int mpi_rank, case BYROW_SELECTUNBALANCE: /* The first one-third of the number of processes only select top half of the domain, The rest will select the bottom - half of the domain. */ + half of the domain. */ block[0] = 1; count[0] = 2; @@ -915,9 +915,9 @@ ccslab_set(int mpi_rank, start[1] = 0; stride[1] = 1; if((mpi_rank *3)<(mpi_size*2)) start[0] = mpi_rank; - else start[0] = 1 + SPACE_DIM1*mpi_size/2 + (mpi_rank-2*mpi_size/3); + else start[0] = 1 + SPACE_DIM1*mpi_size/2 + (mpi_rank-2*mpi_size/3); break; - + case BYROW_SELECTINCHUNK: /* Each process will only select one chunk */ @@ -959,10 +959,10 @@ ccslab_set(int mpi_rank, * Assume dimension rank is 2. */ static void -ccdataset_fill(hsize_t start[], +ccdataset_fill(hsize_t start[], hsize_t stride[], - hsize_t count[], - hsize_t block[], + hsize_t count[], + hsize_t block[], DATATYPE * dataset) { DATATYPE *dataptr = dataset; @@ -994,8 +994,8 @@ ccdataset_fill(hsize_t start[], * Print the first block of the content of the dataset. */ static void -ccdataset_print(hsize_t start[], - hsize_t block[], +ccdataset_print(hsize_t start[], + hsize_t block[], DATATYPE * dataset) { @@ -1025,11 +1025,11 @@ ccdataset_print(hsize_t start[], * Print the content of the dataset. */ static int -ccdataset_vrfy(hsize_t start[], - hsize_t count[], - hsize_t stride[], - hsize_t block[], - DATATYPE *dataset, +ccdataset_vrfy(hsize_t start[], + hsize_t count[], + hsize_t stride[], + hsize_t block[], + DATATYPE *dataset, DATATYPE *original) { hsize_t i, j,k1,k2; diff --git a/testpar/t_mpi.c b/testpar/t_mpi.c index 1ec47f5..a150213 100644 --- a/testpar/t_mpi.c +++ b/testpar/t_mpi.c @@ -833,7 +833,7 @@ static int test_mpio_derived_dtype(char *filename) { Function: test_mpio_special_collective -Test Whether collective IO is still working when more than one process +Test Whether collective IO is still working when more than one process has no contribution to IO. To properly test this case, at least FOUR processes are needed. @@ -867,7 +867,7 @@ static int test_mpio_special_collective(char *filename) { char writedata[2]; char *buf; char expect_val; - int i, irank; + int i, irank; int count,bufcount; int blocklens[2]; MPI_Aint offsets[2]; @@ -883,7 +883,7 @@ static int test_mpio_special_collective(char *filename) { /* create MPI data type */ etype = MPI_BYTE; if(mpi_rank == 0 || mpi_rank == 1) { - count = DIMSIZE; + count = DIMSIZE; bufcount = 1; } else { @@ -895,8 +895,8 @@ static int test_mpio_special_collective(char *filename) { offsets[0] = mpi_rank*count; blocklens[1] = count; offsets[1] = (mpi_size+mpi_rank)*count; - - if(count !=0) { + + if(count !=0) { if((mpi_err= MPI_Type_hindexed(2,blocklens,offsets,etype,&filetype)) != MPI_SUCCESS){ MPI_Error_string(mpi_err, mpi_err_str, &mpi_err_strlen); @@ -969,7 +969,7 @@ static int test_mpio_special_collective(char *filename) { printf("MPI_File_close failed. \n"); return 1; }; - + mpi_err = MPI_Barrier(MPI_COMM_WORLD); #ifdef H5_MPI_SPECIAL_COLLECTIVE_IO_WORKS if(retcode != 0) { @@ -1208,14 +1208,14 @@ main(int argc, char **argv) */ if(mpi_size !=4){ MPI_BANNER("MPIO special collective io test SKIPPED."); - if(mpi_rank == 0){ + if(mpi_rank == 0){ printf("Use FOUR processes to run this test\n"); printf("If you still see the <test SKIPPED>, use <-vh> option to verify the test\n"); } ret_code = 0; goto sc_finish; } - + #ifdef H5_MPI_SPECIAL_COLLECTIVE_IO_WORKS MPI_BANNER("MPIO special collective io test..."); ret_code = test_mpio_special_collective(filenames[0]); @@ -1230,7 +1230,7 @@ main(int argc, char **argv) } #endif -sc_finish: +sc_finish: ret_code = errors_sum(ret_code); if (mpi_rank==0 && ret_code > 0){ printf("***FAILED with %d total errors\n", ret_code); diff --git a/testpar/t_posix_compliant.c b/testpar/t_posix_compliant.c index 6a54cfd..0008b05 100644 --- a/testpar/t_posix_compliant.c +++ b/testpar/t_posix_compliant.c @@ -11,23 +11,23 @@ * http://hdf.ncsa.uiuc.edu/HDF5/doc/Copyright.html. If you do not have * * access to either file, you may request a copy from hdfhelp@ncsa.uiuc.edu. * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * * */ - + /* A series of tests for posix compliance * - * These tests do increasingly complicated sets of writes followed by reads. + * These tests do increasingly complicated sets of writes followed by reads. * POSIX standards say that any read that can be proven to occur after a write * must include the data in that write. These tests attempt to verify whether the * underlying filesystem and i/o layer provide such guarantees. * * There are two sets of tests, one which uses POSIX i/o (fread, fwrite) and one which * uses MPI I/O (MPI_File_read, MPI_File_write). Each set has multiple sub-tests, which - * test varying patters of writes and reads. - * + * test varying patters of writes and reads. + * * * TODO: * Add corresponding posix i/o tests for each MPI i/o test. Currently, not all of the * MPI IO tests are implemented using fwrite/fread. - * + * * Leon Arber * larber@ncsa.uiuc.edu */ @@ -38,7 +38,7 @@ * then run it as an MPI application. E.g., * mpirun -np 3 ./t_posix_compliant */ - + #include <unistd.h> #include <stdio.h> #include <string.h> @@ -90,7 +90,7 @@ static void vrfy_elements(int* a, int* b, int size, int rank); static int find_writesize(int rank, int numprocs, int write_size); -/* All writes are to non-overlapping locations in the file +/* All writes are to non-overlapping locations in the file * Then, each task reads another tasks' data * */ @@ -122,7 +122,7 @@ static int allwrite_allread_blocks(int numprocs, int rank, int write_size) mpio_result = MPI_File_read_at(fh, offset, readbuf, write_size, MPI_INT, &Status); CHECK_SUCCESS(mpio_result); - + vrfy_elements(writebuf, readbuf, write_size, rank); mpio_result = MPI_File_close(&fh); @@ -172,9 +172,9 @@ static int posix_allwrite_allread_blocks(int numprocs, int rank, int write_size) perror("fwrite"); MPI_Abort(MPI_COMM_WORLD, 1); } - + MPI_Barrier(MPI_COMM_WORLD); - + offset = ( (rank+(numprocs-1)) % numprocs)*write_size*sizeof(int); ret = fseek(file, offset, SEEK_SET); @@ -192,7 +192,7 @@ static int posix_allwrite_allread_blocks(int numprocs, int rank, int write_size) perror("Error encountered in fread"); MPI_Abort(MPI_COMM_WORLD, 1); } - + vrfy_elements(writebuf, readbuf, write_size, rank); fclose(file); @@ -251,9 +251,9 @@ static int posix_onewrite_allread_blocks(int numprocs, int rank, int write_size) } } - } + } MPI_Barrier(MPI_COMM_WORLD); - + offset = rank*write_size*sizeof(int); ret = fseek(file, offset, SEEK_SET); @@ -271,7 +271,7 @@ static int posix_onewrite_allread_blocks(int numprocs, int rank, int write_size) perror("Error encountered in fread"); MPI_Abort(MPI_COMM_WORLD, 1); } - + vrfy_elements(writebuf, readbuf, write_size, rank); fclose(file); @@ -331,9 +331,9 @@ static int posix_onewrite_allread_interlaced(int numprocs, int rank, int write_s } } - } + } MPI_Barrier(MPI_COMM_WORLD); - + index = 0; for(offset = rank*sizeof(int); offset<numprocs*write_size*sizeof(int); offset+=(numprocs*sizeof(int))) { @@ -356,7 +356,7 @@ static int posix_onewrite_allread_interlaced(int numprocs, int rank, int write_s index++; } - + for(i=0; i<write_size; i++) writebuf[i] = i; @@ -393,7 +393,7 @@ static int allwrite_allread_interlaced(int numprocs, int rank, int write_size) writebuf[i] = i; - amode = MPI_MODE_CREATE | MPI_MODE_RDWR | MPI_MODE_DELETE_ON_CLOSE; + amode = MPI_MODE_CREATE | MPI_MODE_RDWR | MPI_MODE_DELETE_ON_CLOSE; mpio_result = MPI_File_open(MPI_COMM_WORLD, testfile, amode, MPI_INFO_NULL, &fh); CHECK_SUCCESS(mpio_result); @@ -402,15 +402,15 @@ static int allwrite_allread_interlaced(int numprocs, int rank, int write_size) mpio_result = MPI_Type_commit(&filetype); CHECK_SUCCESS(mpio_result); - + mpio_result = MPI_File_set_view(fh, rank*sizeof(int), MPI_INT, filetype, "native", MPI_INFO_NULL); CHECK_SUCCESS(mpio_result); - + mpio_result = MPI_File_write(fh, writebuf, write_size, MPI_INT, &Status); CHECK_SUCCESS(mpio_result); MPI_Barrier(MPI_COMM_WORLD); - + mpio_result = MPI_File_set_view(fh, 0, MPI_BYTE, MPI_BYTE, "native", MPI_INFO_NULL); CHECK_SUCCESS(mpio_result); @@ -446,7 +446,7 @@ static int allwrite_allread_interlaced(int numprocs, int rank, int write_size) /* Overlapping pattern works as follows (this test requires at least 2 procs: * Writes: - * Task 0: 0 2 4 6 etc... + * Task 0: 0 2 4 6 etc... * Task 1: 1 3 5 7 etc... * Task 2: 0 3 6 etc.. * Task 3: 0 4 8 etc... @@ -477,7 +477,7 @@ static int allwrite_allread_interlaced(int numprocs, int rank, int write_size) static int allwrite_allread_overlap(int numprocs, int rank, int write_size) { - + MPI_File fh = MPI_FILE_NULL; int mpio_result; int amode, i, counter = 0; @@ -491,7 +491,7 @@ static int allwrite_allread_overlap(int numprocs, int rank, int write_size) fprintf(stderr, "The allwrite_allread_overlap test requires at least 2 procs\n"); return -1; } - + if(rank == 0) { for(i=0; i<write_size*(numprocs-1); i++) @@ -511,7 +511,7 @@ static int allwrite_allread_overlap(int numprocs, int rank, int write_size) amode = MPI_MODE_CREATE | MPI_MODE_RDWR | MPI_MODE_DELETE_ON_CLOSE; mpio_result = MPI_File_open(MPI_COMM_WORLD, testfile, amode, MPI_INFO_NULL, &fh); CHECK_SUCCESS(mpio_result); - + if( (rank == 0) || (rank == 1) ) mpio_result = MPI_Type_vector(write_size*(numprocs-1), 1, 2, MPI_INT, &filetype); else @@ -520,13 +520,13 @@ static int allwrite_allread_overlap(int numprocs, int rank, int write_size) mpio_result = MPI_Type_commit(&filetype); CHECK_SUCCESS(mpio_result); - + if( rank == 1) mpio_result = MPI_File_set_view(fh, sizeof(int), MPI_INT, filetype, "native", MPI_INFO_NULL); - else + else mpio_result = MPI_File_set_view(fh, 0, MPI_INT, filetype, "native", MPI_INFO_NULL); CHECK_SUCCESS(mpio_result); - + if( rank == (numprocs - 1)) mpio_result = MPI_File_write(fh, writebuf, write_size, MPI_INT, &Status); else @@ -544,7 +544,7 @@ static int allwrite_allread_overlap(int numprocs, int rank, int write_size) for(i=0; i<write_size*(numprocs-1); i++) { if(i != readbuf[i]) - { + { if( (rank == 0) && (counter == 0)) printf("\n"); if(counter++ < max_err_print) @@ -603,7 +603,7 @@ static int onewrite_allread_blocks(int numprocs, int rank, int write_size) CHECK_SUCCESS(mpio_result); vrfy_elements(writebuf, readbuf, write_size, rank); - + mpio_result = MPI_File_close(&fh); CHECK_SUCCESS(mpio_result); HDfree(writebuf); @@ -634,17 +634,17 @@ static int onewrite_allread_interlaced(int numprocs, int rank, int write_size) amode = MPI_MODE_CREATE | MPI_MODE_RDWR | MPI_MODE_DELETE_ON_CLOSE; mpio_result = MPI_File_open(MPI_COMM_WORLD, testfile, amode, MPI_INFO_NULL, &fh); CHECK_SUCCESS(mpio_result); - + mpio_result = MPI_Type_vector(write_size, 1, numprocs, MPI_INT, &filetype); CHECK_SUCCESS(mpio_result); mpio_result = MPI_Type_commit(&filetype); CHECK_SUCCESS(mpio_result); - + if(rank == (rand() % numprocs)) { for(i=0; i<write_size; i++) - { + { int j; for(j=0; j<numprocs; j++) writebuf[j] = i; @@ -655,7 +655,7 @@ static int onewrite_allread_interlaced(int numprocs, int rank, int write_size) } MPI_Barrier(MPI_COMM_WORLD); - + mpio_result = MPI_File_set_view(fh, rank*sizeof(int), MPI_INT, filetype, "native", MPI_INFO_NULL); CHECK_SUCCESS(mpio_result); @@ -664,7 +664,7 @@ static int onewrite_allread_interlaced(int numprocs, int rank, int write_size) for(i=0; i<write_size; i++) writebuf[i] = i; - + vrfy_elements(writebuf, readbuf, write_size, rank); mpio_result = MPI_File_close(&fh); @@ -684,9 +684,9 @@ int main(int argc, char* argv[]) int write_size = 0; char optstring[] = "h x m p: s: v:"; char *prefix; - + err_flag = 0; - + MPI_Init(&argc, &argv); MPI_Comm_size(MPI_COMM_WORLD, &numprocs); MPI_Comm_rank(MPI_COMM_WORLD, &rank); @@ -728,7 +728,7 @@ int main(int argc, char* argv[]) break; } } - + if( (optind < argc) && (rank == 0)) fprintf(stderr, "Unkown command-line argument passed. Continuing anyway...\n"); @@ -749,7 +749,7 @@ int main(int argc, char* argv[]) } } printf("Proc %d: testfile=%s\n", rank, testfile); - + if(write_size == 0) { lb = 1024; @@ -765,7 +765,7 @@ int main(int argc, char* argv[]) ub = write_size+1; inc = 2; } - + #ifndef STANDALONE /* set alarm. */ ALARM_ON; @@ -775,7 +775,7 @@ int main(int argc, char* argv[]) { if(rank == 0) printf("\nTesting size %d\n", write_size); - + if(mpi_tests) { if(rank == 0) @@ -801,7 +801,7 @@ int main(int argc, char* argv[]) onewrite_allread_blocks(numprocs, rank, write_size/(numprocs*sizeof(int))); PRINT_RESULT(); MPI_Barrier(MPI_COMM_WORLD); - + if(rank == 0) printf("Testing onewrite_allread_interlaced with MPI IO\t\t"); fflush(stdout); onewrite_allread_interlaced(numprocs, rank, write_size/(numprocs*sizeof(int))); @@ -828,7 +828,7 @@ int main(int argc, char* argv[]) posix_onewrite_allread_interlaced(numprocs, rank, write_size/(numprocs*sizeof(int))); PRINT_RESULT(); MPI_Barrier(MPI_COMM_WORLD); - + /* if(rank == 0) printf("Testing allwrite_allread_overlap with POSIX IO\t\t"); fflush(stdout); posix_allwrite_allread_overlap(numprocs, rank, write_size); @@ -860,7 +860,7 @@ static int find_writesize(int rank, int numprocs, int size) int tmp = (size-1)*numprocs; int x = 0; int write_size = 0; - + /* Find largest multiple not greater than tmp */ while(x <= tmp) { @@ -871,7 +871,7 @@ static int find_writesize(int rank, int numprocs, int size) write_size++; } - + return write_size; } diff --git a/testpar/t_span_tree.c b/testpar/t_span_tree.c index 3ead9d8..8c60da3 100644 --- a/testpar/t_span_tree.c +++ b/testpar/t_span_tree.c @@ -45,7 +45,7 @@ static void coll_read_test(int chunk_factor); /*------------------------------------------------------------------------- * Function: coll_irregular_cont_write * - * Purpose: Wrapper to test the collectively irregular hyperslab write in + * Purpose: Wrapper to test the collectively irregular hyperslab write in contiguous storage * * Return: Success: 0 @@ -72,7 +72,7 @@ coll_irregular_cont_write(void) /*------------------------------------------------------------------------- * Function: coll_irregular_cont_read * - * Purpose: Wrapper to test the collectively irregular hyperslab read in + * Purpose: Wrapper to test the collectively irregular hyperslab read in contiguous storage * * Return: Success: 0 @@ -98,7 +98,7 @@ coll_irregular_cont_read(void) /*------------------------------------------------------------------------- * Function: coll_irregular_simple_chunk_write * - * Purpose: Wrapper to test the collectively irregular hyperslab write in + * Purpose: Wrapper to test the collectively irregular hyperslab write in chunk storage(1 chunk) * * Return: Success: 0 @@ -206,7 +206,7 @@ coll_irregular_complex_chunk_read(void) * Purpose: To test the collectively irregular hyperslab write in chunk storage * Input: number of chunks on each dimension - if number is equal to 0, contiguous storage + if number is equal to 0, contiguous storage * Return: Success: 0 * * Failure: -1 @@ -282,7 +282,7 @@ void coll_write_test(int chunk_factor) mdim[1] = MSPACE_DIM2*mpi_size; fsdim[0] = FSPACE_DIM1; fsdim[1] = FSPACE_DIM2*mpi_size; - + vector = (int*)HDmalloc(sizeof(int)*mdim1[0]*mpi_size); matrix_out = (int*)HDmalloc(sizeof(int)*mdim[0]*mdim[1]*mpi_size); matrix_out1 = (int*)HDmalloc(sizeof(int)*mdim[0]*mdim[1]*mpi_size); @@ -341,7 +341,7 @@ void coll_write_test(int chunk_factor) /* The First selection for FILE * * block (3,2) - * stride(4,3) + * stride(4,3) * count (1,768/mpi_size) * start (0,1+768*3*mpi_rank/mpi_size) * @@ -360,10 +360,10 @@ void coll_write_test(int chunk_factor) VRFY((ret >= 0),"hyperslab selection succeeded"); /* The Second selection for FILE - * + * * block (3,768) * stride (1,1) - * count (1,1) + * count (1,1) * start (4,768*mpi_rank/mpi_size) * */ @@ -454,11 +454,11 @@ void coll_write_test(int chunk_factor) * Open the file. */ - /*** - - For testing collective hyperslab selection write + /*** + + For testing collective hyperslab selection write In this test, we are using independent read to check - the correctedness of collective write compared with + the correctedness of collective write compared with independent write, In order to throughly test this feature, we choose @@ -496,7 +496,7 @@ void coll_write_test(int chunk_factor) /* The First selection for FILE to read * * block (1,1) - * stride(1.1) + * stride(1.1) * count (3,768/mpi_size) * start (1,2+768*mpi_rank/mpi_size) * @@ -522,7 +522,7 @@ void coll_write_test(int chunk_factor) /* The Second selection for FILE to read * * block (1,1) - * stride(1.1) + * stride(1.1) * count (3,1536/mpi_size) * start (2,4+1536*mpi_rank/mpi_size) * @@ -560,7 +560,7 @@ void coll_write_test(int chunk_factor) * Only the starting point is different. * The first selection * block (1,1) - * stride(1.1) + * stride(1.1) * count (3,768/mpi_size) * start (0,768*mpi_rank/mpi_size) * @@ -585,7 +585,7 @@ void coll_write_test(int chunk_factor) * Only the starting point is different. * The second selection * block (1,1) - * stride(1,1) + * stride(1,1) * count (3,1536/mpi_size) * start (1,2+1536*mpi_rank/mpi_size) * @@ -616,7 +616,7 @@ void coll_write_test(int chunk_factor) H5P_DEFAULT, matrix_out); VRFY((ret >= 0),"H5D independent read succeed"); - + ret = H5Dread(dataseti, H5T_NATIVE_INT, mspaceid, fspaceid, H5P_DEFAULT, matrix_out1); VRFY((ret >= 0),"H5D independent read succeed"); @@ -627,7 +627,7 @@ void coll_write_test(int chunk_factor) if(matrix_out[i]!=matrix_out1[i]) ret = -1; if(ret < 0) break; } - + VRFY((ret >= 0),"H5D irregular collective write succeed"); /* @@ -670,7 +670,7 @@ void coll_write_test(int chunk_factor) * Purpose: To test the collectively irregular hyperslab read in chunk storage * Input: number of chunks on each dimension - if number is equal to 0, contiguous storage + if number is equal to 0, contiguous storage * Return: Success: 0 * * Failure: -1 @@ -679,8 +679,8 @@ void coll_write_test(int chunk_factor) * Dec 2nd, 2004 * * Modifications: Oct 18th, 2005 - * Note: This test must be used with the correpsonding - coll_write_test. + * Note: This test must be used with the correpsonding + coll_write_test. *------------------------------------------------------------------------- */ void coll_read_test(int chunk_factor) @@ -700,7 +700,7 @@ void coll_read_test(int chunk_factor) dataset on the disk */ #endif - hsize_t mdim[2]; + hsize_t mdim[2]; hsize_t start[2]; /* Start of hyperslab */ hsize_t stride[2]; /* Stride of hyperslab */ hsize_t count[2]; /* Block count */ @@ -733,7 +733,7 @@ void coll_read_test(int chunk_factor) /* Initialize the buffer */ - + mdim[0] = MSPACE_DIM1; mdim[1] = MSPACE_DIM2*mpi_size; matrix_out =(int*)HDmalloc(sizeof(int)*MSPACE_DIM1*MSPACE_DIM2*mpi_size); @@ -766,7 +766,7 @@ void coll_read_test(int chunk_factor) /* The First selection for FILE to read * * block (1,1) - * stride(1.1) + * stride(1.1) * count (3,768/mpi_size) * start (1,2+768*mpi_rank/mpi_size) * @@ -786,7 +786,7 @@ void coll_read_test(int chunk_factor) /* The Second selection for FILE to read * * block (1,1) - * stride(1.1) + * stride(1.1) * count (3,1536/mpi_size) * start (2,4+1536*mpi_rank/mpi_size) * @@ -815,7 +815,7 @@ void coll_read_test(int chunk_factor) * Only the starting point is different. * The first selection * block (1,1) - * stride(1.1) + * stride(1.1) * count (3,768/mpi_size) * start (0,768*mpi_rank/mpi_size) * @@ -838,7 +838,7 @@ void coll_read_test(int chunk_factor) * Only the starting point is different. * The second selection * block (1,1) - * stride(1,1) + * stride(1,1) * count (3,1536/mpi_size) * start (1,2+1536*mpi_rank/mpi_size) * diff --git a/testpar/testphdf5.c b/testpar/testphdf5.c index 9b56b24..8aa3d0d 100644 --- a/testpar/testphdf5.c +++ b/testpar/testphdf5.c @@ -412,10 +412,10 @@ int main(int argc, char **argv) "independent group and dataset read", &collngroups_params); /* By default, do not run big dataset on WIN32. */ #ifdef WIN32 - AddTest("-bigdset", big_dataset, NULL, + AddTest("-bigdset", big_dataset, NULL, "big dataset test", PARATESTFILE); #else - AddTest("bigdset", big_dataset, NULL, + AddTest("bigdset", big_dataset, NULL, "big dataset test", PARATESTFILE); #endif AddTest("fill", dataset_fillvalue, NULL, @@ -453,8 +453,8 @@ int main(int argc, char **argv) AddTest((mpi_size < 3)? "-cchunk10" : "cchunk10", coll_chunk10,NULL, "multiple chunk collective IO transferring to independent IO",PARATESTFILE); - - + + /* irregular collective IO tests*/ AddTest("ccontw", |