summaryrefslogtreecommitdiffstats
path: root/testpar/t_cache.c
diff options
context:
space:
mode:
Diffstat (limited to 'testpar/t_cache.c')
-rw-r--r--testpar/t_cache.c78
1 files changed, 48 insertions, 30 deletions
diff --git a/testpar/t_cache.c b/testpar/t_cache.c
index 8696092..8559afb 100644
--- a/testpar/t_cache.c
+++ b/testpar/t_cache.c
@@ -126,7 +126,7 @@ int total_writes = 0;
*
* local_pinned: Boolean flag that is set to true iff the entry
* has been pinned in the local cache, but probably not all
- * caches. Such pins will typically not be consistant across
+ * caches. Such pins will typically not be consistent across
* processes, and thus cannot be marked as dirty unless they
* happen to overlap some collective operation.
*
@@ -205,7 +205,7 @@ struct datum data[NUM_DATA_ENTRIES];
* even divisor of NUM_DATA_ENTRIES. So far, all tests have been with
* powers of 10 that meet these criteria.
*
- * Further, this value must be consistant across all processes.
+ * Further, this value must be consistent across all processes.
*/
#define STD_VIRT_NUM_DATA_ENTRIES NUM_DATA_ENTRIES
@@ -252,7 +252,7 @@ int data_index[NUM_DATA_ENTRIES];
*
* Strangely, at least on Phoenix, the first solution runs faster by a
* rather large margin. However, I can imagine this changing with
- * different OS's and MPI implementatins.
+ * different OS's and MPI implementations.
*
* Thus I have left code supporting the second solution in place.
*
@@ -1282,7 +1282,7 @@ reset_server_counters(void)
* Function: server_main()
*
* Purpose: Main function for the server process. This process exists
- * to provide an independant view of the data array.
+ * to provide an independent view of the data array.
*
* The function handles request from the other processes in
* the test until the count of done messages received equals
@@ -3962,7 +3962,7 @@ setup_cache_for_test(hid_t *fid_ptr, H5F_t **file_ptr_ptr, H5C_t **cache_ptr_ptr
* Purpose: Verify that the indicated entries have been written exactly
* once each, and that the indicated total number of writes
* has been processed by the server process. Flag an error if
- * discrepency is noted. Finally reset the counters maintained
+ * discrepancy is noted. Finally reset the counters maintained
* by the server process.
*
* This function should only be called by the metadata cache
@@ -4050,7 +4050,7 @@ verify_writes(unsigned num_writes, haddr_t *written_entries_tbl)
/* final barrier to ensure that all processes think that the server
* counters have been reset before we leave the sync point. This
- * barrier is probaby not necessary at this point in time (5/9/10),
+ * barrier is probably not necessary at this point in time (5/9/10),
* but I can think of at least one likely change to the metadata write
* strategies that will require it -- hence its insertion now.
*/
@@ -4794,7 +4794,7 @@ server_smoke_check(void)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5141,7 +5141,7 @@ smoke_check_1(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5305,7 +5305,7 @@ smoke_check_2(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5453,7 +5453,7 @@ smoke_check_2(int metadata_write_strategy)
*
* Purpose: Third smoke check for the parallel cache.
*
- * Use random reads to vary the loads on the diffferent
+ * Use random reads to vary the loads on the different
* processors. Also force different cache size adjustments.
*
* In this test, load process 0 heavily, and the other
@@ -5510,7 +5510,7 @@ smoke_check_3(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -5738,7 +5738,7 @@ smoke_check_3(int metadata_write_strategy)
*
* Purpose: Fourth smoke check for the parallel cache.
*
- * Use random reads to vary the loads on the diffferent
+ * Use random reads to vary the loads on the different
* processors. Also force different cache size adjustments.
*
* In this test, load process 0 lightly, and the other
@@ -5795,7 +5795,7 @@ smoke_check_4(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6070,7 +6070,7 @@ smoke_check_5(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6355,7 +6355,7 @@ trace_file_check(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose)
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6488,7 +6488,7 @@ trace_file_check(int metadata_write_strategy)
} /* end if */
if (nerrors == 0) {
- HDsprintf(trace_file_name, "t_cache_trace.txt.%d", (int)file_mpi_rank);
+ HDsnprintf(trace_file_name, sizeof(trace_file_name), "t_cache_trace.txt.%d", (int)file_mpi_rank);
if ((trace_file_ptr = HDfopen(trace_file_name, "r")) == NULL) {
@@ -6623,13 +6623,15 @@ trace_file_check(int metadata_write_strategy)
static hbool_t
smoke_check_6(int metadata_write_strategy)
{
- hbool_t success = TRUE;
- int i;
- int max_nerrors;
- hid_t fid = -1;
- H5F_t * file_ptr = NULL;
- H5C_t * cache_ptr = NULL;
- struct mssg_t mssg;
+ H5P_coll_md_read_flag_t md_reads_file_flag;
+ hbool_t md_reads_context_flag;
+ hbool_t success = TRUE;
+ int i;
+ int max_nerrors;
+ hid_t fid = -1;
+ H5F_t * file_ptr = NULL;
+ H5C_t * cache_ptr = NULL;
+ struct mssg_t mssg;
switch (metadata_write_strategy) {
@@ -6660,7 +6662,7 @@ smoke_check_6(int metadata_write_strategy)
if (!server_main()) {
- /* some error occured in the server -- report failure */
+ /* some error occurred in the server -- report failure */
nerrors++;
if (verbose) {
HDfprintf(stdout, "%d:%s: server_main() failed.\n", world_mpi_rank, __func__);
@@ -6685,7 +6687,9 @@ smoke_check_6(int metadata_write_strategy)
virt_num_data_entries = NUM_DATA_ENTRIES;
/* insert the first half collectively */
- H5CX_set_coll_metadata_read(TRUE);
+ md_reads_file_flag = H5P_USER_TRUE;
+ md_reads_context_flag = TRUE;
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
for (i = 0; i < virt_num_data_entries / 2; i++) {
struct datum *entry_ptr;
entry_ptr = &(data[i]);
@@ -6704,9 +6708,13 @@ smoke_check_6(int metadata_write_strategy)
H5_CHECK_OVERFLOW(cache_ptr->max_cache_size, size_t, double);
HDassert((double)cache_ptr->max_cache_size * 0.8 > cache_ptr->coll_list_size);
}
+ /* Restore collective metadata reads state */
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
/* insert the other half independently */
- H5CX_set_coll_metadata_read(FALSE);
+ md_reads_file_flag = H5P_USER_FALSE;
+ md_reads_context_flag = FALSE;
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
for (i = virt_num_data_entries / 2; i < virt_num_data_entries; i++) {
struct datum *entry_ptr;
entry_ptr = &(data[i]);
@@ -6716,7 +6724,7 @@ smoke_check_6(int metadata_write_strategy)
if (FALSE != entry_ptr->header.coll_access) {
nerrors++;
if (verbose) {
- HDfprintf(stdout, "%d:%s: Entry inserted indepedently marked as collective.\n",
+ HDfprintf(stdout, "%d:%s: Entry inserted independently marked as collective.\n",
world_mpi_rank, __func__);
}
}
@@ -6724,6 +6732,8 @@ smoke_check_6(int metadata_write_strategy)
/* Make sure coll entries do not cross the 80% threshold */
HDassert((double)cache_ptr->max_cache_size * 0.8 > cache_ptr->coll_list_size);
}
+ /* Restore collective metadata reads state */
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
/* flush the file */
if (H5Fflush(fid, H5F_SCOPE_GLOBAL) < 0) {
@@ -6734,7 +6744,9 @@ smoke_check_6(int metadata_write_strategy)
}
/* Protect the first half of the entries collectively */
- H5CX_set_coll_metadata_read(TRUE);
+ md_reads_file_flag = H5P_USER_TRUE;
+ md_reads_context_flag = TRUE;
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
for (i = 0; i < (virt_num_data_entries / 2); i++) {
struct datum *entry_ptr;
entry_ptr = &(data[i]);
@@ -6752,9 +6764,13 @@ smoke_check_6(int metadata_write_strategy)
/* Make sure coll entries do not cross the 80% threshold */
HDassert((double)cache_ptr->max_cache_size * 0.8 > cache_ptr->coll_list_size);
}
+ /* Restore collective metadata reads state */
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
/* protect the other half independently */
- H5CX_set_coll_metadata_read(FALSE);
+ md_reads_file_flag = H5P_USER_FALSE;
+ md_reads_context_flag = FALSE;
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
for (i = virt_num_data_entries / 2; i < virt_num_data_entries; i++) {
struct datum *entry_ptr;
entry_ptr = &(data[i]);
@@ -6764,7 +6780,7 @@ smoke_check_6(int metadata_write_strategy)
if (FALSE != entry_ptr->header.coll_access) {
nerrors++;
if (verbose) {
- HDfprintf(stdout, "%d:%s: Entry inserted indepedently marked as collective.\n",
+ HDfprintf(stdout, "%d:%s: Entry inserted independently marked as collective.\n",
world_mpi_rank, __func__);
}
}
@@ -6772,6 +6788,8 @@ smoke_check_6(int metadata_write_strategy)
/* Make sure coll entries do not cross the 80% threshold */
HDassert((double)cache_ptr->max_cache_size * 0.8 > cache_ptr->coll_list_size);
}
+ /* Restore collective metadata reads state */
+ H5F_set_coll_metadata_reads(file_ptr, &md_reads_file_flag, &md_reads_context_flag);
for (i = 0; i < (virt_num_data_entries); i++) {
unlock_entry(file_ptr, i, H5AC__NO_FLAGS_SET);